Search results for: fixed live camera images
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 5302

Search results for: fixed live camera images

4852 Application Methodology for the Generation of 3D Thermal Models Using UAV Photogrammety and Dual Sensors for Mining/Industrial Facilities Inspection

Authors: Javier Sedano-Cibrián, Julio Manuel de Luis-Ruiz, Rubén Pérez-Álvarez, Raúl Pereda-García, Beatriz Malagón-Picón

Abstract:

Structural inspection activities are necessary to ensure the correct functioning of infrastructures. Unmanned Aerial Vehicle (UAV) techniques have become more popular than traditional techniques. Specifically, UAV Photogrammetry allows time and cost savings. The development of this technology has permitted the use of low-cost thermal sensors in UAVs. The representation of 3D thermal models with this type of equipment is in continuous evolution. The direct processing of thermal images usually leads to errors and inaccurate results. A methodology is proposed for the generation of 3D thermal models using dual sensors, which involves the application of visible Red-Blue-Green (RGB) and thermal images in parallel. Hence, the RGB images are used as the basis for the generation of the model geometry, and the thermal images are the source of the surface temperature information that is projected onto the model. Mining/industrial facilities representations that are obtained can be used for inspection activities.

Keywords: aerial thermography, data processing, drone, low-cost, point cloud

Procedia PDF Downloads 143
4851 Comparative Evaluation of Vanishing Interfacial Tension Approach for Minimum Miscibility Pressure Determination

Authors: Waqar Ahmad Butt, Gholamreza Vakili Nezhaad, Ali Soud Al Bemani, Yahya Al Wahaibi

Abstract:

Minimum miscibility pressure (MMP) plays a great role in determining the displacement efficiency of different gas injection processes. Experimental techniques for MMP determination include industrially recommended slim tube, vanishing interfacial tension (VIT) and rising bubble apparatus (RBA). In this paper, MMP measurement study using slim tube and VIT experimental techniques for two different crude oil samples (M and N) both in live and stock tank oil forms is being presented. VIT measured MMP values for both 'M' and 'N' live crude oils were close to slim tube determined MMP values with 6.4 and 5 % deviation respectively. Whereas for both oil samples in stock tank oil form, VIT measured MMP showed a higher unacceptable deviation from slim tube determined MMP. This higher difference appears to be related to high stabilized crude oil heavier fraction and lack of multiple contacts miscibility. None of the different nine deployed crude oil and CO2 MMP computing correlations could result in reliable MMP, close to slim tube determined MMP. Since VIT determined MMP values for both considered live crude oils are in close match with slim tube determined MMP values, it confirms reliable, reproducible, rapid and cheap alternative for live crude oil MMP determination. Whereas VIT MMP determination for stock tank oil case needed further investigation about stabilization / destabilization mechanism of oil heavier ends and multiple contacts miscibility development issues.

Keywords: minimum miscibility pressure, interfacial tension, multiple contacts miscibility, heavier ends

Procedia PDF Downloads 268
4850 Spatio-Temporal Dynamic of Woody Vegetation Assessment Using Oblique Landscape Photographs

Authors: V. V. Fomin, A. P. Mikhailovich, E. M. Agapitov, V. E. Rogachev, E. A. Kostousova, E. S. Perekhodova

Abstract:

Ground-level landscape photos can be used as a source of objective data on woody vegetation and vegetation dynamics. We proposed a method for processing, analyzing, and presenting ground photographs, which has the following advantages: 1) researcher has to form holistic representation of the study area in form of a set of interlapping ground-level landscape photographs; 2) it is necessary to define or obtain characteristics of the landscape, objects, and phenomena present on the photographs; 3) it is necessary to create new or supplement existing textual descriptions and annotations for the ground-level landscape photographs; 4) single or multiple ground-level landscape photographs can be used to develop specialized geoinformation layers, schematic maps or thematic maps; 5) it is necessary to determine quantitative data that describes both images as a whole, and displayed objects and phenomena, using algorithms for automated image analysis. It is suggested to match each photo with a polygonal geoinformation layer, which is a sector consisting of areas corresponding with parts of the landscape visible in the photos. Calculation of visibility areas is performed in a geoinformation system within a sector using a digital model of a study area relief and visibility analysis functions. Superposition of the visibility sectors corresponding with various camera viewpoints allows matching landscape photos with each other to create a complete and wholesome representation of the space in question. It is suggested to user-defined data or phenomenons on the images with the following superposition over the visibility sector in the form of map symbols. The technology of geoinformation layers’ spatial superposition over the visibility sector creates opportunities for image geotagging using quantitative data obtained from raster or vector layers within the sector with the ability to generate annotations in natural language. The proposed method has proven itself well for relatively open and clearly visible areas with well-defined relief, for example, in mountainous areas in the treeline ecotone. When the polygonal layers of visibility sectors for a large number of different points of photography are topologically superimposed, a layer of visibility of sections of the entire study area is formed, which is displayed in the photographs. Also, as a result of this overlapping of sectors, areas that did not appear in the photo will be assessed as gaps. According to the results of this procedure, it becomes possible to obtain information about the photos that display a specific area and from which points of photography it is visible. This information may be obtained either as a query on the map or as a query for the attribute table of the layer. The method was tested using repeated photos taken from forty camera viewpoints located on Ray-Iz mountain massif (Polar Urals, Russia) from 1960 until 2023. It has been successfully used in combination with other ground-based and remote sensing methods of studying the climate-driven dynamics of woody vegetation in the Polar Urals. Acknowledgment: This research was collaboratively funded by the Russian Ministry for Science and Education project No. FEUG-2023-0002 (image representation) and Russian Science Foundation project No. 24-24-00235 (automated textual description).

Keywords: woody, vegetation, repeated, photographs

Procedia PDF Downloads 89
4849 Bhumastra “Unmanned Ground Vehicle”

Authors: Vivek Krishna, Nikhil Jain, A. Mary Posonia A., Albert Mayan J

Abstract:

Terrorism and insurgency are significant global issues that require constant attention and effort from governments and scientists worldwide. To combat these threats, nations invest billions of dollars in developing new defensive technologies to protect civilians. Breakthroughs in vehicle automation have led to the use of sophisticated machines for many dangerous and critical anti-terrorist activities. Our concept of an "Unmanned Ground Vehicle" can carry out tasks such as border security, surveillance, mine detection, and active combat independently or in tandem with human control. The robot's movement can be wirelessly controlled by a person in a distant location or can travel to a pre-programmed destination autonomously in situations where personal control is not feasible. Our defence system comprises two units: the control unit that regulates mobility and the motion tracking unit. The remote operator robot uses the camera's live visual feed to manually operate both units, and the rover can automatically detect movement. The rover is operated by manpower who controls it using a joystick or mouse, and a wireless modem enables a soldier in a combat zone to control the rover via an additional controller feature.

Keywords: robotics, computer vision, Machine learning, Artificial intelligence, future of AI

Procedia PDF Downloads 125
4848 Regression Model Evaluation on Depth Camera Data for Gaze Estimation

Authors: James Purnama, Riri Fitri Sari

Abstract:

We investigate the machine learning algorithm selection problem in the term of a depth image based eye gaze estimation, with respect to its essential difficulty in reducing the number of required training samples and duration time of training. Statistics based prediction accuracy are increasingly used to assess and evaluate prediction or estimation in gaze estimation. This article evaluates Root Mean Squared Error (RMSE) and R-Squared statistical analysis to assess machine learning methods on depth camera data for gaze estimation. There are 4 machines learning methods have been evaluated: Random Forest Regression, Regression Tree, Support Vector Machine (SVM), and Linear Regression. The experiment results show that the Random Forest Regression has the lowest RMSE and the highest R-Squared, which means that it is the best among other methods.

Keywords: gaze estimation, gaze tracking, eye tracking, kinect, regression model, orange python

Procedia PDF Downloads 538
4847 Reliability Based Analysis of Multi-Lane Reinforced Concrete Slab Bridges

Authors: Ali Mahmoud, Shadi Najjar, Mounir Mabsout, Kassim Tarhini

Abstract:

Empirical expressions for estimating the wheel load distribution and live-load bending moment are typically specified in highway bridge codes such as the AASHTO procedures. The purpose of this paper is to analyze the reliability levels that are inherent in reinforced concrete slab bridges that are designed based on the simplified empirical live load equations in the AASHTO LRFD procedures. To achieve this objective, bridges with multi-lanes (three and four lanes) and different spans are modeled using finite-element analysis (FEA) subjected to HS20 truck loading, tandem loading, and standard lane loading per AASHTO LRFD procedures. The FEA results are compared with the AASHTO LRFD moments in order to quantify the biases that might result from the simplifying assumptions adopted in AASHTO. A reliability analysis is conducted to quantify the reliability index for bridges designed using AASHTO procedures. To reach a consistent level of safety for three- and four-lane bridges, following a previous study restricted to one- and two-lane bridges, the live load factor in the design equation proposed by AASHTO LRFD will be assessed and revised if needed by alternating the live load factor for these lanes. The results will provide structural engineers with more consistent provisions to design concrete slab bridges or evaluate the load-carrying capacity of existing bridges.

Keywords: reliability analysis of concrete bridges, finite element modeling, reliability analysis, reinforced concrete bridge design, load carrying capacity

Procedia PDF Downloads 341
4846 School Curriculum Incorporating Rights to Live in Clean and Healthy Environment: Assessing Its Effectiveness

Authors: Sitaram Dahal

Abstract:

Among many strategic and practical needs in overcoming the threats and challenges being experienced in the global environment, constitutional provision for Rights to live in clean and healthy environment is one and so is the school curriculum incorporating information on such rights. Government of Nepal has also introduced information on rights to live in clean and healthy environment, as provisioned in its interim constitution of 2007, in the secondary level curriculum of formal education. As the predetermined specific objective of such curriculum is to prepare students who are conscious of citizens’ rights and responsibilities and are able to adopt functions, duties and rights of the rights holders and duty bearers; the study was designed to assess the effectiveness of such curriculum. The study was conducted in one private school and a community school to assess the effectiveness of such curriculum. The study shows that such curriculum has been able to make students responsible duty bearers as they were aware of their habits towards environment. Whereas only very few students are aware enough as being rights holders. Students of community schools were aware rights holders as they complain if they are not satisfied with the environment of the school itself. But private school is far behind in this case. It can be said that only curriculum with very few portion of information on such rights might not be capable enough to meet its objective.

Keywords: curriculum, environmental rights, constitution, effectiveness

Procedia PDF Downloads 326
4845 An Empirical Analysis of the Impact of Selected Macroeconomic Variables on Capital Formation in Libya (1970–2010)

Authors: Khaled Ramadan Elbeydi

Abstract:

This study is carried out to provide an insight into the analysis of the impact of selected macro-economic variables on gross fixed capital formation in Libya using annual data over the period (1970-2010). The importance of this study comes from the ability to show the relative important factors that impact the Libyan gross fixed capital formation. This understanding would give indications to decision makers on which policy they must focus to stimulate the economy. An Auto-Regressive Distributed Lag (ARDL) modeling process is employed to investigate the impact of the gross domestic product, monetary base, and trade openness on gross fixed capital formation in Libya. The results of this study reveal that there is an equilibrium relationship between capital formation and its determinants. The results also indicate that GDP and trade openness largely explain the pattern of capital formation in Libya. The findings and recommendations provide vital information relevant for policy formulation and implementation aimed to improve capital formation in Libya.

Keywords: ARDL, bounds test, capital formation, co-integration, Libya

Procedia PDF Downloads 365
4844 RoboWeedSupport-Semi-Automated Unmanned Aerial System for Cost Efficient High Resolution in Sub-Millimeter Scale Acquisition of Weed Images

Authors: Simon L. Madsen, Mads Dyrmann, Morten S. Laursen, Rasmus N. Jørgensen

Abstract:

Recent advances in the Unmanned Aerial System (UAS) safety and perception systems enable safe low altitude autonomous terrain following flights recently demonstrated by the consumer DJI Mavic PRO and Phamtom 4 Pro drones. This paper presents the first prototype system utilizing this functionality in form of semi-automated UAS based collection of crop/weed images where the embedded perception system ensures a significantly safer and faster gathering of weed images with sub-millimeter resolution. The system is to be used when the weeds are at cotyledon stage and prior to the harvest recognizing the grass weed species, which cannot be discriminated at the cotyledon stage.

Keywords: weed mapping, UAV, DJI SDK, automation, cotyledon plants

Procedia PDF Downloads 309
4843 The Effect of Saccharomyces cerevisiae Live Yeast Culture on Microbial Nitrogen Supply to Small Intestine in Male Kivircik Yearlings Fed with Different Ratio of Forage and Concentrate

Authors: Nurcan Cetinkaya, Nadide Hulya Ozdemir

Abstract:

The aim of the study was to investigate the effect of Saccharomyces cerevisiae (SC) live yeast culture on microbial protein supply to the small intestine in Kivircik male yearlings when fed with different ratio of forage and concentrate diets. Four Kivircik male yearlings with permanent rumen canula were used in the experiment. . The treatments were allocated to a 4x4 Latin square design. Diet I consisted of 70% alfalfa hay and 30% concentrate, Diet II consisted of 30% alfalfa hay and 70% concentrate, Diet I and II were supplemented with a SC. Daily urine was collected and stored at -20°C until analysis. Calorimetric methods were used for the determination of urinary allantoin and creatinin levels. The estimated microbial N supply to small intestine for Diets I, I+SC, II and II+SC were 2.51, 2.64, 2.95 and 3.43 g N/d respectively. Supplementation of Diets I and II with SC significantly affected the allantoin levels in µmol/W0. 75 (p<0.05). Mean creatinine values in µmol/W0. 75 and allantoin:creatinin ratios were not significantly different among diets. In conclusion, supplementation with SC live yeast culture had a significant effect on urinary allantoin excretion and microbial protein supply to small intestine in Kivircik yearlings fed with high concentrate Diet II (P<0.05). Hence urinary allantoin excretion may be used as a tool for estimating microbial protein supply in Kivircık yearlings. However, further studies are necessary to understand the metabolism of Saccharomyces cerevisiae live yeast culture with different forage: concentrate ratio in Kıvırcık Yearlings.

Keywords: allantoin, creatinin, Kivircik yearling, microbial nitrogen, Saccharomyces cerevisia

Procedia PDF Downloads 413
4842 Image Denoising Using Spatial Adaptive Mask Filter for Medical Images

Authors: R. Sumalatha, M. V. Subramanyam

Abstract:

In medical image processing the quality of the image is degraded in the presence of noise. Especially in ultra sound imaging and Magnetic resonance imaging the data was corrupted by signal dependent noise known as salt and pepper noise. Removal of noise from the medical images is a critical issue for researchers. In this paper, a new type of technique Adaptive Spatial Mask Filter (ASMF) has been proposed. The proposed filter is used to increase the quality of MRI and ultra sound images. Experimental results show that the proposed filter outperforms the implementation of mean, median, adaptive median filters in terms of MSE and PSNR.

Keywords: salt and pepper noise, ASMF, PSNR, MSE

Procedia PDF Downloads 436
4841 Spatially Encoded Hyperspectral Compressive Microscope for Broadband VIS/NIR Imaging

Authors: Lukáš Klein, Karel Žídek

Abstract:

Hyperspectral imaging counts among the most frequently used multidimensional sensing methods. While there are many approaches to capturing a hyperspectral data cube, optical compression is emerging as a valuable tool to reduce the setup complexity and the amount of data storage needed. Hyperspectral compressive imagers have been created in the past; however, they have primarily focused on relatively narrow sections of the electromagnetic spectrum. A broader spectral study of samples can provide helpful information, especially for applications involving the harmonic generation and advanced material characterizations. We demonstrate a broadband hyperspectral microscope based on the single-pixel camera principle. Captured spatially encoded data are processed to reconstruct a hyperspectral cube in a combined visible and near-infrared spectrum (from 400 to 2500 nm). Hyperspectral cubes can be reconstructed with a spectral resolution of up to 3 nm and spatial resolution of up to 7 µm (subject to diffraction) with a high compressive ratio.

Keywords: compressive imaging, hyperspectral imaging, near-infrared spectrum, single-pixel camera, visible spectrum

Procedia PDF Downloads 89
4840 Protocol for Dynamic Load Distributed Low Latency Web-Based Augmented Reality and Virtual Reality

Authors: Rohit T. P., Sahil Athrij, Sasi Gopalan

Abstract:

Currently, the content entertainment industry is dominated by mobile devices. As the trends slowly shift towards Augmented/Virtual Reality applications the computational demands on these devices are increasing exponentially and we are already reaching the limits of hardware optimizations. This paper proposes a software solution to this problem. By leveraging the capabilities of cloud computing we can offload the work from mobile devices to dedicated rendering servers that are way more powerful. But this introduces the problem of latency. This paper introduces a protocol that can achieve high-performance low latency Augmented/Virtual Reality experience. There are two parts to the protocol, 1) In-flight compression The main cause of latency in the system is the time required to transmit the camera frame from client to server. The round trip time is directly proportional to the amount of data transmitted. This can therefore be reduced by compressing the frames before sending. Using some standard compression algorithms like JPEG can result in minor size reduction only. Since the images to be compressed are consecutive camera frames there won't be a lot of changes between two consecutive images. So inter-frame compression is preferred. Inter-frame compression can be implemented efficiently using WebGL but the implementation of WebGL limits the precision of floating point numbers to 16bit in most devices. This can introduce noise to the image due to rounding errors, which will add up eventually. This can be solved using an improved interframe compression algorithm. The algorithm detects changes between frames and reuses unchanged pixels from the previous frame. This eliminates the need for floating point subtraction thereby cutting down on noise. The change detection is also improved drastically by taking the weighted average difference of pixels instead of the absolute difference. The kernel weights for this comparison can be fine-tuned to match the type of image to be compressed. 2) Dynamic Load distribution Conventional cloud computing architectures work by offloading as much work as possible to the servers, but this approach can cause a hit on bandwidth and server costs. The most optimal solution is obtained when the device utilizes 100% of its resources and the rest is done by the server. The protocol balances the load between the server and the client by doing a fraction of the computing on the device depending on the power of the device and network conditions. The protocol will be responsible for dynamically partitioning the tasks. Special flags will be used to communicate the workload fraction between the client and the server and will be updated in a constant interval of time ( or frames ). The whole of the protocol is designed so that it can be client agnostic. Flags are available to the client for resetting the frame, indicating latency, switching mode, etc. The server can react to client-side changes on the fly and adapt accordingly by switching to different pipelines. The server is designed to effectively spread the load and thereby scale horizontally. This is achieved by isolating client connections into different processes.

Keywords: 2D kernelling, augmented reality, cloud computing, dynamic load distribution, immersive experience, mobile computing, motion tracking, protocols, real-time systems, web-based augmented reality application

Procedia PDF Downloads 74
4839 Performences of Type-2 Fuzzy Logic Control and Neuro-Fuzzy Control Based on DPC for Grid Connected DFIG with Fixed Switching Frequency

Authors: Fayssal Amrane, Azeddine Chaiba

Abstract:

In this paper, type-2 fuzzy logic control (T2FLC) and neuro-fuzzy control (NFC) for a doubly fed induction generator (DFIG) based on direct power control (DPC) with a fixed switching frequency is proposed for wind generation application. First, a mathematical model of the doubly-fed induction generator implemented in d-q reference frame is achieved. Then, a DPC algorithm approach for controlling active and reactive power of DFIG via fixed switching frequency is incorporated using PID. The performance of T2FLC and NFC, which is based on the DPC algorithm, are investigated and compared to those obtained from the PID controller. Finally, simulation results demonstrate that the NFC is more robust, superior dynamic performance for wind power generation system applications.

Keywords: doubly fed induction generator (DFIG), direct power control (DPC), neuro-fuzzy control (NFC), maximum power point tracking (MPPT), space vector modulation (SVM), type 2 fuzzy logic control (T2FLC)

Procedia PDF Downloads 420
4838 A Fuzzy Approach to Liver Tumor Segmentation with Zernike Moments

Authors: Abder-Rahman Ali, Antoine Vacavant, Manuel Grand-Brochier, Adélaïde Albouy-Kissi, Jean-Yves Boire

Abstract:

In this paper, we present a new segmentation approach for liver lesions in regions of interest within MRI (Magnetic Resonance Imaging). This approach, based on a two-cluster Fuzzy C-Means methodology, considers the parameter variable compactness to handle uncertainty. Fine boundaries are detected by a local recursive merging of ambiguous pixels with a sequential forward floating selection with Zernike moments. The method has been tested on both synthetic and real images. When applied on synthetic images, the proposed approach provides good performance, segmentations obtained are accurate, their shape is consistent with the ground truth, and the extracted information is reliable. The results obtained on MR images confirm such observations. Our approach allows, even for difficult cases of MR images, to extract a segmentation with good performance in terms of accuracy and shape, which implies that the geometry of the tumor is preserved for further clinical activities (such as automatic extraction of pharmaco-kinetics properties, lesion characterization, etc).

Keywords: defuzzification, floating search, fuzzy clustering, Zernike moments

Procedia PDF Downloads 452
4837 Computer Aided Diagnostic System for Detection and Classification of a Brain Tumor through MRI Using Level Set Based Segmentation Technique and ANN Classifier

Authors: Atanu K Samanta, Asim Ali Khan

Abstract:

Due to the acquisition of huge amounts of brain tumor magnetic resonance images (MRI) in clinics, it is very difficult for radiologists to manually interpret and segment these images within a reasonable span of time. Computer-aided diagnosis (CAD) systems can enhance the diagnostic capabilities of radiologists and reduce the time required for accurate diagnosis. An intelligent computer-aided technique for automatic detection of a brain tumor through MRI is presented in this paper. The technique uses the following computational methods; the Level Set for segmentation of a brain tumor from other brain parts, extraction of features from this segmented tumor portion using gray level co-occurrence Matrix (GLCM), and the Artificial Neural Network (ANN) to classify brain tumor images according to their respective types. The entire work is carried out on 50 images having five types of brain tumor. The overall classification accuracy using this method is found to be 98% which is significantly good.

Keywords: brain tumor, computer-aided diagnostic (CAD) system, gray-level co-occurrence matrix (GLCM), tumor segmentation, level set method

Procedia PDF Downloads 512
4836 Unsupervised Images Generation Based on Sloan Digital Sky Survey with Deep Convolutional Generative Neural Networks

Authors: Guanghua Zhang, Fubao Wang, Weijun Duan

Abstract:

Convolution neural network (CNN) has attracted more and more attention on recent years. Especially in the field of computer vision and image classification. However, unsupervised learning with CNN has received less attention than supervised learning. In this work, we use a new powerful tool which is deep convolutional generative adversarial networks (DCGANs) to generate images from Sloan Digital Sky Survey. Training by various star and galaxy images, it shows that both the generator and the discriminator are good for unsupervised learning. In this paper, we also took several experiments to choose the best value for hyper-parameters and which could help to stabilize the training process and promise a good quality of the output.

Keywords: convolution neural network, discriminator, generator, unsupervised learning

Procedia PDF Downloads 268
4835 Multiscale Connected Component Labelling and Applications to Scientific Microscopy Image Processing

Authors: Yayun Hsu, Henry Horng-Shing Lu

Abstract:

In this paper, a new method is proposed to extending the method of connected component labeling from processing binary images to multi-scale modeling of images. By using the adaptive threshold of multi-scale attributes, this approach minimizes the possibility of missing those important components with weak intensities. In addition, the computational cost of this approach remains similar to that of the typical approach of component labeling. Then, this methodology is applied to grain boundary detection and Drosophila Brain-bow neuron segmentation. These demonstrate the feasibility of the proposed approach in the analysis of challenging microscopy images for scientific discovery.

Keywords: microscopic image processing, scientific data mining, multi-scale modeling, data mining

Procedia PDF Downloads 435
4834 Routing Medical Images with Tabu Search and Simulated Annealing: A Study on Quality of Service

Authors: Mejía M. Paula, Ramírez L. Leonardo, Puerta A. Gabriel

Abstract:

In telemedicine, the image repository service is important to increase the accuracy of diagnostic support of medical personnel. This study makes comparison between two routing algorithms regarding the quality of service (QoS), to be able to analyze the optimal performance at the time of loading and/or downloading of medical images. This study focused on comparing the performance of Tabu Search with other heuristic and metaheuristic algorithms that improve QoS in telemedicine services in Colombia. For this, Tabu Search and Simulated Annealing heuristic algorithms are chosen for their high usability in this type of applications; the QoS is measured taking into account the following metrics: Delay, Throughput, Jitter and Latency. In addition, routing tests were carried out on ten images in digital image and communication in medicine (DICOM) format of 40 MB. These tests were carried out for ten minutes with different traffic conditions, reaching a total of 25 tests, from a server of Universidad Militar Nueva Granada (UMNG) in Bogotá-Colombia to a remote user in Universidad de Santiago de Chile (USACH) - Chile. The results show that Tabu search presents a better QoS performance compared to Simulated Annealing, managing to optimize the routing of medical images, a basic requirement to offer diagnostic images services in telemedicine.

Keywords: medical image, QoS, simulated annealing, Tabu search, telemedicine

Procedia PDF Downloads 219
4833 KCBA, A Method for Feature Extraction of Colonoscopy Images

Authors: Vahid Bayrami Rad

Abstract:

In recent years, the use of artificial intelligence techniques, tools, and methods in processing medical images and health-related applications has been highlighted and a lot of research has been done in this regard. For example, colonoscopy and diagnosis of colon lesions are some cases in which the process of diagnosis of lesions can be improved by using image processing and artificial intelligence algorithms, which help doctors a lot. Due to the lack of accurate measurements and the variety of injuries in colonoscopy images, the process of diagnosing the type of lesions is a little difficult even for expert doctors. Therefore, by using different software and image processing, doctors can be helped to increase the accuracy of their observations and ultimately improve their diagnosis. Also, by using automatic methods, the process of diagnosing the type of disease can be improved. Therefore, in this paper, a deep learning framework called KCBA is proposed to classify colonoscopy lesions which are composed of several methods such as K-means clustering, a bag of features and deep auto-encoder. Finally, according to the experimental results, the proposed method's performance in classifying colonoscopy images is depicted considering the accuracy criterion.

Keywords: colorectal cancer, colonoscopy, region of interest, narrow band imaging, texture analysis, bag of feature

Procedia PDF Downloads 57
4832 Relationship between Mannheimia haemolytica and the Fertility Characteristics of Boer Goats

Authors: Muhammad Naveed Ali

Abstract:

A study was conducted to determine the effects on the severity of pneumonia due to Mannheimia haemolytica and its relation with the fertility of bucks. A total of 12 crossbred Boer bucks of 3 treatment groups of equal number (4 goats per group, aged 12-14 months) were selected in this study. Group A was intranasally inoculated live M. haemolytica 1 × 105 (cfu). Group B was first immunized subcutaneously M. haemolytica killed vaccine (2 ml) two week before intranasal inoculation of M. haemolytica 1 × 105 (cfu). Group C was treated with normal saline (PBS) as control. Electro-ejaculator was used for semen collection once per week whilst scrotal circumference was measured before and after challenge. The semen volume, sperm concentration, sperm motility, live/dead percentage and morphology were evaluated. From the semen evaluation, goats in Group A exhibited significant decrease in the semen volume, sperm concentration, motility and live/dead sperm compared with vaccinated group B. The scrotal circumference was significantly decreased in group A compared to B. There were non-significant differences in scrotal circumferences of group B and C. The results suggested that M. haemolytica infection has negative effects on the fertility of Boer bucks.

Keywords: Boer bucks, Mannheimia haemolytica, semen evaluation, vaccination

Procedia PDF Downloads 153
4831 Experimental Modeling of Spray and Water Sheet Formation Due to Wave Interactions with Vertical and Slant Bow-Shaped Model

Authors: Armin Bodaghkhani, Bruce Colbourne, Yuri S. Muzychka

Abstract:

The process of spray-cloud formation and flow kinematics produced from breaking wave impact on vertical and slant lab-scale bow-shaped models were experimentally investigated. Bubble Image Velocimetry (BIV) and Image Processing (IP) techniques were applied to study the various types of wave-model impacts. Different wave characteristics were generated in a tow tank to investigate the effects of wave characteristics, such as wave phase velocity, wave steepness on droplet velocities, and behavior of the process of spray cloud formation. The phase ensemble-averaged vertical velocity and turbulent intensity were computed. A high-speed camera and diffused LED backlights were utilized to capture images for further post processing. Various pressure sensors and capacitive wave probes were used to measure the wave impact pressure and the free surface profile at different locations of the model and wave-tank, respectively. Droplet sizes and velocities were measured using BIV and IP techniques to trace bubbles and droplets in order to measure their velocities and sizes by correlating the texture in these images. The impact pressure and droplet size distributions were compared to several previously experimental models, and satisfactory agreements were achieved. The distribution of droplets in front of both models are demonstrated. Due to the highly transient process of spray formation, the drag coefficient for several stages of this transient displacement for various droplet size ranges and different Reynolds number were calculated based on the ensemble average method. From the experimental results, the slant model produces less spray in comparison with the vertical model, and the droplet velocities generated from the wave impact with the slant model have a lower velocity as compared with the vertical model.

Keywords: spray charachteristics, droplet size and velocity, wave-body interactions, bubble image velocimetry, image processing

Procedia PDF Downloads 300
4830 Estimating Cyclone Intensity Using INSAT-3D IR Images Based on Convolution Neural Network Model

Authors: Divvela Vishnu Sai Kumar, Deepak Arora, Sheenu Rizvi

Abstract:

Forecasting a cyclone through satellite images consists of the estimation of the intensity of the cyclone and predicting it before a cyclone comes. This research work can help people to take safety measures before the cyclone comes. The prediction of the intensity of a cyclone is very important to save lives and minimize the damage caused by cyclones. These cyclones are very costliest natural disasters that cause a lot of damage globally due to a lot of hazards. Authors have proposed five different CNN (Convolutional Neural Network) models that estimate the intensity of cyclones through INSAT-3D IR images. There are a lot of techniques that are used to estimate the intensity; the best model proposed by authors estimates intensity with a root mean squared error (RMSE) of 10.02 kts.

Keywords: estimating cyclone intensity, deep learning, convolution neural network, prediction models

Procedia PDF Downloads 128
4829 Optimization of Surface Roughness in Turning Process Utilizing Live Tooling via Taguchi Methodology

Authors: Weinian Wang, Joseph C. Chen

Abstract:

The objective of this research is to optimize the process of cutting cylindrical workpieces utilizing live tooling on a HAAS ST-20 lathe. Surface roughness (Ra) has been investigated as the indicator of quality characteristics for machining process. Aluminum alloy was used to conduct experiments due to its wide range usages in engineering structures and components where light weight or corrosion resistance is required. In this study, Taguchi methodology is utilized to determine the effects that each of the parameters has on surface roughness (Ra). A total of 18 experiments of each process were designed according to Taguchi’s L9 orthogonal array (OA) with four control factors at three levels of each and signal-to-noise ratios (S/N) were computed with Smaller the better equation for minimizing the system. The optimal parameters identified for the surface roughness of the turning operation utilizing live tooling were a feed rate of 3 inches/min(A3); a spindle speed of 1300 rpm(B3); a 2-flute titanium nitrite coated 3/8” endmill (C1); and a depth of cut of 0.025 inches (D2). The mean surface roughness of the confirmation runs in turning operation was 8.22 micro inches. The final results demonstrate that Taguchi methodology is a sufficient way of process improvement in turning process on surface roughness.

Keywords: CNC milling operation, CNC turning operation, surface roughness, Taguchi parameter design

Procedia PDF Downloads 176
4828 Thermal Image Segmentation Method for Stratification of Freezing Temperatures

Authors: Azam Fazelpour, Saeed R. Dehghani, Vlastimil Masek, Yuri S. Muzychka

Abstract:

The study uses an image analysis technique employing thermal imaging to measure the percentage of areas with various temperatures on a freezing surface. An image segmentation method using threshold values is applied to a sequence of image recording the freezing process. The phenomenon is transient and temperatures vary fast to reach the freezing point and complete the freezing process. Freezing salt water is subjected to the salt rejection that makes the freezing point dynamic and dependent on the salinity at the phase interface. For a specific area of freezing, nucleation starts from one side and end to another side, which causes a dynamic and transient temperature in that area. Thermal cameras are able to reveal a difference in temperature due to their sensitivity to infrared radiance. Using Experimental setup, a video is recorded by a thermal camera to monitor radiance and temperatures during the freezing process. Image processing techniques are applied to all frames to detect and classify temperatures on the surface. Image processing segmentation method is used to find contours with same temperatures on the icing surface. Each segment is obtained using the temperature range appeared in the image and correspond pixel values in the image. Using the contours extracted from image and camera parameters, stratified areas with different temperatures are calculated. To observe temperature contours on the icing surface using the thermal camera, the salt water sample is dropped on a cold surface with the temperature of -20°C. A thermal video is recorded for 2 minutes to observe the temperature field. Examining the results obtained by the method and the experimental observations verifies the accuracy and applicability of the method.

Keywords: ice contour boundary, image processing, image segmentation, salt ice, thermal image

Procedia PDF Downloads 321
4827 Machine Learning Approach for Mutation Testing

Authors: Michael Stewart

Abstract:

Mutation testing is a type of software testing proposed in the 1970s where program statements are deliberately changed to introduce simple errors so that test cases can be validated to determine if they can detect the errors. Test cases are executed against the mutant code to determine if one fails, detects the error and ensures the program is correct. One major issue with this type of testing was it became intensive computationally to generate and test all possible mutations for complex programs. This paper used reinforcement learning and parallel processing within the context of mutation testing for the selection of mutation operators and test cases that reduced the computational cost of testing and improved test suite effectiveness. Experiments were conducted using sample programs to determine how well the reinforcement learning-based algorithm performed with one live mutation, multiple live mutations and no live mutations. The experiments, measured by mutation score, were used to update the algorithm and improved accuracy for predictions. The performance was then evaluated on multiple processor computers. With reinforcement learning, the mutation operators utilized were reduced by 50 – 100%.

Keywords: automated-testing, machine learning, mutation testing, parallel processing, reinforcement learning, software engineering, software testing

Procedia PDF Downloads 198
4826 Comprehensive Evaluation of COVID-19 Through Chest Images

Authors: Parisa Mansour

Abstract:

The coronavirus disease 2019 (COVID-19) was discovered and rapidly spread to various countries around the world since the end of 2019. Computed tomography (CT) images have been used as an important alternative to the time-consuming RT. PCR test. However, manual segmentation of CT images alone is a major challenge as the number of suspected cases increases. Thus, accurate and automatic segmentation of COVID-19 infections is urgently needed. Because the imaging features of the COVID-19 infection are different and similar to the background, existing medical image segmentation methods cannot achieve satisfactory performance. In this work, we try to build a deep convolutional neural network adapted for the segmentation of chest CT images with COVID-19 infections. First, we maintain a large and novel chest CT image database containing 165,667 annotated chest CT images from 861 patients with confirmed COVID-19. Inspired by the observation that the boundary of an infected lung can be improved by global intensity adjustment, we introduce a feature variable block into the proposed deep CNN, which adjusts the global features of features to segment the COVID-19 infection. The proposed PV array can effectively and adaptively improve the performance of functions in different cases. We combine features of different scales by proposing a progressive atrocious space pyramid fusion scheme to deal with advanced infection regions with various aspects and shapes. We conducted experiments on data collected in China and Germany and showed that the proposed deep CNN can effectively produce impressive performance.

Keywords: chest, COVID-19, chest Image, coronavirus, CT image, chest CT

Procedia PDF Downloads 57
4825 The Driving Force for Taiwan Social Innovation Business Model Transformation: A Case Study of Social Innovation Internet Celebrity Training Project

Authors: Shih-Jie Ma, Jui-Hsu Hsiao, Ming-Ying Hsieh, Shin-Yan Yang, Chun-Han Yeh, Kuo-Chun Su

Abstract:

In Taiwan, social enterprises and non-profit organizations (NPOs) are not familiar with innovative business models, such as live streaming. In 2019, a brand new course called internet celebrity training project is introduced to them by the Social Innovation Lab. The Goal of this paper is to evaluate the effect of this project, to explore the role of new technology (internet live stream) in business process management (BPM), and to analyze how live stream programs can assist social enterprises in creating new business models. Social Innovation, with the purpose to solve social issues in innovative ways, is one of the most popular topics in the world. Social Innovation Lab was established in 2017 by Executive Yuan in Taiwan. The vision of Social Innovation Lab is to exploit technology, innovation and experimental methods to solve social issues, and to maximize the benefits from government investment. Social Innovation Lab aims at creating a platform for both supply and demand sides of social issues, to make social enterprises and start-ups communicate with each other, and to build an eco-system in which stakeholders can make a social impact. Social Innovation Lab keeps helping social enterprises and NPOs to gain better publicity and to enhance competitiveness by facilitating digital transformation. In this project, Social Innovation Lab exerted the influence of social media such as YouTube and Facebook, to make social enterprises and start-ups adjust their business models by using the live stream of social media, which becomes one of the tools to expand their market and diversify their sales channels. Internet live stream training courses were delivered in different regions of Taiwan in 2019, including Taitung, Taichung, Kaohsiung and Hualien. Through these courses, potential groups and enterprises were cultivated to become so-called internet celebrities. With their concern about social issues in mind, these internet celebrities know how to manipulate social media to make a social impact in different fields, such as aboriginal people, food and agriculture, LOHAS (Lifestyles of Health and Sustainability), environmental protection and senior citizens. Participants of live stream training courses in Taiwan are selected to take in-depth interviews and questionnaire surveys. Results indicate that the digital transformation process of social enterprises and NPOs can be successful by implementing business process reengineering, a significant change made by social innovation internet celebrities. Therefore, this project can be the new driving force to facilitate the business model transformation in Taiwan.

Keywords: business process management, digital transformation, live stream, social innovation

Procedia PDF Downloads 146
4824 Segmenting 3D Optical Coherence Tomography Images Using a Kalman Filter

Authors: Deniz Guven, Wil Ward, Jinming Duan, Li Bai

Abstract:

Over the past two decades or so, Optical Coherence Tomography (OCT) has been used to diagnose retina and optic nerve diseases. The retinal nerve fibre layer, for example, is a powerful diagnostic marker for detecting and staging glaucoma. With the advances in optical imaging hardware, the adoption of OCT is now commonplace in clinics. More and more OCT images are being generated, and for these OCT images to have clinical applicability, accurate automated OCT image segmentation software is needed. Oct image segmentation is still an active research area, as OCT images are inherently noisy, with the multiplicative speckling noise. Simple edge detection algorithms are unsuitable for detecting retinal layer boundaries in OCT images. Intensity fluctuation, motion artefact, and the presence of blood vessels also decrease further OCT image quality. In this paper, we introduce a new method for segmenting three-dimensional (3D) OCT images. This involves the use of a Kalman filter, which is commonly used in computer vision for object tracking. The Kalman filter is applied to the 3D OCT image volume to track the retinal layer boundaries through the slices within the volume and thus segmenting the 3D image. Specifically, after some pre-processing of the OCT images, points on the retinal layer boundaries in the first image are identified, and curve fitting is applied to them such that the layer boundaries can be represented by the coefficients of the curve equations. These coefficients then form the state space for the Kalman Filter. The filter then produces an optimal estimate of the current state of the system by updating its previous state using the measurements available in the form of a feedback control loop. The results show that the algorithm can be used to segment the retinal layers in OCT images. One of the limitations of the current algorithm is that the curve representation of the retinal layer boundary does not work well when the layer boundary is split into two, e.g., at the optic nerve, the layer boundary split into two. This maybe resolved by using a different approach to representing the boundaries, such as b-splines or level sets. The use of a Kalman filter shows promise to developing accurate and effective 3D OCT segmentation methods.

Keywords: optical coherence tomography, image segmentation, Kalman filter, object tracking

Procedia PDF Downloads 482
4823 Intertidal Fixed Stake Net Trap (Hadrah) Fishery in Kuwait, Distribution, Catch Rate, and Species Composition

Authors: Ali F. Al-Baz, Mohsen M. Al-Husaini, James M. Bishop

Abstract:

Intertidal fixed stake net trap (hadrah) is one of the oldest fishing gears used throughout the Arabian Gulf countries since 1800s and also one of most efficient methods of capturing fish from the intertidal area. This study described the hadrah fishery in Kuwait. From October 2001 to December 2002, more than 37,372 specimens representing 95 species (89 fish, 2 mollusks, 4 crustaceans) were measured from hadrah located in three different areas along Kuwait's coast. In Kuwait Bay, catch rates averaged 62 kg/sir day (range 14 kg/sir-day in February to 160 kg/sir-day in October 2002). Commercial species accounted for 41% of the catches. Catches from Failakah Island averaged 96 kg/sir-day from June through September, with 61% of the catch being commercial species. In the southern area, catches averaged only 32 kg/sir-day, and only 34% were commercially important. Forty percent of the hadrah catches were juveniles which shows that the shallow intertidal waters are prime nursery habitat, particularly in Kuwait Bay. To maintain ecosystem biodiversity and recruitment success of the fishes, we recommended that all hadrah should be removed from Kuwait Bay. In the future, removal of hadrah in other locations should be considered.

Keywords: catch and effort, hadrah, intertidal fixed stake net, Kuwait, species composition

Procedia PDF Downloads 501