Search results for: image guided radiation therapy
1639 Effects of Intrauterine and Extrauterine Exposure to 1800 MHz GSM-Like Radiofrequency Radiation on Liver Regulatory Enzymes Activities in Infant Female Rabbits
Authors: A. Tomruk, G. Guler, B. Tandogan, E. Ozgur, N.N. Ulusu, N. Seyhan
Abstract:
In the present study, we aimed to design the intrauterine and extrauterine exposure to 1800 MHz GSM-like RF radiation and investigate its possible bio-effects on infant female rabbits. Totally thirty-six New Zealand White female rabbits, onemonth old, were randomly divided into four groups which are composed of 9 rabbits; i. Group I [Intrauterine (IU) exposure(-); Extrauterine (EU) exposure (-)], Group II [IU exposure (-); EU exposure (+)], Group III [IU exposure(+);EU exposure(-)], Group IV [IU exposure (+);EU exposure(+)]. The master regulatory enzymes activities of pentose phosphate pathway (glucose-6-phosphate dehydrogenase, G-6PD; 6-phosphogluconate dehydrogenase, 6- PGDH) and glutathione-dependent metabolism (glutathione peroxidase, GSH-Px; glutathione reductase, GR; glutathione Stransferase, GST, thioredoxin reductase, TRx) were analyzed in liver tissues of young female rabbits. Decreased G-6PD, 6-PGD, GSH-Px, GR activities were found in Group III compared to Group I (p<0.05, Mann Whitney). Increased GSH-px and TRx activities were found in Group IV compared to Group I (p<0.05, Mann Whitney). It can be concluded that the intrauterine and extrauterine exposure to GSMlike RF radiation may influence the liver regulatory enzymes activities.Keywords: Radiofrequency (RF), intrauterine (IU) andextrauterine (EU) exposure, infant female rabbits.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23191638 Enhance Performance of Secure Image Using Wavelet Compression
Authors: Goh Han Keat, Azman Samsudin Zurinahni Zainol
Abstract:
The increase popularity of multimedia application especially in image processing places a great demand on efficient data storage and transmission techniques. Network communication such as wireless network can easily be intercepted and cause of confidential information leaked. Unfortunately, conventional compression and encryption methods are too slow; it is impossible to carry out real time secure image processing. In this research, Embedded Zerotree Wavelet (EZW) encoder which specially designs for wavelet compression is examined. With this algorithm, three methods are proposed to reduce the processing time, space and security protection that will be secured enough to protect the data.
Keywords: Embedded Zerotree Wavelet (EZW), Imagecompression, Wavelet encoder, Entropy encoder, Encryption.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16711637 Biometric Steganography Using Variable Length Embedding
Authors: Souvik Bhattacharyya, Indradip Banerjee, Anumoy Chakraborty, Gautam Sanyal
Abstract:
Recent growth in digital multimedia technologies has presented a lot of facilities in information transmission, reproduction and manipulation. Therefore, the concept of information security is one of the superior articles in the present day situation. The biometric information security is one of the information security mechanisms. It has the advantages as well as disadvantages. The biometric system is at risk to a range of attacks. These attacks are anticipated to bypass the security system or to suspend the normal functioning. Various hazards have been discovered while using biometric system. Proper use of steganography greatly reduces the risks in biometric systems from the hackers. Steganography is one of the fashionable information hiding technique. The goal of steganography is to hide information inside a cover medium like text, image, audio, video etc. through which it is not possible to detect the existence of the secret information. Here in this paper a new security concept has been established by making the system more secure with the help of steganography along with biometric security. Here the biometric information has been embedded to a skin tone portion of an image with the help of proposed steganographic technique.
Keywords: Biometrics, Skin tone detection, Series, Polynomial, Cover Image, Stego Image.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 26701636 Grooved Linear Microstrip Patch Antenna Array
Authors: Ayesha Aslam, F A Bhatti
Abstract:
A simple impedance matching technique for inset feed grooved microstrip patch antenna based on the concept of coplanar waveguide feed line has been developed and investigated for a printed antenna at X-Band frequency of 10GHz. The proposed technique has been used in the design of Linear Grooved Microstrip patch antenna array. The characteristics of the antenna are determined in terms of Return loss, VSWR, gain, radiation pattern etc. The measured and simulated results presented are found to be in good agreement.Keywords: Gain, Microstrip patch, return loss, VSWR, Radiation pattern, CPW Feed, Inset feed.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28071635 Effectiveness of Contourlet vs Wavelet Transform on Medical Image Compression: a Comparative Study
Authors: Negar Riazifar, Mehran Yazdi
Abstract:
Discrete Wavelet Transform (DWT) has demonstrated far superior to previous Discrete Cosine Transform (DCT) and standard JPEG in natural as well as medical image compression. Due to its localization properties both in special and transform domain, the quantization error introduced in DWT does not propagate globally as in DCT. Moreover, DWT is a global approach that avoids block artifacts as in the JPEG. However, recent reports on natural image compression have shown the superior performance of contourlet transform, a new extension to the wavelet transform in two dimensions using nonseparable and directional filter banks, compared to DWT. It is mostly due to the optimality of contourlet in representing the edges when they are smooth curves. In this work, we investigate this fact for medical images, especially for CT images, which has not been reported yet. To do that, we propose a compression scheme in transform domain and compare the performance of both DWT and contourlet transform in PSNR for different compression ratios (CR) using this scheme. The results obtained using different type of computed tomography images show that the DWT has still good performance at lower CR but contourlet transform performs better at higher CR.Keywords: Computed Tomography (CT), DWT, Discrete Contourlet Transform, Image Compression.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27991634 Pre-Malignant Breast Lesions, Methods of Treatment and Outcome
Authors: Ahmed Mostafa, Mohamed Mahmoud, Nesreen H. Hafez, Mohamed Fahim
Abstract:
This retrospective study includes 60 patients with pre-invasive breast cancer. Aim of the study: Evaluation of premalignant lesions of the breast (DCIS), different treatment methods and outcome. Patients and methods: 60 patients with DCIS were studied from the period between 2005 to 2012, for 38 patients the primary surgical method was wide local resection (WLE) (63.3%) and the other cases (22 patients, 36.7%) had mastectomy, fourteen cases from those who underwent local excision received radiotherapy, while no adjuvant radiotherapy was given for those who underwent mastectomy. In case of hormonal receptor positive DCIS lesions hormonal treatment (Tamoxifen) was given after local control. Results: No difference in overall survival between mastectomy & breast conserving therapy (wide local excision and adjuvant radiotherapy), however local recurrence rate is higher in case of breast conserving therapy, also no role of Axillary evacuation in case of DCIS. The use of hormonal therapy decreases the incidence of local recurrence by about 98%. Conclusion: The main management of DCIS is local treatment (wide local excision and radiotherapy) with hormonal treatment in case of hormone receptor positive lesions.
Keywords: Ductal carcinoma in situ, surgical treatment, radiotherapy, breast conserving therapy, hormonal treatment.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14371633 Algorithm for Path Recognition in-between Tree Rows for Agricultural Wheeled-Mobile Robots
Authors: Anderson Rocha, Pedro Miguel de Figueiredo Dinis Oliveira Gaspar
Abstract:
Machine vision has been widely used in recent years in agriculture, as a tool to promote the automation of processes and increase the levels of productivity. The aim of this work is the development of a path recognition algorithm based on image processing to guide a terrestrial robot in-between tree rows. The proposed algorithm was developed using the software MATLAB, and it uses several image processing operations, such as threshold detection, morphological erosion, histogram equalization and the Hough transform, to find edge lines along tree rows on an image and to create a path to be followed by a mobile robot. To develop the algorithm, a set of images of different types of orchards was used, which made possible the construction of a method capable of identifying paths between trees of different heights and aspects. The algorithm was evaluated using several images with different characteristics of quality and the results showed that the proposed method can successfully detect a path in different types of environments.
Keywords: Agricultural mobile robot, image processing, path recognition, Hough transform.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17891632 Implementation of RC5 Block Cipher Algorithm for Image Cryptosystems
Authors: Hossam El-din H. Ahmed, Hamdy M. Kalash, Osama S. Farag Allah
Abstract:
This paper examines the implementation of RC5 block cipher for digital images along with its detailed security analysis. A complete specification for the method of application of the RC5 block cipher to digital images is given. The security analysis of RC5 block cipher for digital images against entropy attack, bruteforce, statistical, and differential attacks is explored from strict cryptographic viewpoint. Experiments and results verify and prove that RC5 block cipher is highly secure for real-time image encryption from cryptographic viewpoint. Thorough experimental tests are carried out with detailed analysis, demonstrating the high security of RC5 block cipher algorithm.
Keywords: Image encryption, security analysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 36761631 Image Segmentation Using the K-means Algorithm for Texture Features
Authors: Wan-Ting Lin, Chuen-Horng Lin, Tsung-Ho Wu, Yung-Kuan Chan
Abstract:
This study aims to segment objects using the K-means algorithm for texture features. Firstly, the algorithm transforms color images into gray images. This paper describes a novel technique for the extraction of texture features in an image. Then, in a group of similar features, objects and backgrounds are differentiated by using the K-means algorithm. Finally, this paper proposes a new object segmentation algorithm using the morphological technique. The experiments described include the segmentation of single and multiple objects featured in this paper. The region of an object can be accurately segmented out. The results can help to perform image retrieval and analyze features of an object, as are shown in this paper.Keywords: k-mean, multiple objects, segmentation, texturefeatures.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 28221630 A Comparison Study of the Removal of Selected Pharmaceuticals in Waters by Chemical Oxidation Treatments
Authors: F. Javier Benitez, Juan Luis Acero, Francisco J. Real, Gloria Roldan, Francisco Casas
Abstract:
The degradation of selected pharmaceuticals in some water matrices was studied by using several chemical treatments. The pharmaceuticals selected were the beta-blocker metoprolol, the nonsteroidal anti-inflammatory naproxen, the antibiotic amoxicillin, and the analgesic phenacetin; and their degradations were conducted by using UV radiation alone, ozone, Fenton-s reagent, Fenton-like system, photo-Fenton system, and combinations of UV radiation and ozone with H2O2, TiO2, Fe(II), and Fe(III). The water matrices, in addition to ultra-pure water, were a reservoir water, a groundwater, and two secondary effluents from two municipal WWTP. The results reveal that the presence of any second oxidant enhanced the oxidation rates, with the systems UV/TiO2 and O3/TiO2 providing the highest degradation rates. It is also observed in most of the investigated oxidation systems that the degradation rate followed the sequence: amoxicillin > naproxen > metoprolol > phenacetin. Lower rates were obtained with the pharmaceuticals dissolved in natural waters and secondary effluents due to the organic matter present which consume some amounts of the oxidant agents.Keywords: Pharmaceuticals, UV radiation, ozone, advancedoxidation processes, water matrices, degradation rates
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22371629 Adaptive Block State Update Method for Separating Background
Authors: Youngsuck Ji, Youngjoon Han, Hernsoo Hahn
Abstract:
In this paper, we proposed the robust mobile object detection method for light effect in the night street image block based updating reference background model using block state analysis. Experiment image is acquired sequence color video from steady camera. When suddenly appeared artificial illumination, reference background model update this information such as street light, sign light. Generally natural illumination is change by temporal, but artificial illumination is suddenly appearance. So in this paper for exactly detect artificial illumination have 2 state process. First process is compare difference between current image and reference background by block based, it can know changed blocks. Second process is difference between current image-s edge map and reference background image-s edge map, it possible to estimate illumination at any block. This information is possible to exactly detect object, artificial illumination and it was generating reference background more clearly. Block is classified by block-state analysis. Block-state has a 4 state (i.e. transient, stationary, background, artificial illumination). Fig. 1 is show characteristic of block-state respectively [1]. Experimental results show that the presented approach works well in the presence of illumination variance.Keywords: Block-state, Edge component, Reference backgroundi, Artificial illumination.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13211628 Mining Image Features in an Automatic Two-Dimensional Shape Recognition System
Authors: R. A. Salam, M.A. Rodrigues
Abstract:
The number of features required to represent an image can be very huge. Using all available features to recognize objects can suffer from curse dimensionality. Feature selection and extraction is the pre-processing step of image mining. Main issues in analyzing images is the effective identification of features and another one is extracting them. The mining problem that has been focused is the grouping of features for different shapes. Experiments have been conducted by using shape outline as the features. Shape outline readings are put through normalization and dimensionality reduction process using an eigenvector based method to produce a new set of readings. After this pre-processing step data will be grouped through their shapes. Through statistical analysis, these readings together with peak measures a robust classification and recognition process is achieved. Tests showed that the suggested methods are able to automatically recognize objects through their shapes. Finally, experiments also demonstrate the system invariance to rotation, translation, scale, reflection and to a small degree of distortion.Keywords: Image mining, feature selection, shape recognition, peak measures.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14581627 Region Segmentation based on Gaussian Dirichlet Process Mixture Model and its Application to 3D Geometric Stricture Detection
Authors: Jonghyun Park, Soonyoung Park, Sanggyun Kim, Wanhyun Cho, Sunworl Kim
Abstract:
In general, image-based 3D scenes can now be found in many popular vision systems, computer games and virtual reality tours. So, It is important to segment ROI (region of interest) from input scenes as a preprocessing step for geometric stricture detection in 3D scene. In this paper, we propose a method for segmenting ROI based on tensor voting and Dirichlet process mixture model. In particular, to estimate geometric structure information for 3D scene from a single outdoor image, we apply the tensor voting and Dirichlet process mixture model to a image segmentation. The tensor voting is used based on the fact that homogeneous region in an image are usually close together on a smooth region and therefore the tokens corresponding to centers of these regions have high saliency values. The proposed approach is a novel nonparametric Bayesian segmentation method using Gaussian Dirichlet process mixture model to automatically segment various natural scenes. Finally, our method can label regions of the input image into coarse categories: “ground", “sky", and “vertical" for 3D application. The experimental results show that our method successfully segments coarse regions in many complex natural scene images for 3D.
Keywords: Region segmentation, tensor voting, image-based 3D, geometric structure, Gaussian Dirichlet process mixture model
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18911626 A New High Speed Neural Model for Fast Character Recognition Using Cross Correlation and Matrix Decomposition
Authors: Hazem M. El-Bakry
Abstract:
Neural processors have shown good results for detecting a certain character in a given input matrix. In this paper, a new idead to speed up the operation of neural processors for character detection is presented. Such processors are designed based on cross correlation in the frequency domain between the input matrix and the weights of neural networks. This approach is developed to reduce the computation steps required by these faster neural networks for the searching process. The principle of divide and conquer strategy is applied through image decomposition. Each image is divided into small in size sub-images and then each one is tested separately by using a single faster neural processor. Furthermore, faster character detection is obtained by using parallel processing techniques to test the resulting sub-images at the same time using the same number of faster neural networks. In contrast to using only faster neural processors, the speed up ratio is increased with the size of the input image when using faster neural processors and image decomposition. Moreover, the problem of local subimage normalization in the frequency domain is solved. The effect of image normalization on the speed up ratio of character detection is discussed. Simulation results show that local subimage normalization through weight normalization is faster than subimage normalization in the spatial domain. The overall speed up ratio of the detection process is increased as the normalization of weights is done off line.Keywords: Fast Character Detection, Neural Processors, Cross Correlation, Image Normalization, Parallel Processing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15371625 Computer Countenanced Diagnosis of Skin Nodule Detection and Histogram Augmentation: Extracting System for Skin Cancer
Authors: S. Zith Dey Babu, S. Kour, S. Verma, C. Verma, V. Pathania, A. Agrawal, V. Chaudhary, A. Manoj Puthur, R. Goyal, A. Pal, T. Danti Dey, A. Kumar, K. Wadhwa, O. Ved
Abstract:
Background: Skin cancer is now is the buzzing button in the field of medical science. The cyst's pandemic is drastically calibrating the body and well-being of the global village. Methods: The extracted image of the skin tumor cannot be used in one way for diagnosis. The stored image contains anarchies like the center. This approach will locate the forepart of an extracted appearance of skin. Partitioning image models has been presented to sort out the disturbance in the picture. Results: After completing partitioning, feature extraction has been formed by using genetic algorithm and finally, classification can be performed between the trained and test data to evaluate a large scale of an image that helps the doctors for the right prediction. To bring the improvisation of the existing system, we have set our objectives with an analysis. The efficiency of the natural selection process and the enriching histogram is essential in that respect. To reduce the false-positive rate or output, GA is performed with its accuracy. Conclusions: The objective of this task is to bring improvisation of effectiveness. GA is accomplishing its task with perfection to bring down the invalid-positive rate or outcome. The paper's mergeable portion conflicts with the composition of deep learning and medical image processing, which provides superior accuracy. Proportional types of handling create the reusability without any errors.
Keywords: Computer-aided system, detection, image segmentation, morphology.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5441624 Low Computational Image Compression Scheme based on Absolute Moment Block Truncation Coding
Authors: K.Somasundaram, I.Kaspar Raj
Abstract:
In this paper we have proposed three and two stage still gray scale image compressor based on BTC. In our schemes, we have employed a combination of four techniques to reduce the bit rate. They are quad tree segmentation, bit plane omission, bit plane coding using 32 visual patterns and interpolative bit plane coding. The experimental results show that the proposed schemes achieve an average bit rate of 0.46 bits per pixel (bpp) for standard gray scale images with an average PSNR value of 30.25, which is better than the results from the exiting similar methods based on BTC.Keywords: Bit plane, Block Truncation Coding, Image compression, lossy compression, quad tree segmentation
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17521623 FPGA Implementation of a Vision-Based Blind Spot Warning System
Authors: Yu Ren Lin, Yu Hong Li
Abstract:
Vision-based intelligent vehicle applications often require large amounts of memory to handle video streaming and image processing, which in turn increases complexity of hardware and software. This paper presents an FPGA implement of a vision-based blind spot warning system. Using video frames, the information of the blind spot area turns into one-dimensional information. Analysis of the estimated entropy of image allows the detection of an object in time. This idea has been implemented in the XtremeDSP video starter kit. The blind spot warning system uses only 13% of its logic resources and 95k bits block memory, and its frame rate is over 30 frames per sec (fps).
Keywords: blind-spot area, image, FPGA
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18361622 Survey Gamma Radiation Measurements in Commercially-used Natural Tiling Rocks in Iran
Authors: A.Abbasi, F.Mirekhtiary
Abstract:
The gamma radiation in samples of a variety of natural tiling rocks (granites) produced and imported in Iran use in the building industry was measured, employing high-resolution Gamma-ray spectroscopy. The rock samples were pulverized, sealed in 0.5 liter plastic Marinelli beakers, and measured in the laboratory with an accumulating time between 50000 and 80000 second each. From the measured Gamma-ray spectra, activity concentrations were determined for 232Th (range from 6.5 to 172.2 Bq kg-1), 238U (from 7.5 to 178.1 Bq kg-1 ),226Ra( from 3.8 to 94.2 Bq kg-1 ) 40K (from 556.9 to 1539.2 Bq kg-1). From the 29 samples measured in this study, “Nehbndan ( Berjand )" appears to present the highest concentrations for 232Th,“Big Red Flower (China) "for 238U , “ Khoram dareh" for 226 Ra and “ Peranshahr" for 40K , respectively.Keywords: activity concentration, natural radioactivity, tilingrocks (granites)
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14091621 Experimental and Theoretical Investigation of Rough Rice Drying in Infrared-assisted Hot Air Dryer Using Artificial Neural Network
Authors: D. Zare, H. Naderi, A. A. Jafari
Abstract:
Drying characteristics of rough rice (variety of lenjan) with an initial moisture content of 25% dry basis (db) was studied in a hot air dryer assisted by infrared heating. Three arrival air temperatures (30, 40 and 500C) and four infrared radiation intensities (0, 0.2 , 0.4 and 0.6 W/cm2) and three arrival air speeds (0.1, 0.15 and 0.2 m.s-1) were studied. Bending strength of brown rice kernel, percentage of cracked kernels and time of drying were measured and evaluated. The results showed that increasing the drying arrival air temperature and radiation intensity of infrared resulted decrease in drying time. High bending strength and low percentage of cracked kernel was obtained when paddy was dried by hot air assisted infrared dryer. Between this factors and their interactive effect were a significant difference (p<0.01). An intensity level of 0.2 W/cm2 was found to be optimum for radiation drying. Furthermore, in the present study, the application of Artificial Neural Network (ANN) for predicting the moisture content during drying (output parameter for ANN modeling) was investigated. Infrared Radiation intensity, drying air temperature, arrival air speed and drying time were considered as input parameters for the model. An ANN model with two hidden layers with 8 and 14 neurons were selected for studying the influence of transfer functions and training algorithms. The results revealed that a network with the Tansig (hyperbolic tangent sigmoid) transfer function and trainlm (Levenberg-Marquardt) back propagation algorithm made the most accurate predictions for the paddy drying system. Mean square error (MSE) was calculated and found that the random errors were within and acceptable range of ±5% with coefficient of determination (R2) of 99%.
Keywords: Rough rice, Infrared-hot air, Artificial Neural Network
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18261620 Motion Detection Techniques Using Optical Flow
Authors: A. A. Shafie, Fadhlan Hafiz, M. H. Ali
Abstract:
Motion detection is very important in image processing. One way of detecting motion is using optical flow. Optical flow cannot be computed locally, since only one independent measurement is available from the image sequence at a point, while the flow velocity has two components. A second constraint is needed. The method used for finding the optical flow in this project is assuming that the apparent velocity of the brightness pattern varies smoothly almost everywhere in the image. This technique is later used in developing software for motion detection which has the capability to carry out four types of motion detection. The motion detection software presented in this project also can highlight motion region, count motion level as well as counting object numbers. Many objects such as vehicles and human from video streams can be recognized by applying optical flow technique.Keywords: Background modeling, Motion detection, Optical flow, Velocity smoothness constant, motion trajectories.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 53861619 Fragile Watermarking for Color Images Using Thresholding Technique
Authors: Kuo-Cheng Liu
Abstract:
In this paper, we propose ablock-wise watermarking scheme for color image authentication to resist malicious tampering of digital media. The thresholding technique is incorporated into the scheme such that the tampered region of the color image can be recovered with high quality while the proofing result is obtained. The watermark for each block consists of its dual authentication data and the corresponding feature information. The feature information for recovery iscomputed bythe thresholding technique. In the proofing process, we propose a dual-option parity check method to proof the validity of image blocks. In the recovery process, the feature information of each block embedded into the color image is rebuilt for high quality recovery. The simulation results show that the proposed watermarking scheme can effectively proof the tempered region with high detection rate and can recover the tempered region with high quality.
Keywords: thresholding technique, tamper proofing, tamper recovery
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16321618 Laser Excited Nuclear γ -Source of High Spectral Brightness
Authors: A. Аndreev, Yu. Rozhdestvenskii, К. Platonov, R. Salomaa
Abstract:
This paper considers various channels of gammaquantum generation via an ultra-short high-power laser pulse interaction with different targets.We analyse the possibilities to create a pulsed gamma-radiation source using laser triggering of some nuclear reactions and isomer targets. It is shown that sub-MeV monochromatic short pulse of gamma-radiation can be obtained with pulse energy of sub-mJ level from isomer target irradiated by intense laser pulse. For nuclear reaction channel in light- atom materials, it is shown that sub-PW laser pulse gives rise to formation about million gamma-photons of multi-MeV energy.Keywords: High power laser, short pulse, fast particles, isomertarget.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18261617 Medical Image Segmentation and Detection of MR Images Based on Spatial Multiple-Kernel Fuzzy C-Means Algorithm
Authors: J. Mehena, M. C. Adhikary
Abstract:
In this paper, a spatial multiple-kernel fuzzy C-means (SMKFCM) algorithm is introduced for segmentation problem. A linear combination of multiples kernels with spatial information is used in the kernel FCM (KFCM) and the updating rules for the linear coefficients of the composite kernels are derived as well. Fuzzy cmeans (FCM) based techniques have been widely used in medical image segmentation problem due to their simplicity and fast convergence. The proposed SMKFCM algorithm provides us a new flexible vehicle to fuse different pixel information in medical image segmentation and detection of MR images. To evaluate the robustness of the proposed segmentation algorithm in noisy environment, we add noise in medical brain tumor MR images and calculated the success rate and segmentation accuracy. From the experimental results it is clear that the proposed algorithm has better performance than those of other FCM based techniques for noisy medical MR images.Keywords: Clustering, fuzzy C-means, image segmentation, MR images, multiple kernels.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21291616 Dual Pyramid of Agents for Image Segmentation
Authors: K. Idir, H. Merouani, Y. Tlili.
Abstract:
An effective method for the early detection of breast cancer is the mammographic screening. One of the most important signs of early breast cancer is the presence of microcalcifications. For the detection of microcalcification in a mammography image, we propose to conceive a multiagent system based on a dual irregular pyramid. An initial segmentation is obtained by an incremental approach; the result represents level zero of the pyramid. The edge information obtained by application of the Canny filter is taken into account to affine the segmentation. The edge-agents and region-agents cooper level by level of the pyramid by exploiting its various characteristics to provide the segmentation process convergence.Keywords: Dual Pyramid, Image Segmentation, Multi-agent System, Region/Edge Cooperation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19191615 The Water Level Detection Algorithm Using the Accumulated Histogram with Band Pass Filter
Authors: Sangbum Park, Namki Lee, Youngjoon Han, Hernsoo Hahn
Abstract:
In this paper, we propose the robust water level detection method based on the accumulated histogram under small changed image which is acquired from water level surveillance camera. In general surveillance system, this is detecting and recognizing invasion from searching area which is in big change on the sequential images. However, in case of a water level detection system, these general surveillance techniques are not suitable due to small change on the water surface. Therefore the algorithm introduces the accumulated histogram which is emphasizing change of water surface in sequential images. Accumulated histogram is based on the current image frame. The histogram is cumulating differences between previous images and current image. But, these differences are also appeared in the land region. The band pass filter is able to remove noises in the accumulated histogram Finally, this algorithm clearly separates water and land regions. After these works, the algorithm converts from the water level value on the image space to the real water level on the real space using calibration table. The detected water level is sent to the host computer with current image. To evaluate the proposed algorithm, we use test images from various situations.Keywords: accumulated histogram, water level detection, band pass filter.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 20001614 Deficiencies of Lung Segmentation Techniques using CT Scan Images for CAD
Authors: Nisar Ahmed Memon, Anwar Majid Mirza, S.A.M. Gilani
Abstract:
Segmentation is an important step in medical image analysis and classification for radiological evaluation or computer aided diagnosis. This paper presents the problem of inaccurate lung segmentation as observed in algorithms presented by researchers working in the area of medical image analysis. The different lung segmentation techniques have been tested using the dataset of 19 patients consisting of a total of 917 images. We obtained datasets of 11 patients from Ackron University, USA and of 8 patients from AGA Khan Medical University, Pakistan. After testing the algorithms against datasets, the deficiencies of each algorithm have been highlighted.Keywords: Computer Aided Diagnosis (CAD), MathematicalMorphology, Medical Image Analysis, Region Growing, Segmentation, Thresholding,
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23401613 The Use of Complex Contourlet Transform on Fusion Scheme
Authors: Dipeng Chen, Qi Li
Abstract:
Image fusion aims to enhance the perception of a scene by combining important information captured by different sensors. Dual-Tree Complex Wavelet (DT-CWT) has been thouroughly investigated for image fusion, since it takes advantages of approximate shift invariance and direction selectivity. But it can only handle limited direction information. To allow a more flexible directional expansion for images, we propose a novel fusion scheme, referred to as complex contourlet transform (CCT). It successfully incorporates directional filter banks (DFB) into DT-CWT. As a result it efficiently deal with images containing contours and textures, whereas it retains the property of shift invariance. Experimental results demonstrated that the method features high quality fusion performance and can facilitate many image processing applications.Keywords: Complex contourlet transform, Complex wavelettransform, Fusion.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15961612 A Compact Wearable Slot Antenna for LTE and WLAN Applications
Authors: Haider K. Raad
Abstract:
In this paper, a compact wide-band, ultra-thin and flexible slot antenna intended for wearable applications is presented. The presented antenna is designed to provide Wireless Local Area Network (WLAN) and Long Term Evolution (LTE) connectivity. The presented design exhibits a relatively wide bandwidth (1600-3500 MHz below -6 dB impedance bandwidth limit). The antenna is positioned on a 33 mm x 30 mm flexible substrate with a thickness of 50 µm. Antenna properties, such as the far-field radiation patterns, scattering parameter S11 are provided. The presented compact, thin and flexible design along with excellent radiation characteristics are deemed suitable for integration into flexible and wearable devices.Keywords: Wearable Electronics, Slot Antenna, LTE, WLAN.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11941611 Noise Reduction in Image Sequences using an Effective Fuzzy Algorithm
Authors: Mahmoud Saeidi, Khadijeh Saeidi, Mahmoud Khaleghi
Abstract:
In this paper, we propose a novel spatiotemporal fuzzy based algorithm for noise filtering of image sequences. Our proposed algorithm uses adaptive weights based on a triangular membership functions. In this algorithm median filter is used to suppress noise. Experimental results show when the images are corrupted by highdensity Salt and Pepper noise, our fuzzy based algorithm for noise filtering of image sequences, are much more effective in suppressing noise and preserving edges than the previously reported algorithms such as [1-7]. Indeed, assigned weights to noisy pixels are very adaptive so that they well make use of correlation of pixels. On the other hand, the motion estimation methods are erroneous and in highdensity noise they may degrade the filter performance. Therefore, our proposed fuzzy algorithm doesn-t need any estimation of motion trajectory. The proposed algorithm admissibly removes noise without having any knowledge of Salt and Pepper noise density.Keywords: Image Sequences, Noise Reduction, fuzzy algorithm, triangular membership function
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18811610 Near-Lossless Image Coding based on Orthogonal Polynomials
Authors: Krishnamoorthy R, Rajavijayalakshmi K, Punidha R
Abstract:
In this paper, a near lossless image coding scheme based on Orthogonal Polynomials Transform (OPT) has been presented. The polynomial operators and polynomials basis operators are obtained from set of orthogonal polynomials functions for the proposed transform coding. The image is partitioned into a number of distinct square blocks and the proposed transform coding is applied to each of these individually. After applying the proposed transform coding, the transformed coefficients are rearranged into a sub-band structure. The Embedded Zerotree (EZ) coding algorithm is then employed to quantize the coefficients. The proposed transform is implemented for various block sizes and the performance is compared with existing Discrete Cosine Transform (DCT) transform coding scheme.Keywords: Near-lossless Coding, Orthogonal Polynomials Transform, Embedded Zerotree Coding
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1946