Region-Based Image Fusion with Artificial Neural Network
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 32797
Region-Based Image Fusion with Artificial Neural Network

Authors: Shuo-Li Hsu, Peng-Wei Gau, I-Lin Wu, Jyh-Horng Jeng

Abstract:

For most image fusion algorithms separate relationship by pixels in the image and treat them more or less independently. In addition, they have to be adjusted different parameters in different time or weather. In this paper, we propose a region–based image fusion which combines aspects of feature and pixel-level fusion method to replace only by pixel. The basic idea is to segment far infrared image only and to add information of each region from segmented image to visual image respectively. Then we determine different fused parameters according different region. At last, we adopt artificial neural network to deal with the problems of different time or weather, because the relationship between fused parameters and image features are nonlinear. It render the fused parameters can be produce automatically according different states. The experimental results present the method we proposed indeed have good adaptive capacity with automatic determined fused parameters. And the architecture can be used for lots of applications.

Keywords: Image fusion, Region-based fusion, Segmentation, Neural network, Multi-sensor.

Digital Object Identifier (DOI): doi.org/10.5281/zenodo.1334015

Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2194

References:


[1] Z. Wang, D. Ziou, C. Armenakis, D. Li, and Q. Li, "A Comparative Analysis of Image Fusion Methods," Geoscience and Remote Sensing, vol. 43, no. 6, pp. 1391-1402, June 2006.
[2] J. G. Liu, "Smoothing filter-based intensity modulation: A spectral preserve image fusion technique for improving spatial details," Int. J. Remote Sensing, vol. 21, no. 18, pp. 3461-3472, 2000.
[3] M. Li, W. Cai, and Z. Tan, "A region-based multi-sensor image fusion scheme using pulse-coupled neural network," Pattern Recognition Letters, vol. 27, pp. 1948-1956, 2006.
[4] L. J. Guo and J. M. Moore, "Pixel block intensity modulation: adding spatial detail to TM band 6 thermal imagery," Int. J. Remote Sensing., vol. 19, no. 13, pp. 2477-2491, 1988.
[5] P. S. Chavez and J. A. Bowell, "Comparison of the spectral information content of Landsat thematic mapper and SPOT for three different sites in the Phoenix, Arizona region," Photogramm. Eng. Remote Sensing., vol. 54, no.12, pp. 1699-1708, 1988.
[6] A. R. Gillespie, A. B. Kahle, and R. E. Walker, "Color enhancement of highly Correlated images- . Channel ratio and chromaticity transformation Techniques," Remote Sensing Environment, vol. 22, pp. 343-365, 1987.
[7] J. Sun, J. Li and J. Li, "Multi-source remote sensing image fusion," INT. J. Remote Sensing, vol. 2, no. 1, pp. 323-328, Feb. 1998.
[8] W. J. Carper, T. M. Lillesand, and R. W. Kiefer, "The use of Intensity- Hue-Saturation transformation for merging SPOT panchromatic and multispectral image data," Photogramm. Eng. Remote Sensing, vol. 56, no. 4, pp. 459-467, 1990.
[9] K. Edwards and P. A. Davis, "The use of Intensity-Hue-Saturation transformation for producing color shaded-relief images," Photogramm. Eng. Remote Sensing, vol. 60, no. 11, pp. 1369-1374, 1994.
[10] E. M. Schetselaar, "Fusion by the IHS transform: Should we use cylindrical or Spherical coordinates?," Int. J. Remote Sensing, vol. 19, no. 4, pp. 759-765, 1998.
[11] J. Zhou, D. L. Civco, and J. A. Silander, "A wavelet transform method to merge Landsat TM and SPOT panchromatic data," Int. J. Remote Sensing, vol. 19, no. 4, pp. 743-757, 1998.
[12] S. Li, J. T. Kwok, Y. Wang, "Multifocus image fusion using artificial neural networks," Pattern Recognition Letters, vol. 23, pp. 985-997, 2002.
[13] Q. Yuan, C.Y. Dong, Q. Wang, "An adaptive fusion algorithm based on ANFIS for radar/infrared system," Expert Systems with Applications, vol. 36, pp. 111-120, 2009.