International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016), pp. 31-44 http://dx.doi.org/10.14257/ijseia.2016.10.2.04 Visible and Infrared Data Fusion Gwanggil Jeon Department of Embedded Systems Engineering, Incheon National University 119 Academy-ro, Yeonsu-gu, Incheon 406-772, Korea [email protected] Abstract This paper shows data fusion method with visible light image and infrared image. We assume an image is captured by FLIR image, which provides two images: visible light and infrared images. The details of the image are obtained in visible light image. This information is added to original low resolution infrared image and we assume this image as the detail strengthen infrared image. Experimental results section provides PSNR and MSE results. It is obvious that that proposed method with p=0.03 gives visually satisfactory results. Keywords: color image, membership function, quality enhancement, fuzzy model 1. Introduction An optical image is captured by sensor which is transformed into a digital signal [1]. Thus, the camera is an optical instrument for remembering data, which could be saved in a particular place, transmitted and transformed to another place [2, 3]. Recent photographic camera developed from the camera obscura [4]. The functioning of the camera is very close to the operating of the human eye. There are two cameras: visible light camera and infrared camera [5]. The visible light camera is ordinary camera, while infrared camera captures signal acts like thermometer [6]. The infrared radiation was found in early 19th century and learned a type of invisible radiation in the spectrum lower in energy than red signal, via its accomplish upon a thermometer. For visible camera, it can work with the light of the visible spectrum or with other locations of the electromagnetic spectrum [7, 8]. On the other hand, infrared camera captures invisible radiant energy, electromagnetic radiation with longer wavelengths than that of visible light. In general, most of the thermal radiation sent out by objects near room temperature is infrared [9-11]. Generally speaking, infrared image has lower resolution while visible image has higher resolution. Although infrared image can be enlarged by super resolution technique, most important details are still missing. To achieve better performance, in this paper, we use higher resolution visible image. We obtain details from visible image, and then this information is added to upsampled infrared image. We first transform color image into YCbCr. Details are obtained from luminance (Y) information. This paper is arranged as follows. Section 2 describes the proposed method. Section 3 describes simulation results where objective and subjective performance are provided. Conclusion remarks are provided in Section 4. 2. Proposed Method Figure 1 shows an example of infrared images with different colormap. Four options were used gray, hot, cold, and copper. The FLIR infrared camera has two outputs: one is visible light image and the other is infrared image. Normally, visible light image has ISSN: 1738-9984 IJSEIA Copyright ⓒ 2016 SERSC International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) higher resolution while infrared image has lower resolution. Therefore, object details can be found from visible light image and which can be added to infrared image. (a) (b) (c) (d) Figure 1. Examples of Infrared Images with Various Colormap: (a) Gray, (b) Hot, (c) Cold, and (d) Copper Figure 2 shows the flowchart of the proposed method. First of all, object is captured by FLIR camera, which provides two images, infrared and visible light images. As visible light image has higher resolution, detail of the images is obtained by using high pass filter. On the other hand, the infrared image has lower resolution, it is required to upsample the image to meet the size of visible light image. After upsampling it with super resolution method, details from the high pass filter are added to provide visually satisfactory image. There are few methods that we consider for super resolution method: nearest neighbor (hNN), bilinear (hBI), and bicubic (hBC) as shown in Eqs. (1-3). Note that, we assume upsampling rate is 2. 32 Copyright ⓒ 2016 SERSC International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) Object High resolution Low resolution Visible camera Infrared camera High pass filter Detail achievement with factor of p Image upsampling to meet visible camera image size Image fusion p High resolution infrared image Figure 2. Flowchart of the Proposed Method 1 hNN 1 0 1 2 1 hBI hBC 1 0 1 0 , 0 0 2 1 4 2 2 1 (1) (2) , 4 1 4 6 4 1 4 16 24 16 4 6 24 36 24 6 4 16 24 16 4 1 4 6 4 1 . 64 (3) To apply Eqs. (1-3), we first create zero-interleaved version of the matrix. Then the matrix becomes double the size of the original and will include mainly zeros. We now substitute the zeros by applying given spatial filters to this matrix. Copyright ⓒ 2016 SERSC 33 International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) (a) (b) (c) (d) (e) (f) Figure 3. Edge Results to be Added in Low Resolution Image: (a) p=0.03, (b) p=0.05, (c) p=0.07, (d) p=0.10, (e) p=0.20, and (f) p=0.30. To achieve high pass filter results, we generated new 3-by-3 matrix hHPF, which is obtained by Eq. (4). hHPF a p d g b e h c f . i (3) where a+b+c+d+f+g+h+i=-e, and a, b, c, d, f, g, h, i are negative numbers and e is positive number. We assume b, d, f, and h have more effect than a, c, g, i. Parameter p is high pass filter strength parameter, and in this paper we assume it p=0, p=0.03, p=0.05, p=0.07, p=0.10, p=0.20, and p=0.30, as shown in Figure 3. 34 Copyright ⓒ 2016 SERSC International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) It can be found from Fig. 4 that bigger p results strong edges, small p provides weak edges. By adding Figure 3(b) to original infrared image Figure 4(b), we obtain edge strengthened image, Figure 4(c). (a) (b) (c) Figure 4. (a) Original Image, (b) Zoomed Result of Infrared Image, (c) Proposed Method Applied Image (p=0.05) The output image ImIRout is calculated as Eq. (X). ImV represents original visible light image and hHPF*ImV shows convoluted results with hHPF. Imout IR Im IR hHPF *ImV . (4) 3. Experimental Results The proposed method was simulated under MATLAB 7.0 environment. Test images are 25 Zahra dataset, which has 768-by-512 size. Infrared image has one channel, while color image has 3 channels. Visual performance was compared and presented. First of all 768-by-512 size infrared image is downsampled and upsampled using bicubic method. Decimation factor was 2, therefore downsampled image size was 384by-256. High pass filter parameters were a=-2, b=-3, c=-2, d=-3, f=-3, g=-2, h=-3, i=-2, and therefore e=20. Copyright ⓒ 2016 SERSC 35 International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) (a) (b) (c) Figure 5. Three Original Images in Zahra Set: (a) #18, (b) #19, and (c) #20 Figure 5 shows three original test images: #18, #19, and #20, which are infrared images. The subjective performance comparisons on Figure 5 are shown in Figures 6-8. Each figure has six images: low resolution image (p=0), edge enhanced images with parameter values p=0.03, p=0.05, p=0.07, p=0.10, and p=0.20. It can be found in all (a) images that details in these images were missing. As p increases, details of images are getting distinct. However, big p does not always give the best performance. As can be seen in (f) image, some pixels were over-sharpened. (a) 36 (b) Copyright ⓒ 2016 SERSC International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) (c) (d) (e) (f) Figure 6. Subjective Performance Comparison on #18 Image: (a) Low Resolution Image, (b) p=0.03, (c) p=0.05, (d) p=0.07, (e) p=0.10, and (f) p=0.20 (a) Copyright ⓒ 2016 SERSC (b) 37 International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) (c) (d) (e) (f) Figure 7. Subjective Performance Comparison on #19 image: (a) Low Resolution Ismage, (b) p=0.03, (c) p=0.05, (d) p=0.07, (e) p=0.10, and (f) p=0.20 (a) 38 (b) Copyright ⓒ 2016 SERSC International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) (c) (d) (e) (f) Figure 8. Subjective Performance Comparison on #24 image: (a) Low Resolution Image, (b) p=0.03, (c) p=0.05, (d) p=0.07, (e) p=0.10, and (f) p=0.20 Two objective performances metrics were used in this paper. One is PSNR and the other is SME. Table 1 shows PSNR results for 25 Zahra dataset. The last row shows average results of each parameter p. It can be found from Tables 1 and 2 that p=0.03 provides the best performance. The second best p was p=0.05. Table 1. PSNR Results (in dB) for 25 Zahra Images 1 2 3 4 5 6 7 8 9 10 11 12 p=0 p=0.03 p=0.05 p=0.07 p=0.10 p=0.20 p=0.30 32.929 35.524 33.412 30.487 26.896 19.938 16.053 34.666 35.762 35.516 34.568 32.586 26.807 23.058 29.806 32.042 32.588 31.922 29.629 22.787 18.674 32.026 34.139 35.245 35.656 34.646 28.098 23.720 27.980 29.134 29.260 28.788 27.339 21.963 18.197 32.539 33.195 33.279 33.053 32.238 28.250 24.876 28.151 30.809 32.188 32.429 30.532 22.953 18.515 34.524 36.281 35.840 34.335 31.639 25.105 21.216 31.081 32.129 32.228 31.785 30.430 25.253 21.541 30.135 32.967 33.392 32.071 29.001 21.677 17.549 29.102 30.556 30.984 30.781 29.513 23.955 20.025 34.060 35.323 35.646 35.416 34.236 28.995 25.168 Copyright ⓒ 2016 SERSC 39 International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) 13 14 15 16 17 18 19 20 21 22 23 24 25 Avg. 27.066 28.867 28.357 26.765 23.996 17.417 13.527 31.566 32.278 31.017 29.184 26.509 20.378 16.676 32.366 36.854 39.725 38.991 34.208 24.997 20.489 30.785 34.416 36.969 37.858 34.689 25.325 20.645 34.763 34.897 33.754 32.201 29.870 24.187 20.613 33.849 35.477 35.962 35.716 34.270 28.373 24.374 31.560 34.016 34.738 34.120 31.709 24.615 20.438 38.297 39.116 38.434 37.111 34.843 28.983 25.293 39.813 42.589 41.703 39.344 35.893 28.803 24.836 45.281 39.370 35.614 32.885 29.883 23.918 20.401 37.565 39.680 38.878 36.887 33.788 26.995 23.085 27.119 27.046 26.130 24.874 22.888 17.650 14.197 36.571 38.236 37.632 36.016 33.272 26.767 22.904 32.944 34.428 34.340 33.330 30.980 24.567 20.643 Tables 1 and 2 are re-drawn in figures as shown in Fig. 9. It is noted that results of p=0.20 and p=0.30 are excluded as their MSE value is too large to show. (a) 40 Copyright ⓒ 2016 SERSC International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) (b) Figure 9. Objective Performance Comparison for Zahra set: (a) PSNR (in dB), (b) MSE. It is noted that results of p=0.20 and p=0.30 are not Included as their MSE Value is Very Large. Table 2. MSE Results for 25 Zahra Images 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 p=0 p=0.03 p=0.05 p=0.07 p=0.10 p=0.20 p=0.30 33.126 18.225 29.639 58.133 132.897 659.642 1613.361 22.206 17.255 18.261 22.713 35.851 135.633 321.553 67.994 40.637 35.831 41.771 70.830 342.313 882.443 40.787 25.069 19.435 17.678 22.311 100.747 276.095 103.545 79.366 77.112 85.950 120.005 413.773 984.848 36.241 31.157 30.561 32.198 38.842 97.291 211.588 99.545 53.975 39.292 37.168 57.527 329.456 915.333 22.944 15.311 16.948 23.966 44.581 200.734 491.402 50.698 39.829 38.930 43.107 58.892 194.007 456.044 63.037 32.841 29.776 40.363 81.840 441.945 1143.351 79.964 57.208 51.842 54.318 72.738 261.584 646.501 25.530 19.090 17.719 18.685 24.519 81.953 197.832 127.772 84.416 94.921 136.951 259.107 1178.595 2886.234 45.341 38.481 51.451 78.457 145.281 596.101 1397.798 37.714 13.418 6.928 8.204 24.678 205.784 581.034 54.278 23.524 13.067 10.647 22.087 190.819 560.475 21.714 21.054 27.398 39.169 67.002 247.968 564.613 26.801 18.424 16.476 17.436 24.329 94.568 237.518 45.400 25.793 21.840 25.181 43.870 224.701 587.892 9.624 7.970 9.325 12.648 21.320 82.187 192.225 Copyright ⓒ 2016 SERSC 41 International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) 21 22 23 24 25 Avg. 6.789 3.582 4.393 7.563 16.741 85.664 213.559 1.928 7.518 17.851 33.468 66.802 263.785 592.878 11.392 7.000 8.419 13.315 27.180 129.906 319.568 126.241 128.384 158.534 211.660 334.433 1117.049 2474.090 14.322 9.760 11.218 16.275 30.611 136.889 333.156 46.997 32.771 33.887 43.481 73.771 312.524 763.256 4. Conclusions In this paper, we studied image data fusion method. We consider FLIR camera which provides visible light and infrared images. We assume image edges (or details) can be obtained from visible light image. The obtained edges are finally added to original low resolution infrared image. Simulation results show that parameter p=0.03 gives the best performance. Acknowledgment This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT and Future Planning(2013R1A1A1010797). This paper is a revised and expanded version of a paper entitled “Data Fusion for Infrared Image” presented at MulGraB2015, November 25-28, 2015 at International Center, Jeju National University, Jeju Island, Korea. References [1] H. Li, B.S. Manjunath and S.K. Mitra, “Multi-sensor image fusion using the wavelet transform,” IEEE ICIP1994, vol. 1, (1994), pp. 51-55. [2] . Proen a, “On the feasi ilit on the visi le wavelength, at-a-distance and on-the-move iris recognition,” IEEE IEEE S mposium Series on Computational Intelligence in Biometrics: Theor , Algorithms, and Applications, (2009). [3] J. G. Daugman, “New methods in iris recognition,” IEEE Transactions on S stems, Man, and Cybernetics - Part B: Cybernetics, vol. 37, no. 5, (2007), pp. 1167-1175. [4] K. Park and J. Kim, “A real-time focusing algorithm for iris recognition camera”, IEEE Transactions on Systems, Man and Cybernetics, vol. 35, no. 3, (2005) August, pp. 441-444. [5] L. Ma, T. Tan, Y. Wang and D. Zhang, “Local intensit variations analysis for iris recognition”, Pattern recognition, vol. 37, no. 6, (2004), pp. 1287-1298. [6] Y. He, J. Cui, T. Tan and Y. Wang, “Ke techniques and methods for imaging iris in focus”, in Proceedings of the IEEE International Conference on Pattern Recognition, pp. 557-561, (2006) August. [7] J. R. Matley, D. Ackerman, J. Bergen and M. Tinker, “Iris recognition in less constrained environments,” Springer Advances in Biometrics: Sensors, Algorithms and S stems, (2007) October, pp. 107-131. [8] J. G. Daugman, “ ow iris recognition works”, IEEE Transactions on Circuits and Systems for Video Technology, vol. 14, no. 1, (2004) January, pp. 21-30. [9] J. L. Cam ier, “Iridian large database performance”, Iridian Technologies, Tech. Rep., http://iridiantech.com, (2007). [10] G. Jeon, “Contrast Intensification in NTSC YIQ,” IJCA vol. 6, no. 4, (2013) August, pp. 157-166. [11] G. Jeon, “Measuring and Comparison of Edge Detectors in Color Spaces,” IJCA vol. 6, no. 5, (2013) October, pp. 21-30. Author Gwanggil Jeon received the BS, MS, and PhD (summa cum laude) degrees in Department of Electronics and Computer Engineering from Hanyang University, Seoul, Korea, in 2003, 2005, and 2008, respectively. From 2008 to 2009, he was with the Department of Electronics and Computer Engineering, Hanyang University, from 2009 to 2011, he was with the School of 42 Copyright ⓒ 2016 SERSC International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) Information Technology and Engineering (SITE), University of Ottawa, as a postdoctoral fellow, and from 2011 to 2012, he was with the Graduate School of Science & Technology, Niigata University, as an assistant professor. He is currently an assistant professor with the Department of Embedded Systems Engineering, Incheon National University, Incheon, Korea. His research interests fall under the umbrella of image processing, particularly image compression, motion estimation, demosaicking, and image enhancement as well as computational intelligence such as fuzzy and rough sets theories. He was the recipient of the IEEE Chester Sall Award in 2007 and the 2008 ETRI Journal Paper Award. Copyright ⓒ 2016 SERSC 43 International Journal of Software Engineering and Its Applications Vol. 10, No. 2 (2016) 44 Copyright ⓒ 2016 SERSC
© Copyright 2026 Paperzz