|
[1]J. Ma, J. Zhao, Y. Ma, and J. Tian, "Non-rigid visible and infrared face registration via regularized Gaussian fields criterion," Pattern Recognition, vol. 48, no. 3, pp. 772-784, 2015. [2]A. Toet, J. K. IJspeert, A. M. Waxman, and M. Aguilar, "Fusion of visible and thermal imagery improves situational awareness," Displays, vol. 18, no. 2, pp. 85-95, 1997. [3] J. Kocić, N. Jovičić, and V. Drndarević, "Sensors and sensor fusion in autonomous vehicles," in 2018 26th Telecommunications Forum (TELFOR), 2018: IEEE, pp. 420-425. [4]S. Li, X. Kang, L. Fang, J. Hu, and H. Yin, "Pixel-level image fusion: A survey of the state of the art," information Fusion, vol. 33, pp. 100-112, 2017. [5]Y. Liu, L. Wang, J. Cheng, C. Li, and X. Chen, "Multi-focus image fusion: A survey of the state of the art," Information Fusion, vol. 64, pp. 71-91, 2020. [6]P. Chai, X. Luo, and Z. Zhang, "Image fusion using quaternion wavelet transform and multiple features," IEEE access, vol. 5, pp. 6724-6734, 2017. [7]A. P. James and B. V. Dasarathy, "Medical image fusion: A survey of the state of the art," Information fusion, vol. 19, pp. 4-19, 2014. [8]Y. Yang, L. Wu, S. Huang, W. Wan, and Y. Que, "Remote sensing image fusion based on adaptively weighted joint detail injection," IEEE Access, vol. 6, pp. 6849-6864, 2018. [9]H. Ghassemian, "A review of remote sensing image fusion methods," Information Fusion, vol. 32, pp. 75-89, 2016. [10]Z. Wang, D. Ziou, C. Armenakis, D. Li, and Q. Li, "A comparative analysis of image fusion methods," IEEE transactions on geoscience and remote sensing, vol. 43, no. 6, pp. 1391-1402, 2005. [11] R. A. Newcombe et al., "Kinectfusion: Real-time dense surface mapping and tracking," in 2011 10th IEEE international symposium on mixed and augmented reality, 2011: IEEE, pp. 127-136. [12]G. Xiao, D. P. Bavirisetti, G. Liu, and X. Zhang, Image Fusion. Springer Singapore, 2020. [13]A. Toet, "TNO Image Fusion Dataset." [Online]. Available: https://figshare.com/articles/dataset/TNO_Image_Fusion_Dataset/1008029/1 [14]G. Petschnigg, R. Szeliski, M. Agrawala, M. Cohen, H. Hoppe, and K. Toyama, "Digital photography with flash and no-flash image pairs," ACM transactions on graphics (TOG), vol. 23, no. 3, pp. 664-672, 2004. [15]Y. Zhang, L. Zhang, X. Bai, and L. Zhang, "Infrared and visual image fusion through infrared feature extraction and visual information preservation," Infrared Physics & Technology, vol. 83, pp. 227-237, 2017. [16]L. Meylan and S. Susstrunk, "High dynamic range image rendering with a retinex-based adaptive filter," IEEE Transactions on image processing, vol. 15, no. 9, pp. 2820-2830, 2006. [17]J. Ma, Y. Ma, and C. Li, "Infrared and visible image fusion methods and applications: A survey," Information Fusion, vol. 45, pp. 153-178, 2019. [18]Z. Zhou, M. Dong, X. Xie, and Z. Gao, "Fusion of infrared and visible images for night-vision context enhancement," Applied optics, vol. 55, no. 23, pp. 6480-6490, 2016. [19]Y. Zhou, K. Gao, Z. Dou, Z. Hua, and H. Wang, "Target-aware fusion of infrared and visible images," IEEE Access, vol. 6, pp. 79039-79049, 2018. [20]J. Ma, W. Yu, P. Liang, C. Li, and J. Jiang, "FusionGAN: A generative adversarial network for infrared and visible image fusion," Information Fusion, vol. 48, pp. 11-26, 2019. [21]J. Ma and Y. Zhou, "Infrared and visible image fusion via gradientlet filter," Computer Vision and Image Understanding, vol. 197, p. 103016, 2020. [22]S. Li, X. Kang, and J. Hu, "Image fusion with guided filtering," IEEE Transactions on Image processing, vol. 22, no. 7, pp. 2864-2875, 2013. [23]S. M. Pizer et al., "Adaptive histogram equalization and its variations," Computer vision, graphics, and image processing, vol. 39, no. 3, pp. 355-368, 1987. [24]A. M. Reza, "Realization of the contrast limited adaptive histogram equalization (CLAHE) for real-time image enhancement," Journal of VLSI signal processing systems for signal, image and video technology, vol. 38, no. 1, pp. 35-44, 2004. [25]J. Ma, X. Fan, S. X. Yang, X. Zhang, and X. Zhu, "Contrast limited adaptive histogram equalization-based fusion in YIQ and HSI color spaces for underwater image enhancement," International Journal of Pattern Recognition and Artificial Intelligence, vol. 32, no. 07, p. 1854018, 2018. [26]H. Ibrahim and N. S. P. Kong, "Brightness preserving dynamic histogram equalization for image contrast enhancement," IEEE Transactions on Consumer Electronics, vol. 53, no. 4, pp. 1752-1758, 2007. [27]P. Gupta, J. Kumare, U. Singh, and R. Singh, "Histogram based image enhancement techniques: a survey," Int J Comput Sci Eng, vol. 5, no. 6, pp. 475-484, 2017. [28]J.-Y. Kim, L.-S. Kim, and S.-H. Hwang, "An advanced contrast enhancement using partially overlapped sub-block histogram equalization," IEEE transactions on circuits and systems for video technology, vol. 11, no. 4, pp. 475-484, 2001. [29]E. D. Pisano et al., "Contrast limited adaptive histogram equalization image processing to improve the detection of simulated spiculations in dense mammograms," Journal of Digital imaging, vol. 11, no. 4, p. 193, 1998. [30] T. Jintasuttisak and S. Intajag, "Color retinal image enhancement by Rayleigh contrast-limited adaptive histogram equalization," in 2014 14th International Conference on Control, Automation and Systems (ICCAS 2014), 2014: IEEE, pp. 692-697. [31] C. Tomasi and R. Manduchi, "Bilateral filtering for gray and color images," in Computer Vision, 1998. Sixth International Conference on, 1998: IEEE, pp. 839-846. [32]Y. Zhang, D. Li, and W. Zhu, "Infrared and Visible Image Fusion with Hybrid Image Filtering," Mathematical Problems in Engineering, vol. 2020, 2020. [33]K. He, J. Sun, and X. Tang, "Guided image filtering," IEEE transactions on pattern analysis and machine intelligence, vol. 35, no. 6, pp. 1397-1409, 2012. [34]K. He and J. Sun, "Fast guided filter," arXiv preprint arXiv:1505.00996, 2015. [35]L. Xu, C. Lu, Y. Xu, and J. Jia, "Image smoothing via L 0 gradient minimization," in Proceedings of the 2011 SIGGRAPH Asia conference, 2011, pp. 1-12. [36]K. Subr, C. Soler, and F. Durand, "Edge-preserving multiscale image decomposition based on local extrema," ACM Transactions on Graphics (TOG), vol. 28, no. 5, pp. 1-9, 2009. [37]Z. Farbman, R. Fattal, D. Lischinski, and R. Szeliski, "Edge-preserving decompositions for multi-scale tone and detail manipulation," ACM Transactions on Graphics (TOG), vol. 27, no. 3, pp. 1-10, 2008. [38]L. Itti, C. Koch, and E. Niebur, "A model of saliency-based visual attention for rapid scene analysis," IEEE Transactions on Pattern Analysis & Machine Intelligence, no. 11, pp. 1254-1259, 1998. [39]J. Chen, K. Wu, Z. Cheng, and L. Luo, "A saliency-based multiscale approach for infrared and visible image fusion," Signal Processing, vol. 182, p. 107936, 2021. [40]Z. Zhou, B. Wang, S. Li, and M. Dong, "Perceptual fusion of infrared and visible images through a hybrid multi-scale decomposition with Gaussian and bilateral filters," Information Fusion, vol. 30, pp. 15-26, 2016. [41]G. Cui, H. Feng, Z. Xu, Q. Li, and Y. Chen, "Detail preserved fusion of visible and infrared images using regional saliency extraction and multi-scale image decomposition," Optics Communications, vol. 341, pp. 199-209, 2015. [42]R. M. Haralick, K. Shanmugam, and I. H. Dinstein, "Textural features for image classification," IEEE Transactions on systems, man, and cybernetics, no. 6, pp. 610-621, 1973. [43]Y. Han, Y. Cai, Y. Cao, and X. Xu, "A new image fusion performance metric based on visual information fidelity," Information fusion, vol. 14, no. 2, pp. 127-135, 2013. [44]W. Xue, L. Zhang, X. Mou, and A. C. Bovik, "Gradient magnitude similarity deviation: A highly efficient perceptual image quality index," IEEE Transactions on Image Processing, vol. 23, no. 2, pp. 684-695, 2013.
|