Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Visible and infrared self-supervised fusion trained on a single example (2307.04100v2)

Published 9 Jul 2023 in cs.CV

Abstract: Multispectral imaging is an important task of image processing and computer vision, which is especially relevant to applications such as dehazing or object detection. With the development of the RGBT (RGB & Thermal) sensor, the problem of visible (RGB) to Near Infrared (NIR) image fusion has become particularly timely. Indeed, while visible images see color, but suffer from noise, haze, and clouds, the NIR channel captures a clearer picture. The proposed approach fuses these two channels by training a Convolutional Neural Network by Self Supervised Learning (SSL) on a single example. For each such pair, RGB and NIR, the network is trained for seconds to deduce the final fusion. The SSL is based on the comparison of the Structure of Similarity and Edge-Preservation losses, where the labels for the SSL are the input channels themselves. This fusion preserves the relevant detail of each spectral channel without relying on a heavy training process. Experiments demonstrate that the proposed approach achieves similar or better qualitative and quantitative multispectral fusion results than other state-of-the-art methods that do not rely on heavy training and/or large datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. Slic superpixels compared to state-of-the-art superpixel methods. IEEE transactions on pattern analysis and machine intelligence, 34(11):2274–2282, 2012.
  2. E. B. Baruch and Y. Keller. Joint detection and matching of feature points in multimodal images. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(10):6585–6593, 2021.
  3. M. Brown and S. Süsstrunk. Multispectral SIFT for scene category recognition. In Computer Vision and Pattern Recognition (CVPR11), pages 177–184, Colorado Springs, June 2011.
  4. J. Canny. A computational approach to edge detection. IEEE Transactions on pattern analysis and machine intelligence, (6):679–698, 1986.
  5. Wavelets and image fusion. In Proceedings., International Conference on Image Processing, volume 3, pages 248–251. IEEE, 1995.
  6. Deep learning-based inverse scattering with structural similarity loss functions. IEEE Sensors Journal, 21(4):4900–4907, 2020.
  7. Spatial transformer networks. Advances in neural information processing systems, 28, 2015.
  8. A survey on contrastive self-supervised learning. Technologies, 9(1):2, 2020.
  9. S. S. Kumar and S. Muttan. Pca-based image fusion. In Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XII, volume 6233, page 62331T. International Society for Optics and Photonics, 2006.
  10. Challenge-aware rgbt tracking. In European Conference on Computer Vision, pages 222–237. Springer, 2020.
  11. Nestfuse: An infrared and visible image fusion architecture based on nest connection and spatial/channel attention models. IEEE Transactions on Instrumentation and Measurement, 69(12):9645–9656, 2020.
  12. Infrared and visible image fusion using a deep learning framework. In 2018 24th international conference on pattern recognition (ICPR), pages 2705–2710. IEEE, 2018.
  13. Target-aware dual adversarial learning and a multi-scenario multi-modality benchmark to fuse infrared and visible for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5802–5811, 2022.
  14. D. G. Lowe. Distinctive image features from scale-invariant keypoints. International journal of computer vision, 60(2):91–110, 2004.
  15. Y. Lu. The level weighted structural similarity loss: A step away from mse. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 9989–9990, 2019.
  16. Multimodality image registration by maximization of mutual information. IEEE transactions on Medical Imaging, 16(2):187–198, 1997.
  17. N. Ofir. Multispectral image fusion based on super pixel segmentation. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5. IEEE, 2023.
  18. N. Ofir and Y. Keller. Multi-scale processing of noisy images using edge preservation losses. In 2020 25th International Conference on Pattern Recognition (ICPR), pages 1–8. IEEE, 2021.
  19. Deep multi-spectral registration using invariant descriptor learning. In 2018 25th IEEE International Conference on Image Processing (ICIP), pages 1238–1242. IEEE, 2018.
  20. Registration and fusion of multi-spectral images using a novel edge descriptor. In 2018 25th IEEE International Conference on Image Processing (ICIP), pages 1857–1861. IEEE, 2018.
  21. U. Patil and U. Mudengudi. Image fusion using hierarchical pca. In 2011 international conference on image information processing, pages 1–6. IEEE, 2011.
  22. Singan: Learning a generative model from a single natural image. In Proceedings of the IEEE/CVF international conference on computer vision, pages 4570–4580, 2019.
  23. “zero-shot” super-resolution using deep internal learning. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3118–3126, 2018.
  24. U-net and its variants for medical image segmentation: A review of theory and applications. Ieee Access, 9:82031–82057, 2021.
  25. Image super-resolution via deep recursive residual network. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3147–3155, 2017.
  26. W. Wang and F. Chang. A multi-focus image fusion method based on laplacian pyramid. J. Comput., 6(12):2559–2566, 2011.
  27. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing, 13(4):600–612, 2004.

Summary

We haven't generated a summary for this paper yet.