Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Overcoming Dimensional Collapse in Self-supervised Contrastive Learning for Medical Image Segmentation (2402.14611v2)

Published 22 Feb 2024 in cs.CV

Abstract: Self-supervised learning (SSL) approaches have achieved great success when the amount of labeled data is limited. Within SSL, models learn robust feature representations by solving pretext tasks. One such pretext task is contrastive learning, which involves forming pairs of similar and dissimilar input samples, guiding the model to distinguish between them. In this work, we investigate the application of contrastive learning to the domain of medical image analysis. Our findings reveal that MoCo v2, a state-of-the-art contrastive learning method, encounters dimensional collapse when applied to medical images. This is attributed to the high degree of inter-image similarity shared between the medical images. To address this, we propose two key contributions: local feature learning and feature decorrelation. Local feature learning improves the ability of the model to focus on the local regions of the image, while feature decorrelation removes the linear dependence among the features. Our experimental findings demonstrate that our contributions significantly enhance the model's performance in the downstream task of medical segmentation, both in the linear evaluation and full fine-tuning settings. This work illustrates the importance of effectively adapting SSL techniques to the characteristics of medical imaging tasks. The source code will be made publicly available at: https://github.com/CAMMA-public/med-moco

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. “nnu-net: Self-adapting framework for u-net-based medical image segmentation,” in MICCAI, 2019.
  2. “Transunet: Transformers make strong encoders for medical image segmentation,” arXiv:2102.04306, 2021.
  3. “Unetr: Transformers for 3d medical image segmentation,” in WACV, 2022.
  4. “Simmim: A simple framework for masked image modeling,” in CVPR, 2022.
  5. “Improved baselines with momentum contrastive learning,” arXiv:2003.04297, 2020.
  6. “A simple framework for contrastive learning of visual representations,” in ICML, 2020.
  7. “Momentum contrast for unsupervised visual representation learning,” in CVPR, 2020.
  8. “Bootstrap your own latent-a new approach to self-supervised learning,” NIPS, 2020.
  9. “On feature decorrelation in self-supervised learning,” in CVPR, 2021.
  10. “Abdomenct-1k: Is abdominal organ segmentation a solved problem?,” IEEE TPAMI, 2021.
  11. “Miccai multi-atlas labeling beyond the cranial vault–workshop and challenge,” in MICCAI Workshop, 2015.
  12. “Deep metric learning via lifted structured feature embedding,” in CVPR, 2016.
  13. “Deep residual learning for image recognition,” in CVPR, 2016.
  14. “Batch normalization: Accelerating deep network training by reducing internal covariate shift,” in ICML, 2015.
  15. “Decorrelated batch normalization,” in CVPR, 2018.
  16. “Iterative normalization: Beyond standardization towards efficient whitening,” in CVPR, 2019.
  17. MMSelfSup Contributors, “MMSelfSup: Openmmlab self-supervised learning toolbox and benchmark,” https://github.com/open-mmlab/mmselfsup, 2021.
  18. MMSegmentation Contributors, “MMSegmentation: Openmmlab semantic segmentation toolbox and benchmark,” https://github.com/open-mmlab/mmsegmentation, 2020.
  19. “Vicregl: Self-supervised learning of local visual features,” NIPS, 2022.
  20. “Dcan: Deep contour-aware networks for accurate gland segmentation,” IEEE TMI, 2018.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Jamshid Hassanpour (1 paper)
  2. Vinkle Srivastav (23 papers)
  3. Didier Mutter (37 papers)
  4. Nicolas Padoy (93 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.