Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 134 tok/s
Gemini 2.5 Pro 41 tok/s Pro
GPT-5 Medium 37 tok/s Pro
GPT-5 High 35 tok/s Pro
GPT-4o 125 tok/s Pro
Kimi K2 203 tok/s Pro
GPT OSS 120B 429 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Investigating the Robustness of Vision Transformers against Label Noise in Medical Image Classification (2402.16734v1)

Published 26 Feb 2024 in eess.IV, cs.CV, and cs.LG

Abstract: Label noise in medical image classification datasets significantly hampers the training of supervised deep learning methods, undermining their generalizability. The test performance of a model tends to decrease as the label noise rate increases. Over recent years, several methods have been proposed to mitigate the impact of label noise in medical image classification and enhance the robustness of the model. Predominantly, these works have employed CNN-based architectures as the backbone of their classifiers for feature extraction. However, in recent years, Vision Transformer (ViT)-based backbones have replaced CNNs, demonstrating improved performance and a greater ability to learn more generalizable features, especially when the dataset is large. Nevertheless, no prior work has rigorously investigated how transformer-based backbones handle the impact of label noise in medical image classification. In this paper, we investigate the architectural robustness of ViT against label noise and compare it to that of CNNs. We use two medical image classification datasets -- COVID-DU-Ex, and NCT-CRC-HE-100K -- both corrupted by injecting label noise at various rates. Additionally, we show that pretraining is crucial for ensuring ViT's improved robustness against label noise in supervised training.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. Quantifying attention flow in transformers. arXiv preprint arXiv:2005.00928, 2020.
  2. Training a neural network based on unreliable human annotation of medical images. In 2018 IEEE 15th International symposium on biomedical imaging (ISBI 2018), pages 39–42. IEEE, 2018.
  3. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
  4. Co-teaching: Robust training of deep neural networks with extremely noisy labels. Advances in neural information processing systems, 31, 2018.
  5. A survey on vision transformer. IEEE transactions on pattern analysis and machine intelligence, 45(1):87–110, 2022.
  6. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16000–16009, 2022.
  7. Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pages 590–597, 2019.
  8. Predicting survival from colorectal cancer histology slides using deep learning: A retrospective multicenter study. PLoS medicine, 16(1):e1002730, 2019.
  9. Improving medical image classification in noisy labels using only self-supervised pretraining. In MICCAI Workshop on Data Engineering in Medical Imaging, pages 78–90. Springer, 2023.
  10. Investigating the impact of class-dependent label noise in medical image classification. In Medical Imaging 2023: Image Processing, volume 12464, pages 728–733. SPIE, 2023.
  11. How does heterogeneous label noise impact generalization in neural nets? In Advances in Visual Computing: 16th International Symposium, ISVC 2021, Virtual Event, October 4-6, 2021, Proceedings, Part II, pages 229–241. Springer, 2021.
  12. Pancreatic cancer detection in whole slide images using noisy label annotations. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd International Conference, Shenzhen, China, October 13–17, 2019, Proceedings, Part I 22, pages 541–549. Springer, 2019.
  13. Robust inference via generative classifiers for handling noisy labels. In International conference on machine learning, pages 3763–3772. PMLR, 2019.
  14. Dividemix: Learning with noisy labels as semi-supervised learning. In International Conference on Learning Representations, 2020.
  15. Selective-supervised contrastive learning with noisy labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 316–325, 2022.
  16. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101, 2017.
  17. A survey of crowdsourcing in medical image analysis. Human Computation, 7:1–26, 2020.
  18. Robustness to label noise depends on the shape of the noise distribution. Advances in Neural Information Processing Systems, 35:35645–35656, 2022.
  19. Interpreting chest x-rays via cnns that exploit hierarchical disease dependencies and uncertainty labels. Neurocomputing, 437:186–194, 2021.
  20. Labelling instructions matter in biomedical image analysis. Nature Machine Intelligence, 5(3):273–283, 2023.
  21. Rochester Institute of Technology. Research computing services, 2022.
  22. Transformers in medical imaging: A survey. Medical Image Analysis, page 102802, 2023.
  23. Covid-19 infection localization and severity grading from chest x-ray images. Computers in biology and medicine, 139:105002, 2021.
  24. Simmim: A simple framework for masked image modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9653–9663, 2022.
  25. Robust learning at noisy labeled medical images: Applied to skin lesion classification. In 2019 IEEE 16th International symposium on biomedical imaging (ISBI 2019), pages 1280–1283. IEEE, 2019.
  26. Robust medical image classification from noisy labeled data with global and local representation guided co-training. IEEE Transactions on Medical Imaging, 41(6):1371–1382, 2022.
  27. Understanding deep learning (still) requires rethinking generalization. Communications of the ACM, 64(3):107–115, 2021.
  28. Combating medical noisy labels by disentangled distribution learning and consistency regularization. Future Generation Computer Systems, 141:567–576, 2023.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 1 like.

Upgrade to Pro to view all of the tweets about this paper: