Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Swin MAE: Masked Autoencoders for Small Datasets (2212.13805v2)

Published 28 Dec 2022 in cs.CV and cs.AI

Abstract: The development of deep learning models in medical image analysis is majorly limited by the lack of large-sized and well-annotated datasets. Unsupervised learning does not require labels and is more suitable for solving medical image analysis problems. However, most of the current unsupervised learning methods need to be applied to large datasets. To make unsupervised learning applicable to small datasets, we proposed Swin MAE, which is a masked autoencoder with Swin Transformer as its backbone. Even on a dataset of only a few thousand medical images and without using any pre-trained models, Swin MAE is still able to learn useful semantic features purely from images. It can equal or even slightly outperform the supervised model obtained by Swin Transformer trained on ImageNet in terms of the transfer learning results of downstream tasks. The code is publicly available at https://github.com/Zian-Xu/Swin-MAE.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Zi'an Xu (3 papers)
  2. Yin Dai (8 papers)
  3. Fayu Liu (5 papers)
  4. Weibing Chen (9 papers)
  5. Yue Liu (257 papers)
  6. Lifu Shi (4 papers)
  7. Sheng Liu (122 papers)
  8. Yuhang Zhou (52 papers)
Citations (20)

Summary

We haven't generated a summary for this paper yet.

Github Logo Streamline Icon: https://streamlinehq.com