Emergent Mind

Abstract

Whole-slide image (WSI) analysis plays a crucial role in cancer diagnosis and treatment. In addressing the demands of this critical task, self-supervised learning (SSL) methods have emerged as a valuable resource, leveraging their efficiency in circumventing the need for a large number of annotations, which can be both costly and time-consuming to deploy supervised methods. Nevertheless, patch-wise representation may exhibit instability in performance, primarily due to class imbalances stemming from patch selection within WSIs. In this paper, we introduce Nearby Patch Contrastive Learning (NearbyPatchCL), a novel self-supervised learning method that leverages nearby patches as positive samples and a decoupled contrastive loss for robust representation learning. Our method demonstrates a tangible enhancement in performance for downstream tasks involving patch-level multi-class classification. Additionally, we curate a new dataset derived from WSIs sourced from the Canine Cutaneous Cancer Histology, thus establishing a benchmark for the rigorous evaluation of patch-level multi-class classification methodologies. Intensive experiments show that our method significantly outperforms the supervised baseline and state-of-the-art SSL methods with top-1 classification accuracy of 87.56%. Our method also achieves comparable results while utilizing a mere 1% of labeled data, a stark contrast to the 100% labeled data requirement of other approaches. Source code: https://github.com/nvtien457/NearbyPatchCL

Overview

  • Whole-slide imaging (WSI) is essential in cancer diagnosis, but it requires automated analysis due to the amount of data generated. Deep learning is a prominent tool in this field, but it often needs extensive labeled datasets.

  • Self-supervised learning (SSL) is a promising approach in WSI analysis because it uses unlabeled data for pre-training and requires fewer labeled samples for fine-tuning.

  • The Nearby Patch Contrastive Learning (NearbyPatchCL) utilizes adjacent patches and a decoupled contrastive loss to learn robust features without extensive annotations.

  • NearbyPatchCL demonstrates superior performance on the P-CATCH dataset, outperforming traditional supervised methods and other SSL techniques, especially when labeled data is scarce.

  • The research provides the community with a new WSI dataset and presents a novel SSL framework adept at handling the imbalanced data commonly found in WSI analysis.

Introduction

Whole-slide image (WSI) analysis is a critical component in modern cancer diagnosis and treatment. The shift toward digitizing tissue slides has increased the need for automated and precise analysis algorithms. In this context, deep learning has taken center stage, often requiring large annotated datasets for training. Since annotations for WSI are time-consuming and expensive, self-supervised learning (SSL) methods, which do not require annotated data, have become a key interest. SSL uses a pre-training phase with unlabeled data, followed by a fine-tuned phase on a smaller labeled dataset. Although this approach is promising, the instability arises when these methods are subjected to imbalanced datasets, a common occurrence in WSI.

Self-Supervised Learning in Whole-Slide Images

Unlike traditional machine learning, SSL does not rely on extensive labeled datasets. It uses unlabeled data to learn representations before fine-tuning on a specific task with limited labeled data. Previous SSL approaches have applied random cropping to generate patches from WSIs. However, due to the diverse and skewed nature of tissue sections, this can result in imbalance problems. Furthermore, while contrastive learning has been employed to improve this process, it is not wholly resistant to issues caused by imbalanced data. To mitigate this, new methods are needed that can be effectively integrated into medical routines.

NearbyPatchCL Methodology

To address these challenges, a novel method called Nearby Patch Contrastive Learning (NearbyPatchCL) is introduced. This method uses a contrastive learning paradigm that treats adjacent patches as positive samples and employs a decoupled contrastive loss (DCL) to assist in learning robust representations. NearbyPatchCL seeks to achieve strong, stable features by viewing nearby patches as affirming, which is in contrast to existing methods that only consider different views of the same patch as positive.

What sets NearbyPatchCL apart is that it effectively benefits from leveraging neighboring patches in WSI without needing extensive annotation. This offers a way to leverage label-like information in a self-supervised setting. By enhancing performance with a fraction of labeled data, it shows great potential for real-world clinical scenarios that often struggle with limited annotations.

Evaluation and Contributions

To validate this method, the researchers curated a new dataset named P-CATCH from public WSIs focused on canine cancer. Rigorous experiments on P-CATCH demonstrated that NearbyPatchCL notably outperformed both the conventional supervised baseline and other advanced SSL methods. Specifically, it achieved top-1 classification accuracy of above 87% even when using only 1% labeled data, underscoring its efficiency.

In summary, this study contributes a novel SSL framework that copes well with the complex challenges of imbalanced data in WSI. It suggests that NearbyPatchCL could indeed transition into clinical practice, offering a workable solution for high-quality WSI analysis with limited annotations. The research also released a new publicly available WSI dataset, fostering further development and benchmarking of patch-level multi-class classification methods in digital pathology.

Create an account to read this summary for free:

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.