Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 27 tok/s
Gemini 2.5 Pro 46 tok/s Pro
GPT-5 Medium 23 tok/s Pro
GPT-5 High 29 tok/s Pro
GPT-4o 70 tok/s Pro
Kimi K2 117 tok/s Pro
GPT OSS 120B 459 tok/s Pro
Claude Sonnet 4 34 tok/s Pro
2000 character limit reached

An Empirical Study of Multi-Task Learning on BERT for Biomedical Text Mining (2005.02799v1)

Published 6 May 2020 in cs.CL

Abstract: Multi-task learning (MTL) has achieved remarkable success in natural language processing applications. In this work, we study a multi-task learning model with multiple decoders on varieties of biomedical and clinical natural language processing tasks such as text similarity, relation extraction, named entity recognition, and text inference. Our empirical results demonstrate that the MTL fine-tuned models outperform state-of-the-art transformer models (e.g., BERT and its variants) by 2.0% and 1.3% in biomedical and clinical domains, respectively. Pairwise MTL further demonstrates more details about which tasks can improve or decrease others. This is particularly helpful in the context that researchers are in the hassle of choosing a suitable model for new problems. The code and models are publicly available at https://github.com/ncbi-nlp/bluebert

Citations (102)

Summary

  • The paper demonstrates that multi-task learning with BERT improves performance on biomedical NLP tasks by sharing representations across eight distinct tasks.
  • The study compares baseline BERT, MT-BERT-Refinement, and MT-BERT-Fine-Tune, reporting a 2.0% enhancement in biomedical and 1.3% in clinical domains.
  • MT-BERT-Fine-Tune achieves state-of-the-art results on BLUE benchmark tasks, underscoring the potential of MTL in data-scarce biomedical settings.

Multi-Task Learning with BERT for Biomedical Text Mining

The empirical paper conducted in this paper investigates the application of Multi-Task Learning (MTL) with BERT-based models for various biomedical and clinical NLP tasks. The paper focuses on improving task performance in data-scarce domains through shared learning. The tasks examined include text similarity, relation extraction, named entity recognition, and text inference, with a particular focus on the Biomedical Language Understanding Evaluation (BLUE) benchmark.

Methodology

The paper outlines a model architecture where shared layers based on BERT are employed alongside task-specific layers. This structure supports the joint learning of eight distinct tasks across different biomedical and clinical datasets. The paper distinguishes three models for comparison: a baseline single-task model utilizing BERT, an MTL refinement model (MT-BERT-Refinement), and a fine-tuned model for each task (MT-BERT-Fine-Tune).

Results

The paper reports a performance enhancement of 2.0% in biomedical domains and 1.3% in clinical domains when using MTL compared to conventional fine-tuned BERT models. Notably, MT-BERT-Fine-Tune achieved new state-of-the-art results on four BLUE benchmark tasks. These improvements signify potential gains for researchers facing the challenge of selecting models for novel problems, particularly when training data is limited.

The research presented intriguing insights into task interactions. Through pairwise MTL analysis, it was observed that specific tasks, such as ShARe/CLEFE benefiting most from the MTL approach, can enhance performance when learned jointly with others. This underlines task suitability as a crucial factor in MTL model gains.

Implications and Future Directions

The findings underscore the effectiveness of MTL in resource-limited biomedical and clinical contexts, suggesting that domain-specific pretraining, especially utilizing datasets like PubMed and MIMIC-III, enhances performance. However, the paper also acknowledges that MTL doesn't universally lead to performance gains across all task combinations within different domains.

Future work could explore deeper analyses of task relationships to understand under which conditions MTL most effectively yields performance improvements. Moreover, the research suggests potential in exploring alternative MTL strategies, such as soft parameter sharing or knowledge distillation, to optimize task interactions and enhance model generalization.

The code and pre-trained models are openly available, offering a resource for further exploration and validation within the research community, potentially fostering advancements in biomedical NLP applications. This work provides a foundation for developing robust, generalizable NLP models that leverage MTL's capabilities to handle multi-faceted biomedical text mining tasks.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Github Logo Streamline Icon: https://streamlinehq.com