Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Few-Shot Text Classification with Triplet Networks, Data Augmentation, and Curriculum Learning (2103.07552v1)

Published 12 Mar 2021 in cs.CL

Abstract: Few-shot text classification is a fundamental NLP task in which a model aims to classify text into a large number of categories, given only a few training examples per category. This paper explores data augmentation -- a technique particularly suitable for training with limited data -- for this few-shot, highly-multiclass text classification setting. On four diverse text classification tasks, we find that common data augmentation techniques can improve the performance of triplet networks by up to 3.0% on average. To further boost performance, we present a simple training strategy called curriculum data augmentation, which leverages curriculum learning by first training on only original examples and then introducing augmented data as training progresses. We explore a two-stage and a gradual schedule, and find that, compared with standard single-stage training, curriculum data augmentation trains faster, improves performance, and remains robust to high amounts of noising from augmentation.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jason Wei (49 papers)
  2. Chengyu Huang (14 papers)
  3. Soroush Vosoughi (90 papers)
  4. Yu Cheng (354 papers)
  5. Shiqi Xu (19 papers)
Citations (60)

Summary

We haven't generated a summary for this paper yet.