Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

When can unlabeled data improve the learning rate? (1905.11866v2)

Published 28 May 2019 in cs.LG and stat.ML

Abstract: In semi-supervised classification, one is given access both to labeled and unlabeled data. As unlabeled data is typically cheaper to acquire than labeled data, this setup becomes advantageous as soon as one can exploit the unlabeled data in order to produce a better classifier than with labeled data alone. However, the conditions under which such an improvement is possible are not fully understood yet. Our analysis focuses on improvements in the minimax learning rate in terms of the number of labeled examples (with the number of unlabeled examples being allowed to depend on the number of labeled ones). We argue that for such improvements to be realistic and indisputable, certain specific conditions should be satisfied and previous analyses have failed to meet those conditions. We then demonstrate examples where these conditions can be met, in particular showing rate changes from $1/\sqrt{\ell}$ to $e{-c\ell}$ and from $1/\sqrt{\ell}$ to $1/\ell$. These results improve our understanding of what is and isn't possible in semi-supervised learning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Christina Göpfert (7 papers)
  2. Shai Ben-David (26 papers)
  3. Olivier Bousquet (33 papers)
  4. Sylvain Gelly (43 papers)
  5. Ilya Tolstikhin (21 papers)
  6. Ruth Urner (13 papers)
Citations (18)

Summary

We haven't generated a summary for this paper yet.