Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Classification vs regression in overparameterized regimes: Does the loss function matter? (2005.08054v2)

Published 16 May 2020 in cs.LG, cs.IT, math.IT, and stat.ML

Abstract: We compare classification and regression tasks in an overparameterized linear model with Gaussian features. On the one hand, we show that with sufficient overparameterization all training points are support vectors: solutions obtained by least-squares minimum-norm interpolation, typically used for regression, are identical to those produced by the hard-margin support vector machine (SVM) that minimizes the hinge loss, typically used for training classifiers. On the other hand, we show that there exist regimes where these interpolating solutions generalize well when evaluated by the 0-1 test loss function, but do not generalize if evaluated by the square loss function, i.e. they approach the null risk. Our results demonstrate the very different roles and properties of loss functions used at the training phase (optimization) and the testing phase (generalization).

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Vidya Muthukumar (33 papers)
  2. Adhyyan Narang (7 papers)
  3. Vignesh Subramanian (6 papers)
  4. Mikhail Belkin (76 papers)
  5. Daniel Hsu (107 papers)
  6. Anant Sahai (49 papers)
Citations (139)

Summary

We haven't generated a summary for this paper yet.