Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning under Distribution Mismatch and Model Misspecification (2102.05695v3)

Published 10 Feb 2021 in cs.IT and math.IT

Abstract: We study learning algorithms when there is a mismatch between the distributions of the training and test datasets of a learning algorithm. The effect of this mismatch on the generalization error and model misspecification are quantified. Moreover, we provide a connection between the generalization error and the rate-distortion theory, which allows one to utilize bounds from the rate-distortion theory to derive new bounds on the generalization error and vice versa. In particular, the rate-distortion based bound strictly improves over the earlier bound by Xu and Raginsky even when there is no mismatch. We also discuss how "auxiliary loss functions" can be utilized to obtain upper bounds on the generalization error.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Saeed Masiha (6 papers)
  2. Amin Gohari (59 papers)
  3. Mohammad Hossein Yassaee (18 papers)
  4. Mohammad Reza Aref (69 papers)
Citations (19)

Summary

We haven't generated a summary for this paper yet.