Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Quantitative CLTs in Deep Neural Networks (2307.06092v5)

Published 12 Jul 2023 in cs.LG, cs.AI, math.PR, and stat.ML

Abstract: We study the distribution of a fully connected neural network with random Gaussian weights and biases in which the hidden layer widths are proportional to a large constant $n$. Under mild assumptions on the non-linearity, we obtain quantitative bounds on normal approximations valid at large but finite $n$ and any fixed network depth. Our theorems show both for the finite-dimensional distributions and the entire process, that the distance between a random fully connected network (and its derivatives) to the corresponding infinite width Gaussian process scales like $n{-\gamma}$ for $\gamma>0$, with the exponent depending on the metric used to measure discrepancy. Our bounds are strictly stronger in terms of their dependence on network width than any previously available in the literature; in the one-dimensional case, we also prove that they are optimal, i.e., we establish matching lower bounds.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Stefano Favaro (74 papers)
  2. Boris Hanin (50 papers)
  3. Domenico Marinucci (58 papers)
  4. Ivan Nourdin (67 papers)
  5. Giovanni Peccati (70 papers)
Citations (7)

Summary

We haven't generated a summary for this paper yet.