Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
149 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Measuring Sharpness in Grokking (2402.08946v1)

Published 14 Feb 2024 in cs.LG

Abstract: Neural networks sometimes exhibit grokking, a phenomenon where perfect or near-perfect performance is achieved on a validation set well after the same performance has been obtained on the corresponding training set. In this workshop paper, we introduce a robust technique for measuring grokking, based on fitting an appropriate functional form. We then use this to investigate the sharpness of transitions in training and validation accuracy under two settings. The first setting is the theoretical framework developed by Levi et al. (2023) where closed form expressions are readily accessible. The second setting is a two-layer MLP trained to predict the parity of bits, with grokking induced by the concealment strategy of Miller et al. (2023). We find that trends between relative grokking gap and grokking sharpness are similar in both settings when using absolute and relative measures of sharpness. Reflecting on this, we make progress toward explaining some trends and identify the need for further study to untangle the various mechanisms which influence the sharpness of grokking.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (11)
  1. Hidden progress in deep learning: SGD learns parities near the computational limit. Advances in Neural Information Processing Systems, 35:21750–21764, 2022.
  2. Unifying grokking and double descent, 2023.
  3. Grokking in linear estimators – a solvable model that groks without understanding, 2023.
  4. Omnigrok: Grokking beyond algorithmic data. International Conference on Learning Representations, 2023.
  5. Dichotomy of early and late phase implicit biases can provably induce grokking, 2023.
  6. A tale of two circuits: Grokking as competition of sparse and dense subnetworks, 2023.
  7. Grokking beyond neural networks: An empirical exploration with model complexity, 2023.
  8. Progress measures for grokking via mechanistic interpretability, 2023.
  9. Grokking: Generalization beyond overfitting on small algorithmic datasets, 2022.
  10. Explaining grokking through circuit efficiency, 2023.
  11. Grokking phase transitions in learning local rules with gradient descent, 2022.

Summary

We haven't generated a summary for this paper yet.