Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 45 tok/s
Gemini 2.5 Pro 54 tok/s Pro
GPT-5 Medium 22 tok/s Pro
GPT-5 High 20 tok/s Pro
GPT-4o 99 tok/s Pro
Kimi K2 183 tok/s Pro
GPT OSS 120B 467 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

SLACK: Stable Learning of Augmentations with Cold-start and KL regularization (2306.09998v1)

Published 16 Jun 2023 in cs.CV and cs.LG

Abstract: Data augmentation is known to improve the generalization capabilities of neural networks, provided that the set of transformations is chosen with care, a selection often performed manually. Automatic data augmentation aims at automating this process. However, most recent approaches still rely on some prior information; they start from a small pool of manually-selected default transformations that are either used to pretrain the network or forced to be part of the policy learned by the automatic data augmentation algorithm. In this paper, we propose to directly learn the augmentation policy without leveraging such prior knowledge. The resulting bilevel optimization problem becomes more challenging due to the larger search space and the inherent instability of bilevel optimization algorithms. To mitigate these issues (i) we follow a successive cold-start strategy with a Kullback-Leibler regularization, and (ii) we parameterize magnitudes as continuous distributions. Our approach leads to competitive results on standard benchmarks despite a more challenging setting, and generalizes beyond natural images.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (31)
  1. Non-convex bilevel games with critical point selection maps. preprint arXiv:2207.04888, 2022.
  2. Autoaugment: Learning augmentation strategies from data. In Proc. CVPR, 2019.
  3. RandAugment: Practical automated data augmentation with a reduced search space. In Proc. NeurIPS, 2020.
  4. ImageNet: A large-scale hierarchical image database. In Proc. CVPR, 2009.
  5. Improved regularization of convolutional neural networks with cutout. arXiv preprint arXiv:1708.04552, 2017.
  6. Michael C Fu. Gradient estimation. Handbooks in operations research and management science, 13:575–616, 2006.
  7. Backpropagation through the void: Optimizing control variates for black-box gradient estimation. In Proc. ICLR, 2018.
  8. In search of lost domain generalization. In Proc. ICLR, 2021.
  9. Faster AutoAugment: Learning augmentation strategies using backpropagation. In Proc. ECCV, 2020.
  10. Meta approach to data augmentation optimization. In Proc. WACV, 2022.
  11. Deep residual learning for image recognition. In Proc. CVPR, 2016.
  12. Population based augmentation: Efficient learning of augmentation policy schedules. In Proc. ICML, pages 2731–2741, 2019.
  13. Learning multiple layers of features from tiny images. 2009.
  14. DADA: differentiable automatic data augmentation. In Proc. ECCV, 2020.
  15. Fast AutoAugment. In Proc. NeurIPS, 2019.
  16. Online hyper-parameter learning for auto-augmentation strategy. In Proc. ICCV, 2019.
  17. Uniformaugment: A search-free probabilistic data augmentation approach. arXiv preprint arXiv:2003.14348, 2020.
  18. Direct differentiable augmentation search. In Proc. ICCV, 2021.
  19. TrivialAugment: tuning-free yet state-of-the-art data augmentation. In Proc. ICCV, 2021.
  20. Transformation pursuit for image classification. In Proc. CVPR, 2014.
  21. Moment matching for multi-source domain adaptation. In Proc. ICCV, pages 1406–1415, 2019.
  22. R Tyrrell Rockafellar. Monotone operators and the proximal point algorithm. SIAM journal on control and optimization, 14(5):877–898, 1976.
  23. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347, 2017.
  24. Teppei Suzuki. Teachaugment: Data augmentation optimization using teacher knowledge. In Proc. CVPR, 2022.
  25. Contrastive multiview coding. arXiv preprint arXiv:1906.05849, 2019.
  26. Addressing model vulnerability to distributional shifts over image transformation sets. In Proc. ICCV, pages 7980–7989, 2019.
  27. DAAS: Differentiable architecture and augmentation policy search. arXiv preprint arXiv:2109.15273, 2021.
  28. Ronald J Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine learning, 8(3):229–256, 1992.
  29. Wide residual networks. In Edwin R. Hancock Richard C. Wilson and William A. P. Smith, editors, Proc. BMVC, pages 87.1–87.12. BMVA Press, September 2016.
  30. Adversarial AutoAugment. In Proc. ICLR, 2020.
  31. Deep AutoAugmentation. In Proc. ICLR, 2022.
Citations (4)

Summary

We haven't generated a summary for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Lightbulb On Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube