Emergent Mind

VeLO: Training Versatile Learned Optimizers by Scaling Up

(2211.09760)
Published Nov 17, 2022 in cs.LG , math.OC , and stat.ML

Abstract

While deep learning models have replaced hand-designed features across many domains, these models are still trained with hand-designed optimizers. In this work, we leverage the same scaling approach behind the success of deep learning to learn versatile optimizers. We train an optimizer for deep learning which is itself a small neural network that ingests gradients and outputs parameter updates. Meta-trained with approximately four thousand TPU-months of compute on a wide variety of optimization tasks, our optimizer not only exhibits compelling performance, but optimizes in interesting and unexpected ways. It requires no hyperparameter tuning, instead automatically adapting to the specifics of the problem being optimized. We open source our learned optimizer, meta-training code, the associated train and test data, and an extensive optimizer benchmark suite with baselines at velo-code.github.io.

Optimizer performance on VeLOdrome benchmark: VeLO outperforms learning rate-tuned baselines and NAdamW with extensive tuning.

Overview

  • VeLO is a meta-learned optimizer that outperforms traditional optimizers by using a neural network-based architecture and requiring no hyperparameter tuning.

  • The optimizer employs a hierarchical design combining LSTM and MLP components and is meta-trained using Evolution Strategies across extensive tasks to ensure generalizability.

  • VeLO shows superior performance in benchmarks, often being faster and more effective than the best-tuned Adam optimizers, demonstrating potential for significant advancements in machine learning optimization.

VeLO: Meta-Learned Optimization

Introduction

The presented paper details the creation and evaluation of VeLO, a learned optimizer for deep learning, which demonstrates impressive performance across a variety of tasks. Unlike traditional optimizers that rely on hand-designed algorithms and hyperparameter tuning, VeLO is a neural network-based optimizer meta-trained using approximately four thousand TPU-months of computational resources. It shows remarkable adaptability and efficiency, requiring no hyperparameter tuning and automatically adjusting to the task at hand.

Training and Meta-Training of VeLO

VeLO represents a significant scaling of previous work in learned optimizers. The novel optimizer architecture is hierarchical, consisting of a per-tensor LSTM which processes aggregated information about tensor statistics, and a per-parameter MLP that operates on a small set of features. This design allows VeLO to remain computationally efficient while being expressive enough to capture complex optimization strategies.

The meta-training process employed Evolution Strategies (ES) to estimate gradients for the meta-objective. Meta-training covered an extensive set of tasks, including various network architectures and learning problems, ensuring that VeLO could generalize across different domains. To manage the computational cost, the training curriculum incorporated task-augmentation and rejection sampling, favoring tasks with shorter run times.

Evaluation Metrics and Baselines

VeLO was evaluated using multiple benchmarks:

  1. VeLOdrome: Includes 83 canonical tasks with diverse model architectures.
  2. MLCommons: Tests on a variety of large-scale tasks.
  3. Real-world Tasks: Applications such as object detection, LLMs, and vision transformers.

Baselines included traditional optimizers like Adam, tuned over numerous hyperparameter trials, as well as learned optimizers from prior work. Evaluations revealed that VeLO outperformed these baselines across a broad range of tasks, even when the baselines received extensive hyperparameter tuning.

Numerical Results

The performance evaluation was compelling. On VeLOdrome tasks, VeLO was consistently faster and often achieved lower final losses than the best-tuned Adam optimizations. For example, on more than 50% of tasks, VeLO was over four times faster. In MLCommons benchmarks, VeLO either matched or outperformed heavily tuned Adam optimizers across several large-scale tasks without any hyperparameter tuning. The evaluation demonstrated that VeLO is particularly effective in scenarios where optimization is traditionally challenging, such as training large-scale vision transformers and decision transformers.

Implications and Future Work

This work illustrates a considerable advancement in optimization capabilities through meta-learning. By demonstrating that learned optimizers can outperform hand-designed counterparts without hyperparameter tuning, VeLO sets a new precedent for the capabilities of learned algorithms. Practically, this translates into reduced computational overhead in the training process and more robust performance across diverse and complex modeling tasks.

Theoretically, VeLO's success emphasizes the potential of meta-learning paradigms to discover superior optimization strategies, potentially revealing insights into better hand-design heuristics as well. Future research could expand on several fronts:

  • Enhanced Architectures: Improving the expressiveness and efficiency of the learned optimizer by incorporating advanced neural network designs or leveraging second-order information.
  • Task-specific Customization: Investigating mechanisms for task-specific conditioning to tailor optimization strategies dynamically.
  • Efficiency Improvements: Streamlining the meta-training process to reduce computational requirements and explore partial unroll techniques or analytic gradient methods.
  • Extended Generalization: Addressing current limitations in scaling to very large model sizes or extremely lengthy training steps by expanding the meta-training task distribution and improving the robustness of continuation strategies.

Conclusion

VeLO demonstrates the feasibility and advantages of using meta-learned optimization techniques over traditional hand-designed algorithms. By leveraging significant computational resources and a diverse training curriculum, VeLO achieves unparalleled generalization and performance across a wide array of tasks. This work not only advances the state-of-the-art in optimization but also underscores the broader potential of meta-learning to innovate and enhance various components of future machine learning pipelines.

Create an account to read this summary for free:

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.