Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Meta-learning of Sequential Strategies (1905.03030v2)

Published 8 May 2019 in cs.LG, cs.AI, and stat.ML

Abstract: In this report we review memory-based meta-learning as a tool for building sample-efficient strategies that learn from past experience to adapt to any task within a target class. Our goal is to equip the reader with the conceptual foundations of this tool for building new, scalable agents that operate on broad domains. To do so, we present basic algorithmic templates for building near-optimal predictors and reinforcement learners which behave as if they had a probabilistic model that allowed them to efficiently exploit task structure. Furthermore, we recast memory-based meta-learning within a Bayesian framework, showing that the meta-learned strategies are near-optimal because they amortize Bayes-filtered data, where the adaptation is implemented in the memory dynamics as a state-machine of sufficient statistics. Essentially, memory-based meta-learning translates the hard problem of probabilistic sequential inference into a regression problem.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (24)
  1. Pedro A. Ortega (34 papers)
  2. Jane X. Wang (21 papers)
  3. Mark Rowland (57 papers)
  4. Tim Genewein (25 papers)
  5. Zeb Kurth-Nelson (9 papers)
  6. Razvan Pascanu (138 papers)
  7. Nicolas Heess (139 papers)
  8. Joel Veness (29 papers)
  9. Alex Pritzel (1 paper)
  10. Pablo Sprechmann (25 papers)
  11. Siddhant M. Jayakumar (13 papers)
  12. Tom McGrath (6 papers)
  13. Kevin Miller (17 papers)
  14. Mohammad Azar (3 papers)
  15. Ian Osband (34 papers)
  16. Neil Rabinowitz (7 papers)
  17. Silvia Chiappa (26 papers)
  18. Simon Osindero (45 papers)
  19. Yee Whye Teh (162 papers)
  20. Hado van Hasselt (57 papers)
Citations (90)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com