Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Target Network and Truncation Overcome The Deadly Triad in $Q$-Learning (2203.02628v2)

Published 5 Mar 2022 in cs.LG, math.OC, and stat.ML

Abstract: $Q$-learning with function approximation is one of the most empirically successful while theoretically mysterious reinforcement learning (RL) algorithms, and was identified in Sutton (1999) as one of the most important theoretical open problems in the RL community. Even in the basic linear function approximation setting, there are well-known divergent examples. In this work, we show that \textit{target network} and \textit{truncation} together are enough to provably stabilize $Q$-learning with linear function approximation, and we establish the finite-sample guarantees. The result implies an $O(\epsilon{-2})$ sample complexity up to a function approximation error. Moreover, our results do not require strong assumptions or modifying the problem parameters as in existing literature.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Zaiwei Chen (21 papers)
  2. John Paul Clarke (1 paper)
  3. Siva Theja Maguluri (53 papers)
Citations (17)

Summary

We haven't generated a summary for this paper yet.