Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Offline Reinforcement Learning Hands-On (2011.14379v1)

Published 29 Nov 2020 in cs.LG

Abstract: Offline Reinforcement Learning (RL) aims to turn large datasets into powerful decision-making engines without any online interactions with the environment. This great promise has motivated a large amount of research that hopes to replicate the success RL has experienced in simulation settings. This work ambitions to reflect upon these efforts from a practitioner viewpoint. We start by discussing the dataset properties that we hypothesise can characterise the type of offline methods that will be the most successful. We then verify these claims through a set of experiments and designed datasets generated from environments with both discrete and continuous action spaces. We experimentally validate that diversity and high-return examples in the data are crucial to the success of offline RL and show that behavioural cloning remains a strong contender compared to its contemporaries. Overall, this work stands as a tutorial to help people build their intuition on today's offline RL methods and their applicability.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Louis Monier (1 paper)
  2. Jakub Kmec (4 papers)
  3. Alexandre Laterre (13 papers)
  4. Thomas Pierrot (21 papers)
  5. Valentin Courgeau (4 papers)
  6. Olivier Sigaud (56 papers)
  7. Karim Beguir (11 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.