Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Language Decision Transformers with Exponential Tilt for Interactive Text Environments (2302.05507v2)

Published 10 Feb 2023 in cs.CL, cs.AI, and cs.LG

Abstract: Text-based game environments are challenging because agents must deal with long sequences of text, execute compositional actions using text and learn from sparse rewards. We address these challenges by proposing Language Decision Transformers (LDTs), a framework that is based on transformer LLMs and decision transformers (DTs). Our LDTs extend DTs with 3 components: (1) exponential tilt to guide the agent towards high obtainable goals, (2) novel goal conditioning methods yielding better results than the traditional return-to-go (sum of all future rewards), and (3) a model of future observations that improves agent performance. LDTs are the first to address offline RL with DTs on these challenging games. Our experiments show that LDTs achieve the highest scores among many different types of agents on some of the most challenging Jericho games, such as Enchanter.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Nicolas Gontier (8 papers)
  2. Issam Laradji (37 papers)
  3. David Vazquez (73 papers)
  4. Christopher Pal (97 papers)
  5. Pau Rodriguez (35 papers)
Citations (1)

Summary

We haven't generated a summary for this paper yet.