Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Do Androids Dream of Electric Fences? Safety-Aware Reinforcement Learning with Latent Shielding (2112.11490v1)

Published 21 Dec 2021 in cs.LG, cs.LO, cs.NE, cs.SY, and eess.SY

Abstract: The growing trend of fledgling reinforcement learning systems making their way into real-world applications has been accompanied by growing concerns for their safety and robustness. In recent years, a variety of approaches have been put forward to address the challenges of safety-aware reinforcement learning; however, these methods often either require a handcrafted model of the environment to be provided beforehand, or that the environment is relatively simple and low-dimensional. We present a novel approach to safety-aware deep reinforcement learning in high-dimensional environments called latent shielding. Latent shielding leverages internal representations of the environment learnt by model-based agents to "imagine" future trajectories and avoid those deemed unsafe. We experimentally demonstrate that this approach leads to improved adherence to formally-defined safety specifications.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Chloe He (7 papers)
  2. Borja G. Leon (6 papers)
  3. Francesco Belardinelli (40 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.