Learning Planning Abstractions from Language (2405.03864v1)
Abstract: This paper presents a framework for learning state and action abstractions in sequential decision-making domains. Our framework, planning abstraction from language (PARL), utilizes language-annotated demonstrations to automatically discover a symbolic and abstract action space and induce a latent state abstraction based on it. PARL consists of three stages: 1) recovering object-level and action concepts, 2) learning state abstractions, abstract action feasibility, and transition models, and 3) applying low-level policies for abstract actions. During inference, given the task description, PARL first makes abstract action plans using the latent transition and feasibility functions, then refines the high-level plan using low-level policies. PARL generalizes across scenarios involving novel object instances and environments, unseen concept compositions, and tasks that require longer planning horizons than settings it is trained on.
- TAPS: Task-agnostic policy sequencing. In ICRA, 2023.
- Alignment-based compositional semantics for instruction following. In EMNLP, 2015.
- Modular multitask reinforcement learning with policy sketches. In ICML, 2017.
- Hindsight Experience Replay. In NeurIPS, 2017.
- Learning Neural-Symbolic Descriptive Planning Models via Cube-Space Priors: The Voyage Home (To STRIPS). arXiv:2004.12850, 2020.
- Recent advances in hierarchical reinforcement learning. Discrete event dynamic systems, 13(1):41–77, 2003.
- Learning First-Order Symbolic Representations for Planning from the Structure of the State Space. In ECAI, 2020.
- Learning and planning for temporally extended tasks in unknown environments, 2021.
- Decision Transformer: Reinforcement Learning via Sequence Modeling. In NeurIPS, 2021.
- BabyAI: First Steps Towards Grounded Language Learning With a Human In the Loop. In ICLR, 2019.
- Recurrent Environment Simulators. In ICLR, 2017.
- Learning Neuro-Symbolic Relational Transition Models for Bilevel Planning. arXiv:2105.14074, 2021.
- Modular networks for compositional instruction following. In NAACL-HLT, pp. 1033–1040, 2021.
- Pybullet, a python module for physics simulation in robotics, games and machine learning, 2017.
- Tomás de la Rosa and Sheila McIlraith. Learning domain control knowledge for tlplan and beyond. In ICAPS 2011 Workshop on Planning and Learning, 2011.
- Thomas G Dietterich. Hierarchical reinforcement learning with the maxq value function decomposition. JAIR, 13:227–303, 2000.
- Automated Planning: Theory and Practice. Elsevier, 2004.
- A Theory of Abstraction. Artif. Intell., 57(2-3):323–389, 1992.
- Equivalence Notions and Model Minimization in Markov Decision Processes. Artif. Intell., 147(1-2):163–223, 2003.
- Pct: Point cloud transformer. Computational Visual Media, 7:187–199, 2021.
- Latent space planning for multi-object manipulation with environment-aware relational classifiers. arXiv preprint arXiv:2305.10857, 2023.
- Learning Grounded Relational Symbols from Continuous Data for Abstract Reasoning. In ICRA Workshop, 2013.
- Language as an abstraction for hierarchical deep reinforcement learning. Advances in Neural Information Processing Systems, 32, 2019.
- Leslie Pack Kaelbling. Hierarchical Learning in Stochastic Domains: Preliminary Results. In ICML, 1993.
- Language-driven representation learning for robotics. In RSS, 2023.
- From Skills to Symbols: Learning Symbolic Representations for Abstract High-Level Planning. JAIR, 61:215–289, 2018.
- Towards a Unified Theory of State Abstraction for MDPs. In AI&M, 2006.
- Structformer: Learning spatial structure for language-guided semantic rearrangement of novel objects. In International Conference on Robotics and Automation, 2022.
- A survey of reinforcement learning informed by natural language. In IJCAI, 2019.
- Learning rational subgoals from demonstrations and instructions. In Proceedings of the AAAI Conference on Artificial Intelligence, 2023.
- Liv: Language-image representations and rewards for robotic control. In ICML, 2023.
- Pdsketch: Integrated domain programming, learning, and planning. Advances in Neural Information Processing Systems, 35:36972–36984, 2022.
- What matters in language conditioned robotic imitation learning over unstructured data. IEEE Robotics and Automation Letters, 7(4):11205–11212, 2022.
- Neville Mehta. Hierarchical structure discovery and transfer in sequential decision problems. Oregon State University, 2011.
- Listen, attend, and walk: Neural mapping of navigational instructions to action sequences. In Proceedings of the AAAI Conference on Artificial Intelligence, 2016.
- Mapping instructions and visual observations to actions with reinforcement learning. arXiv preprint arXiv:1704.08795, 2017.
- Asynchronous methods for deep reinforcement learning. In International conference on machine learning, 2016.
- Learning language-conditioned robot behavior from offline data and crowd-sourced annotation. In Conference on Robot Learning, pp. 1303–1315. PMLR, 2022.
- Learning macro-actions for arbitrary planners and domains. In ICAPS, 2007.
- Inferring task goals and constraints using bayesian nonparametric inverse reinforcement learning. In JMLR, 2020.
- Learning Symbolic Models of Stochastic Domains. JAIR, 29:309–352, 2007.
- FiLM: Visual Reasoning with a General Conditioning Layer. In AAAI, 2018.
- Learning transferable visual models from natural language supervision. In ICML, 2021.
- Aggregation and Disaggregation Techniques and Methodology in Optimization. Operations Research, 39(4):553–582, 1991.
- Object scene representation transformer. In Advances in Neural Information Processing Systems, 2022.
- Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model. Nat., 588(7839):604–609, 2020.
- Value function spaces: Skill-centric state abstractions for long-horizon reasoning. In ICLR, 2022.
- Skill induction and planning with latent language. arXiv preprint arXiv:2110.01517, 2021.
- Learning Symbolic Operators for Task and Motion Planning. In IROS, 2021.
- Program guided agent. In ICLR, 2020.
- Between MDPs and semi-MDPs: A Framework for Temporal Abstraction in Reinforcement Learning. Artif. Intell., 112(1-2):181–211, 1999.
- Understanding natural language commands for robotic navigation and mobile manipulation. In AAAI, 2011.
- Robots that use language. Annual Review of Control, Robotics, and Autonomous Systems, 3:25–55, 2020.
- Teaching multiple tasks to an rl agent using ltl. In AAMAS, 2018.
- Generalizable task planning through representation pretraining. IEEE Robotics and Automation Letters, 7(3):8299–8306, 2022.
- Programmatically Grounded, Compositionally Generalizable Robotic Manipulation. In ICLR, 2023.
- Language-mediated, object-centric representation learning. In ACL Findings, 2021.
- Regression planning networks. Advances in Neural Information Processing Systems, 32, 2019.
- Sequence-Based Plan Feasibility Prediction for Efficient Task and Motion Planning. In RSS, 2023.
- Piglet: Language grounding through neuro-symbolic interaction in a 3d world. In ACL, 2021.
- Learning Invariant Representations for Reinforcement Learning without Reconstruction. In ICLR, 2021.
- Glipv2: Unifying localization and vision-language understanding. In Advances in Neural Information Processing Systems, 2022.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.