Papers
Topics
Authors
Recent
2000 character limit reached

On the Prospects of Incorporating Large Language Models (LLMs) in Automated Planning and Scheduling (APS) (2401.02500v2)

Published 4 Jan 2024 in cs.AI

Abstract: Automated Planning and Scheduling is among the growing areas in AI where mention of LLMs has gained popularity. Based on a comprehensive review of 126 papers, this paper investigates eight categories based on the unique applications of LLMs in addressing various aspects of planning problems: language translation, plan generation, model construction, multi-agent planning, interactive planning, heuristics optimization, tool integration, and brain-inspired planning. For each category, we articulate the issues considered and existing gaps. A critical insight resulting from our review is that the true potential of LLMs unfolds when they are integrated with traditional symbolic planners, pointing towards a promising neuro-symbolic approach. This approach effectively combines the generative aspects of LLMs with the precision of classical planning methods. By synthesizing insights from existing literature, we underline the potential of this integration to address complex planning challenges. Our goal is to encourage the ICAPS community to recognize the complementary strengths of LLMs and symbolic planners, advocating for a direction in automated planning that leverages these synergistic capabilities to develop more advanced and intelligent planning systems.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (138)
  1. Llm-deliberation: Evaluating llms with interactive multi-agent negotiation games. arXiv preprint arXiv:2309.17234.
  2. Learning and Leveraging Verifiers to Improve Planning Capabilities of Pre-trained Language Models. arXiv preprint arXiv:2305.17077.
  3. Asai, M. 2018. Photo-Realistic Blocksworld Dataset. arXiv preprint arXiv:1812.01818.
  4. Graph of thoughts: Solving elaborate problems with large language models. arXiv preprint arXiv:2308.09687.
  5. Do as i can, not as i say: Grounding language in robotic affordances. In Conference on Robot Learning, 287–318. PMLR.
  6. A Framework to Generate Neurosymbolic PDDL-compliant Planners. arXiv preprint arXiv:2303.00438.
  7. Grounding large language models in interactive environments with online reinforcement learning. arXiv preprint arXiv:2302.02662.
  8. Learning to reason over scene graphs: a case study of finetuning GPT-2 into a robot language model for grounded task planning. Frontiers in Robotics and AI, 10.
  9. Open-vocabulary queryable scene representations for real world planning. In 2023 IEEE International Conference on Robotics and Automation (ICRA), 11509–11522. IEEE.
  10. Asking Before Action: Gather Information in Embodied Decision Making with Language Models. arXiv preprint arXiv:2305.15695.
  11. AutoTAMP: Autoregressive Task and Motion Planning with LLMs as Translators and Checkers. arXiv preprint arXiv:2306.06531.
  12. Scalable Multi-Robot Collaboration with Large Language Models: Centralized or Decentralized Systems? arXiv preprint arXiv:2309.15943.
  13. Dynamic Planning with a LLM. arXiv preprint arXiv:2308.06391.
  14. Optimal Scene Graph Planning with Large Language Model Guidance. arXiv preprint arXiv:2309.09182.
  15. Plan-Seq-Learn: Language Model Guided RL for Solving Long Horizon Robotics Tasks. In 2nd Workshop on Language and Robot Learning: Language as Grounding.
  16. Emergent Cooperation and Strategy Adaptation in Multi-Agent Systems: An Extended Coevolutionary Theory with LLMs. Electronics, 12(12): 2722.
  17. Integrating Action Knowledge and LLMs for Task Planning and Situation Handling in Open Worlds. arXiv preprint arXiv:2305.17590.
  18. Leveraging Commonsense Knowledge from Large Language Models for Task and Motion Planning. In RSS 2023 Workshop on Learning for Task and Motion Planning.
  19. Task and motion planning with large language models for object rearrangement. arXiv preprint arXiv:2303.06247.
  20. Palm-e: An embodied multimodal language model. arXiv preprint arXiv:2303.03378.
  21. From Static to Dynamic: A Continual Learning Framework for Large Language Models. arXiv preprint arXiv:2310.14248.
  22. Halo: Estimation and reduction of hallucinations in open-source weak large language models. arXiv preprint arXiv:2308.11764.
  23. Fast and Slow Planning. arXiv preprint arXiv:2303.04283.
  24. Alphazero-like tree-search can guide large language model decoding and training. arXiv preprint arXiv:2309.17179.
  25. Strategic Reasoning with Language Models. arXiv preprint arXiv:2305.19165.
  26. Openagi: When llm meets domain experts. arXiv preprint arXiv:2304.04370.
  27. HTN planning: Overview, comparison, and beyond. Artif. Intell., 222: 124–156.
  28. LPG: A Planner Based on Local Search for Planning Graphs with Action Costs. In Aips, volume 2, 281–290.
  29. Automated Planning: Theory and Practice. The Morgan Kaufmann Series in Artificial Intelligence. Amsterdam: Morgan Kaufmann. ISBN 978-1-55860-856-6.
  30. Exploring the Limitations of using Large Language Models to Fix Planning Tasks.
  31. Generating executable action plans with environmentally-aware language models. arXiv preprint arXiv:2210.04964.
  32. GG-LLM: Geometrically Grounding Large Language Models for Zero-shot Human Activity Forecasting in Human-Aware Task Planning. arXiv preprint arXiv:2310.20034.
  33. Conceptgraphs: Open-vocabulary 3d scene graphs for perception and planning. arXiv preprint arXiv:2309.16650.
  34. Leveraging Pre-trained Large Language Models to Construct and Utilize World Models for Model-based Task Planning. arXiv preprint arXiv:2305.14909.
  35. Reasoning with language model is planning with world model. arXiv preprint arXiv:2305.14992.
  36. ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings. arXiv preprint arXiv:2305.11554.
  37. SayCanPay: Heuristic Planning with Large Language Models using Learnable Domain Knowledge. arXiv preprint arXiv:2308.12682.
  38. VAL: automatic plan validation, continuous effects and mixed initiative planning using PDDL. In 16th IEEE International Conference on Tools with Artificial Intelligence, 294–301.
  39. Tool documentation enables zero-shot tool-usage with large language models. arXiv preprint arXiv:2308.00675.
  40. Enabling Efficient Interaction between an Algorithm Agent and an LLM: A Reinforcement Learning Approach. arXiv preprint arXiv:2306.03604.
  41. Chain-of-Symbol Prompting Elicits Planning in Large Langauge Models. arXiv preprint arXiv:2305.10276.
  42. Tree-Planner: Efficient Close-loop Task Planning with Large Language Models. arXiv preprint arXiv:2310.08582.
  43. Tree-of-mixed-thought: Combining fast and slow thinking for multi-hop visual reasoning. arXiv preprint arXiv:2308.09658.
  44. Language Models, Agent Models, and World Models: The LAW for Machine Reasoning and Planning. arXiv:2312.05230.
  45. Language models as zero-shot planners: Extracting actionable knowledge for embodied agents. In International Conference on Machine Learning, 9118–9147. PMLR.
  46. Voxposer: Composable 3d value maps for robotic manipulation with language models. arXiv preprint arXiv:2307.05973.
  47. Grounded decoding: Guiding text generation with grounded models for robot control. arXiv preprint arXiv:2303.00855.
  48. Inner monologue: Embodied reasoning through planning with language models. arXiv preprint arXiv:2207.05608.
  49. Neuro Symbolic Reasoning for Planning: Counterexample Guided Inductive Synthesis using Large Language Models and Satisfiability Solving. arXiv preprint arXiv:2309.16436.
  50. CoPAL: Corrective Planning of Robot Actions with Large Language Models. arXiv preprint arXiv:2310.07263.
  51. SMART-LLM: Smart Multi-Agent Robot Task Planning using Large Language Models. arXiv preprint arXiv:2309.10062.
  52. Housekeep: Tidying virtual households using commonsense reasoning. In European Conference on Computer Vision, 355–373. Springer.
  53. Kautz, H. A. 2022. The third AI summer: AAAI Robert S. Engelmore Memorial Lecture. AI Magazine, 43(1): 105–125.
  54. There and back again: extracting formal domains for controllable neurosymbolic story authoring. In Proceedings of the AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment, volume 19, 64–74.
  55. DynaCon: Dynamic Robot Planner with Contextual Awareness via LLMs. arXiv preprint arXiv:2309.16031.
  56. Exploiting Language Models as a Source of Knowledge for Cognitive Agents. arXiv preprint arXiv:2310.06846.
  57. Getting from generative ai to trustworthy ai: What llms might learn from cyc. arXiv preprint arXiv:2308.04445.
  58. Api-bank: A benchmark for tool-augmented llms. arXiv preprint arXiv:2304.08244.
  59. Human-Centered Planning. arXiv preprint arXiv:2311.04403.
  60. SwiftSage: A Generative Agent with Fast and Slow Thinking for Complex Interactive Tasks. arXiv preprint arXiv:2305.17390.
  61. Videodirectorgpt: Consistent multi-scene video generation via llm-guided planning. arXiv preprint arXiv:2309.15091.
  62. Text2motion: From natural language instructions to feasible plans. arXiv preprint arXiv:2303.12153.
  63. Llm+ p: Empowering large language models with optimal planning proficiency. arXiv preprint arXiv:2304.11477.
  64. Reflect: Summarizing robot experiences for failure explanation and correction. arXiv preprint arXiv:2306.15724.
  65. Code Models are Zero-shot Precondition Reasoners. arXiv preprint arXiv:2311.09601.
  66. Chameleon: Plug-and-play compositional reasoning with large language models. arXiv preprint arXiv:2304.09842.
  67. Neuro-symbolic causal language planning with commonsense prompting. arXiv e-prints, arXiv–2206.
  68. Multimodal Procedural Planning via Dual Text-Image Prompting. arXiv preprint arXiv:2305.01795.
  69. Reasoning on graphs: Faithful and interpretable large language model reasoning. arXiv preprint arXiv:2310.01061.
  70. Roco: Dialectic multi-robot collaboration with large language models. arXiv preprint arXiv:2307.04738.
  71. PDDL-the planning domain definition language.
  72. Evaluating Cognitive Maps and Planning in Large Language Models with CogEval. arXiv preprint arXiv:2309.15129.
  73. Diversity of Thought Improves Reasoning Abilities of Large Language Models. arXiv preprint arXiv:2310.07088.
  74. Do embodied agents dream of pixelated sheep?: Embodied decision making using language guided world modelling. arXiv preprint arXiv:2301.12050.
  75. Plansformer: Generating symbolic plans using transformers. arXiv preprint arXiv:2212.08681.
  76. Understanding the Capabilities of Large Language Models for Automated Planning. arXiv preprint arXiv:2305.16151.
  77. Plansformer Tool: Demonstrating Generation of Symbolic Plans Using Transformers. In IJCAI, volume 2023, 7158–7162. International Joint Conferences on Artificial Intelligence.
  78. Logic-lm: Empowering large language models with symbolic solvers for faithful logical reasoning. arXiv preprint arXiv:2305.12295.
  79. Human-Assisted Continual Robot Learning with Foundation Models. arXiv preprint arXiv:2309.14321.
  80. Improving Generalization in Task-oriented Dialogues with Workflows and Action Plans. arXiv preprint arXiv:2306.01729.
  81. Saynav: Grounding large language models for dynamic planning to navigation in new environments. arXiv preprint arXiv:2309.04077.
  82. Planning with large language models via corrective re-prompting. In NeurIPS 2022 Foundation Models for Decision Making Workshop.
  83. Sayplan: Grounding large language models using 3d scene graphs for scalable task planning. arXiv preprint arXiv:2307.06135.
  84. Robots that ask for help: Uncertainty alignment for large language model planners. arXiv preprint arXiv:2307.01928.
  85. Tptu: Task planning and tool usage of large language model-based ai agents. arXiv preprint arXiv:2308.03427.
  86. Artificial Intelligence, A Modern Approach. Second Edition.
  87. From Cooking Recipes to Robot Task Trees–Improving Planning Correctness and Task Efficiency by Leveraging LLMs with a Knowledge Network. arXiv preprint arXiv:2309.09181.
  88. Evaluation of Pretrained Large Language Models in Embodied Planning Tasks. In International Conference on Artificial General Intelligence, 222–232. Springer.
  89. RoboVQA: Multimodal Long-Horizon Reasoning for Robotics. arXiv preprint arXiv:2311.00899.
  90. Navigation with large language models: Semantic guesswork as a heuristic for planning. arXiv preprint arXiv:2310.10103.
  91. Hugginggpt: Solving ai tasks with chatgpt and its friends in huggingface. arXiv preprint arXiv:2303.17580.
  92. Vision-Language Interpreter for Robot Task Planning. arXiv preprint arXiv:2311.00967.
  93. Generalized Planning in PDDL Domains with Pretrained Large Language Models. arXiv preprint arXiv:2305.11014.
  94. PDDL planning with pretrained large language models. In NeurIPS 2022 Foundation Models for Decision Making Workshop.
  95. ProgPrompt: program generation for situated robot task planning using large language models. Autonomous Robots, 1–14.
  96. Llm-planner: Few-shot grounded planning for embodied agents with large language models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2998–3009.
  97. Cognitive architectures for language agents. arXiv preprint arXiv:2309.02427.
  98. AdaPlanner: Adaptive Planning from Feedback with Language Models. arXiv preprint arXiv:2305.16653.
  99. Large Language Models are In-Context Semantic Reasoners rather than Symbolic Reasoners. arXiv preprint arXiv:2305.14825.
  100. Can Large Language Models Really Improve by Self-critiquing Their Own Plans? arXiv preprint arXiv:2310.08118.
  101. PlanBench: An Extensible Benchmark for Evaluating Large Language Models on Planning and Reasoning about Change. In Thirty-seventh Conference on Neural Information Processing Systems Datasets and Benchmarks Track.
  102. Large Language Models Still Can’t Plan (A Benchmark for LLMs on Planning and Reasoning about Change). arXiv preprint arXiv:2206.10498.
  103. On the planning abilities of large language models (a critical investigation with a proposed benchmark). arXiv preprint arXiv:2302.06706.
  104. Conformal Temporal Logic Planning using Large Language Models: Knowing When to Do What and When to Ask for Help. arXiv preprint arXiv:2309.10092.
  105. Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models. arXiv preprint arXiv:2305.04091.
  106. Guiding language model reasoning with planning tokens. arXiv preprint arXiv:2310.05707.
  107. Describe, explain, plan and select: Interactive planning with large language models enables open-world multi-task agents. arXiv preprint arXiv:2302.01560.
  108. A Prefrontal Cortex-inspired Architecture for Planning in Large Language Models. arXiv preprint arXiv:2310.00194.
  109. Unleashing the Power of Graph Learning through LLM-based Autonomous Agents. arXiv preprint arXiv:2309.04565.
  110. From Word Models to World Models: Translating from Natural Language to the Probabilistic Language of Thought. arXiv preprint arXiv:2306.12672.
  111. Plan, Eliminate, and Track–Language Models are Good Teachers for Embodied Agents. arXiv preprint arXiv:2305.02412.
  112. Integrating Common Sense and Planning with Large Language Models for Room Tidying. In RSS 2023 Workshop on Learning for Task and Motion Planning.
  113. Embodied task planning with large language models. arXiv preprint arXiv:2307.01848.
  114. Translating natural language to planning goals with large-language models. arXiv preprint arXiv:2302.05128.
  115. Gentopia: A collaborative platform for tool-augmented llms. arXiv preprint arXiv:2308.04030.
  116. Xu, L. 1995. Case based reasoning. IEEE Potentials, 13(5): 10–13.
  117. Creative Robot Tool Use with Large Language Models. arXiv preprint arXiv:2310.13065.
  118. LLM-Grounder: Open-Vocabulary 3D Visual Grounding with Large Language Model as an Agent. arXiv preprint arXiv:2309.12311.
  119. OceanChat: Piloting Autonomous Underwater Vehicles in Natural Language. arXiv preprint arXiv:2309.16052.
  120. Learning Automata-Based Task Knowledge Representation from Large-Scale Generative Language Models. arXiv preprint arXiv:2212.01944.
  121. On the Planning, Search, and Memorization Capabilities of Large Language Models. arXiv preprint arXiv:2309.01868.
  122. Yang, Z. 2023. Neuro-Symbolic AI Approaches to Enhance Deep Neural Networks with Logical Reasoning and Knowledge Integration. Ph.D. thesis, Arizona State University.
  123. Coupling Large Language Models with Logic Programming for Robust and General Reasoning from Text. arXiv preprint arXiv:2307.07696.
  124. Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601.
  125. Statler: State-maintaining language models for embodied reasoning. arXiv preprint arXiv:2306.17840.
  126. EIPE-text: Evaluation-Guided Iterative Plan Extraction for Long-Form Narrative Text Generation. arXiv preprint arXiv:2310.08185.
  127. Plan4mc: Skill reinforcement learning and planning for open-world minecraft tasks. arXiv preprint arXiv:2303.16563.
  128. Distilling Script Knowledge from Large Language Models for Constrained Language Planning. arXiv preprint arXiv:2305.05252.
  129. Parsel: Algorithmic Reasoning with Language Models by Composing Decompositions. In Thirty-seventh Conference on Neural Information Processing Systems.
  130. Large language models as zero-shot human models for human-robot interaction. arXiv preprint arXiv:2303.03548.
  131. Prefer: Prompt ensemble learning via feedback-reflect-refine. arXiv preprint arXiv:2308.12033.
  132. Planning with Logical Graph-based Language Model for Instruction Generation. arXiv:2308.13782.
  133. Building cooperative embodied agents modularly with large language models. arXiv preprint arXiv:2307.02485.
  134. Bootstrap your own skills: Learning to solve new tasks with large language model guidance. arXiv preprint arXiv:2310.10021.
  135. Large Language Models as Commonsense Knowledge for Large-Scale Task Planning. arXiv preprint arXiv:2305.14078.
  136. Judging LLM-as-a-judge with MT-Bench and Chatbot Arena. arXiv:2306.05685.
  137. Steve-Eye: Equipping LLM-based Embodied Agents with Visual Perception in Open Worlds. arXiv preprint arXiv:2310.13255.
  138. ISR-LLM: Iterative Self-Refined Large Language Model for Long-Horizon Sequential Task Planning. arXiv preprint arXiv:2308.13724.
Citations (22)

Summary

  • The paper demonstrates eight primary categories of LLM applications in APS by reviewing 126 papers and highlighting language translation, plan generation, and neurosymbolic integration.
  • It explains the potential and limitations of LLMs in key APS tasks such as model construction, interactive planning, and heuristics optimization.
  • The study emphasizes future research directions, advocating for improved training paradigms and effective integration with symbolic planners.

Prospects of Incorporating LLMs in Automated Planning and Scheduling (APS)

Introduction

The intersection of LLMs and Automated Planning and Scheduling (APS) represents an evolutionary step in AI. APS is fundamentally concerned with algorithms that generate action sequences to achieve specified objectives in diverse domains. The recent surge in LLM capabilities extends their application into APS, facilitated through their strength in language translation, plan generation, model construction, and more. A comprehensive review of 126 papers reveals eight primary categories of LLM applications in APS. These categories expose the promising integration of LLMs with APS while identifying significant functional gaps.

LLMs: A Background

LLMs, typified by architectures such as transformers, have grown to encompass complex tasks in NLP, from word prediction to advanced contextual reasoning. Their evolution is marked by the success of architectures that incorporate self-attention mechanisms, enabling models like GPT-4 to excel across areas including writing, reasoning, and coding. Figure 1

Figure 1: Radar chart showcasing the relative performance of six LLMs (GPT-4, Claude-v1, GPT-3.5-turbo, Vicuna-13B, Alpaca-13B, LLama-13B) across key domains.

Automated Planning and Scheduling

Automated Planning and Scheduling focuses on devising sequences of actions to reach defined goals within complex environments. Classical planning problems (CPPs) within APS utilize symbolic representations, traditionally supported by rigorous languages like PDDL. The flexibility offered by LLMs provides an innovative dimension, allowing planning tasks to be expressed in more intuitive natural language constructs.

Literature Review: LLMs Applied in APS

The literature review, surveying 126 papers, organizes applications of LLMs in APS into eight distinct categories: Language Translation, Plan Generation, Model Construction, Multi-agent Planning, Interactive Planning, Heuristics Optimization, Tool Integration, and Brain-Inspired Planning. This classification illustrates the versatility of LLMs while also highlighting their roles and limitations within each category. Figure 2

Figure 2: Chart illustrating the distribution of surveyed papers across various conferences.

Key Findings and Categories in APS

Language Translation

LLMs bridge natural language and structured planning languages in APS. Notable implementations include transforming descriptive tasks into PDDL formats, thereby optimizing human-machine interactions. However, the autonomous translation of LLMs without expert mediation presents a research gap.

Plan Generation

Although LLMs demonstrate capacity in generating plans through causal LLMs, challenges in achieving optimal and comprehensive results persist. Limited adaptability to non-trained datasets underscores this challenge.

Model Construction

LLMs facilitate developing abstract and detailed world models necessary for planning. Despite their strengths in high-level concepts, they falter in handling minute spatial details, limiting their potential in comprehensive model construction.

Multi-agent Planning

LLMs assist in dynamic, multi-agent environments, enhancing strategic interactions among agents. Yet, standardizing inter-agent communications remains a significant challenge.

Interactive Planning

Interactive scenarios benefit from LLMs' dynamic adaptability facilitated through varied feedback mechanisms. However, integrating fast neural processes with symbolic reasoning is a nuanced challenge that remains to be addressed.

Heuristics Optimization

Here, LLMs contribute to refining planning processes, improving efficiency via heuristic guidance. Full potential can be realized through stronger integration with symbolic planners.

Tool Integration

LLMs serve as integrators among diverse planning tools but face challenges related to tool reliance and operational scalability, notably hallucinating nonexistent tools.

Brain-Inspired Planning

Investigating neurosymbolic methodologies, LLMs mimic cognitive processes but often lack-depth understanding, emphasizing the need for holistic integration of neural-symbolic structures. Figure 3

Figure 3: Annual distribution of surveyed papers demonstrating the growth in LLM research.

Conclusion

The integration of LLMs into APS continues to unravel challenges and opportunities, with notable progress demonstrated across various applications, yet they are not without limitations. Future research should aim to strengthen LLM training paradigms for enhanced coherence, leverage neuro-symbolic frameworks, and establish robust performance metrics for LLM-assisted planners. Advanced integration of LLMs with APS, enhancing both human-like interaction and precise analytical planning, holds considerable promise for sophisticated AI planning systems. Figure 4

Figure 4: Word cloud indicating prevalent themes from the paper titles, emphasizing the emergence of neuro-symbolic trends.

Whiteboard

Paper to Video (Beta)

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.