Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 161 tok/s
Gemini 2.5 Pro 42 tok/s Pro
GPT-5 Medium 30 tok/s Pro
GPT-5 High 32 tok/s Pro
GPT-4o 31 tok/s Pro
Kimi K2 192 tok/s Pro
GPT OSS 120B 435 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

Leveraging Environment Interaction for Automated PDDL Translation and Planning with Large Language Models (2407.12979v2)

Published 17 Jul 2024 in cs.LG

Abstract: LLMs have shown remarkable performance in various natural language tasks, but they often struggle with planning problems that require structured reasoning. To address this limitation, the conversion of planning problems into the Planning Domain Definition Language (PDDL) has been proposed as a potential solution, enabling the use of automated planners. However, generating accurate PDDL files typically demands human inputs or correction, which can be time-consuming and costly. In this paper, we propose a novel approach that leverages LLMs and environment feedback to automatically generate PDDL domain and problem description files without the need for human intervention. Our method introduces an iterative refinement process that generates multiple problem PDDL candidates and progressively refines the domain PDDL based on feedback obtained from interacting with the environment. To guide the refinement process, we develop an Exploration Walk (EW) metric, which provides rich feedback signals for LLMs to update the PDDL file. We evaluate our approach on $10$ PDDL environments. We achieve an average task solve rate of 66% compared to a 29% solve rate by GPT-4's intrinsic planning with chain-of-thought prompting. Our work enables the automated modeling of planning environments using LLMs and environment feedback, eliminating the need for human intervention in the PDDL translation process and paving the way for more reliable LLM agents in challenging problems. Our code is available at https://github.com/BorealisAI/LLM-pddl-planning

Citations (1)

Summary

  • The paper introduces an automated method combining LLMs with iterative environment feedback to generate and refine PDDL files without human input.
  • It employs an innovative Exploration Walk metric for in-context learning, demonstrating a task solve rate of 66% versus GPT-4’s 29%.
  • The approach enhances classical planning by integrating autonomous domain modeling with LLM iterations, paving the way for more robust planning systems.

Leveraging Environment Interaction for Automated PDDL Translation and Planning with LLMs

LLMs, renowned for their prowess in natural language processing tasks, confront notable challenges when navigating structured reasoning inherent in planning problems. The transformation of planning challenges into the Planning Domain Definition Language (PDDL) offers potential resolution, facilitating the activation of automated planners. This paper introduces a novel methodology employing LLMs and environment feedback to autonomously generate PDDL domain and problem description files void of human input.

Methodological Approach

The solution revolves around an iterative refinement protocol, generating several PDDL candidates and incrementally perfecting the domain PDDL through environmental feedback assimilation. Central to this refinement is the Exploration Walk (EW) metric, which supplies substantial feedback signals, enabling LLMs to recalibrate the PDDL file effectively. Validation of the approach across 10 PDDL environments revealed a significant improvement in task solve rates, averaging 66%, contrasting with GPT-4's 29% via intrinsic planning. Figure 1

Figure 1: Overview of our method. Right: The process begins with natural language descriptions translated into problem PDDL by the LLM (red arrows). Then a domain is generated and refined through iterative cycles involving exploration walks in the environment, interaction with a classical planner, and feedback from the LLM (blue/black arrows). Left: The iterative refinement process depicted on the right corresponds to single paths in the structures shown on the left. Each node represents a state in the refinement process, with arrows indicating problem translation (red), domain refinement (blue).

Implementation Details

The primary objective was to automate the generation of PDDL domain and problem files using LLMs without human intervention. Integral to this task is the assumption of access to specific environment components, such as the list of objects and the action interfaces, along with the executability and verifiability of actions. The approach is exemplified by generating a PDDL domain via hypothesized “mental models,” iteratively updated through discrepancies observed between action feasibility within these models and actual environmental conditions.

Exploration Walk Metric

One of the novel innovations—Exploration Walk (EW)—materializes as a smooth feedback signal, offering granular feedback to LLMs for in-context learning progression. The EW metric is constructed by analyzing the executability of random action sequences within LLM-generated environments compared to those in the true environment. This metric not only ensures alignment but also facilitates a consistent evaluation framework throughout the refinement process. Figure 2

Figure 2: Correlation between average exploration walk score and average domain difference.

Experimental Results

When evaluated against Intrinsic Planning baselines (both with and without chain-of-thought prompting), the proposed automated method manifested superior performance across multiple environments. With task solve rates substantially surpassing baselines, it demonstrates the efficacy of environment interaction-based feedback and iterative refinement using LLMs. The correlation between EW scores and task solve rates further corroborates the method’s robustness and potential for precise domain generation. Figure 3

Figure 3: Histogram of average number of lines of domains.

Conclusion

This work fundamentally advances the integration of LLMs with classical planning by eliminating human intervention from PDDL generation processes, thus promoting the autonomous modeling of planning environments. The implications of this paper not only enhance AI’s capacity for complex decision-making tasks but also pave the way for future innovations in automatic model formulation in multi-step planning scenarios. Continued exploration in refining EW strategies and extending the application to non-PDDL environments can significantly augment digital or physical agents' capabilities.

Through methodical engagement with exploration walks and architectural independence from continuous human oversight in iterative domain refinement, this research redefines computational interactions within the field of automated planning and large-scale linguistic models.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.