Safe Learning of PDDL Domains with Conditional Effects -- Extended Version (2403.15251v1)
Abstract: Powerful domain-independent planners have been developed to solve various types of planning problems. These planners often require a model of the acting agent's actions, given in some planning domain description language. Manually designing such an action model is a notoriously challenging task. An alternative is to automatically learn action models from observation. Such an action model is called safe if every plan created with it is consistent with the real, unknown action model. Algorithms for learning such safe action models exist, yet they cannot handle domains with conditional or universal effects, which are common constructs in many planning problems. We prove that learning non-trivial safe action models with conditional effects may require an exponential number of samples. Then, we identify reasonable assumptions under which such learning is tractable and propose SAM Learning of Conditional Effects (Conditional-SAM), the first algorithm capable of doing so. We analyze Conditional-SAM theoretically and evaluate it experimentally. Our results show that the action models learned by Conditional-SAM can be used to solve perfectly most of the test set problems in most of the experimented domains.
- Learning action models with minimal observability. Artificial Intelligence, 275: 104–137.
- Learning STRIPS action models with classical planning. In International Conference on Automated Planning and Scheduling (ICAPS).
- Learning partially observable deterministic action models. Journal of Artificial Intelligence Research, 33: 349–402.
- Bacchus, F. 2001. AIPS 2000 planning competition: The fifth international conference on artificial intelligence planning and scheduling systems. Ai magazine, 22(3): 47–47.
- Generalised domain model acquisition from action traces. In International Conference on Automated Planning and Scheduling (ICAPS), 42–49.
- Acquiring planning domain models using LOCM. The Knowledge Engineering Review, 28(2): 195–213.
- STRIPS: A new approach to the application of theorem proving to problem solving. Artificial Intelligence, 2(3-4): 189–208.
- PDDL – The Planning Domain Definition Language. Technical Report, Tech. Rep.
- Automated planning and acting. Cambridge University Press.
- Helmert, M. 2006. The Fast Downward planning system. Journal of Artificial Intelligence Research, 26: 191–246.
- Hoffmann, J. 2001. FF: The fast-forward planning system. AI magazine, 22(3): 57–57.
- VAL: Automatic plan validation, continuous effects and mixed initiative planning using PDDL. In 16th IEEE International Conference on Tools with Artificial Intelligence, 294–301. IEEE.
- Safe Learning of Lifted Action Models. In International Conference on Principles of Knowledge Representation and Reasoning (KR), 379–389.
- Learning Probably Approximately Complete and Safe Action Models for Stochastic Worlds. In AAAI Conference on Artificial Intelligence.
- The 3rd international planning competition: Results and analysis. Journal of Artificial Intelligence Research, 20: 1–59.
- Learning Safe Numeric Action Models. In AAAI Conference on Artificial Intelligence (AAAI).
- Learning planning operators with conditional and probabilistic effects. In Proceedings of the AAAI Spring Symposium on Planning with Incomplete Information for Robot Problems, 86–94.
- Pednault, E. P. 1989. Adl: Exploringthe middle ground between strips and the situation calculus. In Proceedings of the First International Conference on Principles of Knowledge Representation and Reasoning (KR’89), 324–332.
- Interval-based relaxation for general numeric planning. In European Conference on Artificial Intelligence (ECAI), 655–663.
- PDDL Generators. https://doi.org/10.5281/zenodo.6382173.
- Efficient, Safe, and Probably Approximately Complete Learning of Action Models. In International Joint Conference on Artificial Intelligence (IJCAI), 4405–4411.
- The 2014 international planning competition: Progress and trends. AI Magazine, 36(3): 90–98.
- Safe Reinforcement Learning in Constrained Markov Decision Processes. In International Conference on Machine Learning (ICML), 9797–9806.
- Learning action models from plan examples using weighted MAX-SAT. Artificial Intelligence, 171(2-3): 107–143.
- Learning complex action models with quantifiers and logical implications. Artificial Intelligence, 174(18): 1540–1569.