Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 60 tok/s
Gemini 2.5 Pro 51 tok/s Pro
GPT-5 Medium 18 tok/s Pro
GPT-5 High 14 tok/s Pro
GPT-4o 77 tok/s Pro
Kimi K2 159 tok/s Pro
GPT OSS 120B 456 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

A Data Efficient Framework for Learning Local Heuristics (2404.06728v2)

Published 10 Apr 2024 in cs.RO

Abstract: With the advent of machine learning, there have been several recent attempts to learn effective and generalizable heuristics. Local Heuristic A* (LoHA*) is one recent method that instead of learning the entire heuristic estimate, learns a "local" residual heuristic that estimates the cost to escape a region (Veerapaneni et al 2023). LoHA*, like other supervised learning methods, collects a dataset of target values by querying an oracle on many planning problems (in this case, local planning problems). This data collection process can become slow as the size of the local region increases or if the domain requires expensive collision checks. Our main insight is that when an A* search solves a start-goal planning problem it inherently ends up solving multiple local planning problems. We exploit this observation to propose an efficient data collection framework that does <1/10th the amount of work (measured by expansions) to collect the same amount of data in comparison to baselines. This idea also enables us to run LoHA* in an online manner where we can iteratively collect data and improve our model while solving relevant start-goal tasks. We demonstrate the performance of our data collection and online framework on a 4D $(x, y, \theta, v)$ navigation domain.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (11)
  1. Multi-Heuristic A. In Fox, D.; Kavraki, L. E.; and Kurniawati, H., eds., Robotics: Science and Systems X, University of California, Berkeley, USA, July 12-16, 2014.
  2. Hindsight Experience Replay. In Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, December 4-9, 2017, Long Beach, CA, USA, 5048–5058.
  3. Learning Heuristic Search via Imitation. CoRR, abs/1707.03034.
  4. The Closed List is an Obstacle Too. In Ma, H.; and Serina, I., eds., Proceedings of the Fourteenth International Symposium on Combinatorial Search, SOCS 2021, Virtual Conference [Jinan, China], July 26-30, 2021, 121–125. AAAI Press.
  5. Learning heuristic functions for large state spaces. Artificial Intelligence, 175(16): 2075–2098.
  6. Speeding Up Search-Based Motion Planning using Expansion Delay Heuristics. Proceedings of the International Conference on Automated Planning and Scheduling, 31(1): 528–532.
  7. Learning Heuristic A: Efficient Graph Search using Neural Network. In 2020 IEEE International Conference on Robotics and Automation (ICRA), 9542–9547.
  8. Korf, R. E. 1990. Real-time heuristic search. Artificial Intelligence, 42(2): 189–211.
  9. Studies in Semi-Admissible Heuristics. IEEE Transactions on Pattern Analysis and Machine Intelligence, PAMI-4(4): 392–399.
  10. Learning Heuristic Functions for Mobile Robot Path Planning Using Deep Neural Networks. Proceedings of the International Conference on Automated Planning and Scheduling, 29(1): 764–772.
  11. Learning Local Heuristics for Search-Based Navigation Planning. In Proceedings of the Thirty-Third International Conference on Automated Planning and Scheduling, July 8-13, 2023, Prague, Czech Republic, 634–638. AAAI Press.

Summary

We haven't generated a summary for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Lightbulb On Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.