Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
126 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Versatile Navigation under Partial Observability via Value-guided Diffusion Policy (2404.02176v1)

Published 1 Apr 2024 in cs.RO and cs.AI

Abstract: Route planning for navigation under partial observability plays a crucial role in modern robotics and autonomous driving. Existing route planning approaches can be categorized into two main classes: traditional autoregressive and diffusion-based methods. The former often fails due to its myopic nature, while the latter either assumes full observability or struggles to adapt to unfamiliar scenarios, due to strong couplings with behavior cloning from experts. To address these deficiencies, we propose a versatile diffusion-based approach for both 2D and 3D route planning under partial observability. Specifically, our value-guided diffusion policy first generates plans to predict actions across various timesteps, providing ample foresight to the planning. It then employs a differentiable planner with state estimations to derive a value function, directing the agent's exploration and goal-seeking behaviors without seeking experts while explicitly addressing partial observability. During inference, our policy is further enhanced by a best-plan-selection strategy, substantially boosting the planning success rate. Moreover, we propose projecting point clouds, derived from RGB-D inputs, onto 2D grid-based bird-eye-view maps via semantic segmentation, generalizing to 3D environments. This simple yet effective adaption enables zero-shot transfer from 2D-trained policy to 3D, cutting across the laborious training for 3D policy, and thus certifying our versatility. Experimental results demonstrate our superior performance, particularly in navigating situations beyond expert demonstrations, surpassing state-of-the-art autoregressive and diffusion-based baselines for both 2D and 3D scenarios.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (36)
  1. Is conditional generative modeling all you need for decision making? In ICLR, 2023.
  2. Offline reinforcement learning via high-fidelity generative behavior modeling. In ICLR, 2023.
  3. Analog bits: Generating discrete data using diffusion models with self-conditioning. arXiv:2208.04202, 2022.
  4. Diffusion policy: Visuomotor policy learning via action diffusion. In RSS, 2023.
  5. Pointvector: A vector representation in point cloud analysis. In CVPR, 2023.
  6. Diffusion models beat gans on image synthesis. In NeurIPS, 2021.
  7. Implicit generation and generalization in energy-based models. arXiv:1903.08689, 2019.
  8. Compositional visual generation with energy based models. In NeurIPS, 2020.
  9. Denoising diffusion probabilistic models. In NeurIPS, 2020.
  10. Estimation of non-normalized statistical models by score matching. JMLR, 6(4):695–709, 2005.
  11. Towards real-world navigation with deep differentiable planners. In CVPR, 2022.
  12. Planning with diffusion for flexible behavior synthesis. In ICML, 2022.
  13. Planning and acting in partially observable stochastic domains. Artificial intelligence, 101(1-2):99–134, 1998.
  14. Qmdp-net: Deep learning for planning under partial observability. In NeurIPS, 2017.
  15. Elucidating the design space of diffusion-based generative models. In NeurIPS, 2022.
  16. Offline reinforcement learning: Tutorial, review, and perspectives on open problems. arXiv:2005.01643, 2020.
  17. Adaptdiffuser: Diffusion models as adaptive self-evolving planners. In ICML, 2023.
  18. Learning policies for partially observable environments: Scaling up. In ICML, 1995.
  19. Walk these ways: Tuning robot control for generalization with multiplicity of behavior. In CoRL, 2022.
  20. Receding horizon control of nonlinear systems. In CDC, 1988.
  21. Improved denoising diffusion probabilistic models. In ICML, 2021.
  22. Capability iteration network for robot path planning. IJRA, 37(3):266–272, 2022.
  23. Learning non-convergent non-persistent short-run mcmc toward energy-based model. In NeurIPS, 2019.
  24. Imitating human behaviour with diffusion models. In ICLR, 2023.
  25. Joelle Pineau. Tractable planning under uncertainty: exploiting structure. Carnegie Mellon University, 2004.
  26. Value iteration networks on multiple levels of abstraction. In RSS, 2019.
  27. Deep unsupervised learning using nonequilibrium thermodynamics. In ICML, 2015.
  28. Edward J Sondik. The optimal control of partially observable markov decision processes. PhD thesis, Stanford University, 1971.
  29. Generative modeling by estimating gradients of the data distribution. In NeurIPS, 2019.
  30. Score-based generative modeling through stochastic differential equations. In ICLR, 2021.
  31. Value iteration networks. In NeurIPS, 2016.
  32. Diffusebot: Breeding soft robots with physics-augmented generative diffusion models. In NeurIPS, 2023a.
  33. Diffusion policies as an expressive policy class for offline reinforcement learning. In ICLR, 2023b.
  34. Bayesian learning via stochastic gradient langevin dynamics. In ICML, 2011.
  35. David Bruce Wilson. Generating random spanning trees more quickly than the cover time. In STOC, 1996.
  36. Swin3d: A pretrained transformer backbone for 3d indoor scene understanding. arXiv:2304.06906, 2023.
Citations (1)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com