Detail View
Non-differentiable Reward Optimization for Diffusion-based Autonomous Motion Planning
Citations
WEB OF SCIENCE
Citations
SCOPUS
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Lee, Giwon | - |
| dc.contributor.author | Park, Daehee | - |
| dc.contributor.author | Jeong, Jaewoo | - |
| dc.contributor.author | Yoon, Kuk-Jin | - |
| dc.date.accessioned | 2026-02-09T22:40:10Z | - |
| dc.date.available | 2026-02-09T22:40:10Z | - |
| dc.date.created | 2026-01-06 | - |
| dc.date.issued | 2025-10-21 | - |
| dc.identifier.isbn | 9798331543938 | - |
| dc.identifier.issn | 2153-0866 | - |
| dc.identifier.uri | https://scholar.dgist.ac.kr/handle/20.500.11750/59995 | - |
| dc.description.abstract | Safe and effective motion planning is crucial for autonomous robots. Diffusion models excel at capturing complex agent interactions, a fundamental aspect of decision-making in dynamic environments. Recent studies have successfully applied diffusion models to motion planning, demonstrating their competence in handling complex scenarios and accurately predicting multi-modal future trajectories. Despite their effectiveness, diffusion models have limitations in training objectives, as they approximate data distributions rather than explicitly capturing the underlying decision-making dynamics. However, the crux of motion planning lies in non-differentiable downstream objectives, such as safety (collision avoidance) and effectiveness (goal-reaching), which conventional learning algorithms cannot directly optimize. In this paper, we propose a reinforcement learning-based training scheme for diffusion motion planning models, enabling them to effectively learn non-differentiable objectives that explicitly measure safety and effectiveness. Specifically, we introduce a reward-weighted dynamic thresholding algorithm to shape a dense reward signal, facilitating more effective training and outperforming models trained with differentiable objectives. State-of-the-art performance on pedestrian datasets (CrowdNav, ETH-UCY) compared to various baselines demonstrates the versatility of our approach for safe and effective motion planning. | - |
| dc.language | English | - |
| dc.publisher | IEEE Robotics and Automation Society | - |
| dc.relation.ispartof | 2025 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) | - |
| dc.title | Non-differentiable Reward Optimization for Diffusion-based Autonomous Motion Planning | - |
| dc.type | Conference Paper | - |
| dc.identifier.doi | 10.1109/IROS60139.2025.11247412 | - |
| dc.identifier.bibliographicCitation | IEEE/RSJ International Conference on Intelligent Robots and Systems, pp.13215 - 13222 | - |
| dc.identifier.url | https://ras.papercept.net/conferences/conferences/IROS25/program/IROS25_ContentListWeb_2.html#wect4 | - |
| dc.citation.conferenceDate | 2025-10-19 | - |
| dc.citation.conferencePlace | CC | - |
| dc.citation.conferencePlace | Hangzhou | - |
| dc.citation.endPage | 13222 | - |
| dc.citation.startPage | 13215 | - |
| dc.citation.title | IEEE/RSJ International Conference on Intelligent Robots and Systems | - |
File Downloads
- There are no files associated with this item.
공유
Related Researcher
- Park, Daehee박대희
-
Department of Electrical Engineering and Computer Science
Total Views & Downloads
???jsp.display-item.statistics.view???: , ???jsp.display-item.statistics.download???:
