Detail View

DC Field Value Language
dc.contributor.author Lee, Giwon -
dc.contributor.author Park, Daehee -
dc.contributor.author Jeong, Jaewoo -
dc.contributor.author Yoon, Kuk-Jin -
dc.date.accessioned 2026-02-09T22:40:10Z -
dc.date.available 2026-02-09T22:40:10Z -
dc.date.created 2026-01-06 -
dc.date.issued 2025-10-21 -
dc.identifier.isbn 9798331543938 -
dc.identifier.issn 2153-0866 -
dc.identifier.uri https://scholar.dgist.ac.kr/handle/20.500.11750/59995 -
dc.description.abstract Safe and effective motion planning is crucial for autonomous robots. Diffusion models excel at capturing complex agent interactions, a fundamental aspect of decision-making in dynamic environments. Recent studies have successfully applied diffusion models to motion planning, demonstrating their competence in handling complex scenarios and accurately predicting multi-modal future trajectories. Despite their effectiveness, diffusion models have limitations in training objectives, as they approximate data distributions rather than explicitly capturing the underlying decision-making dynamics. However, the crux of motion planning lies in non-differentiable downstream objectives, such as safety (collision avoidance) and effectiveness (goal-reaching), which conventional learning algorithms cannot directly optimize. In this paper, we propose a reinforcement learning-based training scheme for diffusion motion planning models, enabling them to effectively learn non-differentiable objectives that explicitly measure safety and effectiveness. Specifically, we introduce a reward-weighted dynamic thresholding algorithm to shape a dense reward signal, facilitating more effective training and outperforming models trained with differentiable objectives. State-of-the-art performance on pedestrian datasets (CrowdNav, ETH-UCY) compared to various baselines demonstrates the versatility of our approach for safe and effective motion planning. -
dc.language English -
dc.publisher IEEE Robotics and Automation Society -
dc.relation.ispartof 2025 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) -
dc.title Non-differentiable Reward Optimization for Diffusion-based Autonomous Motion Planning -
dc.type Conference Paper -
dc.identifier.doi 10.1109/IROS60139.2025.11247412 -
dc.identifier.bibliographicCitation IEEE/RSJ International Conference on Intelligent Robots and Systems, pp.13215 - 13222 -
dc.identifier.url https://ras.papercept.net/conferences/conferences/IROS25/program/IROS25_ContentListWeb_2.html#wect4 -
dc.citation.conferenceDate 2025-10-19 -
dc.citation.conferencePlace CC -
dc.citation.conferencePlace Hangzhou -
dc.citation.endPage 13222 -
dc.citation.startPage 13215 -
dc.citation.title IEEE/RSJ International Conference on Intelligent Robots and Systems -
Show Simple Item Record

File Downloads

  • There are no files associated with this item.

공유

qrcode
공유하기

Related Researcher

박대희
Park, Daehee박대희

Department of Electrical Engineering and Computer Science

read more

Total Views & Downloads

???jsp.display-item.statistics.view???: , ???jsp.display-item.statistics.download???: