Learning Goal-Directed Object Pushing in Cluttered Scenes With Location-Based Attention

Nils Dengler, Juan Del Aguila Ferrandis, Joao Pousa De Moura, Sethu Vijayakumar, Maren Bennewitz

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

In complex scenarios where typical pick-and-place techniques are insufficient, often non-prehensile manipulation can ensure that a robot is able to fulfill its task. However, nonprehensile manipulation is challenging due to its underactuated nature with hybrid-dynamics, where a robot needs to reason about an object’s long-term behavior and contact-switching, while being robust to contact uncertainty. The presence of clutter in the workspace further complicates this task, introducing the need to include more advanced spatial analysis to avoid unwanted collisions. Building upon prior work on reinforcement learning with multimodal categorical exploration for planar pushing, we propose to incorporate location-based attention to enable robust manipulation in cluttered scenes. Unlike previous approaches addressing this obstacle avoiding pushing task, our framework requires no predefined global paths and considers the desired target orientation of the manipulated object. Experimental results in simulation as well as with a real KUKA iiwa robot arm demonstrate that our learned policy manipulates objects successfully while avoiding collisions through complex obstacle configurations, including dynamic obstacles, to reach the desired target pose.
Original languageEnglish
Title of host publicationProceedings of the 2025 IEEE/RSJ International Conference on Intelligent Robots and Systems
Pages1-7
Number of pages7
Publication statusAccepted/In press - 16 Jun 2025

Publication series

NameProceedings of the International Conference on Intelligent Robots and Systems
PublisherIEEE
ISSN (Print)2153-0858
ISSN (Electronic)2153-0866

Fingerprint

Dive into the research topics of 'Learning Goal-Directed Object Pushing in Cluttered Scenes With Location-Based Attention'. Together they form a unique fingerprint.

Cite this