A novel method for learning policies from variable constraint data

M. Howard, S. Klanke, M. Gienger, C. Goerick, S. Vijayakumar

Research output: Contribution to journalArticlepeer-review

Abstract / Description of output

Many everyday human skills can be framed in terms of performing some task subject to constraints imposed by the environment. Constraints are usually unobservable and frequently change between contexts. In this paper, we present a novel approach for learning (unconstrained) control policies from movement data, where observations come from movements under different constraints. As a key ingredient, we introduce a small but highly effective modification to the standard risk functional, allowing us to make a meaningful comparison between the estimated policy and constrained observations. We demonstrate our approach on systems of varying complexity, including kinematic data from the ASIMO humanoid robot with 27 degrees of freedom, and present results for learning from human demonstration.
Original languageEnglish
Pages (from-to)105-121
Number of pages17
JournalAutonomous Robots
Volume27
Issue number2
DOIs
Publication statusPublished - 2009

Keywords / Materials (for Non-textual outputs)

  • direct policy learning
  • constrained motion
  • Imitation
  • nullspace control

Fingerprint

Dive into the research topics of 'A novel method for learning policies from variable constraint data'. Together they form a unique fingerprint.

Cite this