Abstract
Learning models of user behaviour is an important problem that
is broadly applicable across many application domains requiring human-robot
interaction. In this work, we show that it is possible to learn generative models for distinct user behavioural types, extracted from human demonstrations,
by enforcing clustering of preferred task solutions within the latent space. We
use these models to differentiate between user types and to find cases with
overlapping solutions. Moreover, we can alter an initially guessed solution to
satisfy the preferences that constitute a particular user type by backpropagating through the learned differentiable models. An advantage of structuring
generative models in this way is that we can extract causal relationships between symbols that might form part of the user’s specification of the task,
as manifested in the demonstrations. We further parameterize these specifications through constraint optimization in order to find a safety envelope under
which motion planning can be performed. We show that the proposed method
is capable of correctly distinguishing between three user types, who differ in
degrees of cautiousness in their motion, while performing the task of moving
objects with a kinesthetically driven robot in a tabletop environment. Our
method successfully identifies the correct type, within the specified time, in
99% [97.8 − 99.8] of the cases, which outperforms an IRL baseline. We also
show that our proposed method correctly changes a default trajectory to one
satisfying a particular user specification even with unseen objects. The resulting trajectory is shown to be directly implementable on a PR2 humanoid robot
completing the same task.
Original language | English |
---|---|
Article number | 45 |
Pages (from-to) | 1-19 |
Number of pages | 19 |
Journal | Autonomous Agents and Multi-Agent Systems |
Volume | 34 |
Issue number | 2 |
Early online date | 17 Jun 2020 |
DOIs | |
Publication status | Published - 31 Oct 2020 |
Keywords
- Human-Computer Interaction
- robot learning
- Explainability