Social State Recognition and Knowledge-Level Planning for Human-Robot Interaction in a Bartender Domain

Ronald P. A. Petrick, Mary Ellen Foster, Amy Isard

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract / Description of output

We discuss preliminary work focusing on the problem of combining social interaction with task-based action in a dynamic, multiagent bartending domain, using an embodied robot. We show how the users' spoken input is interpreted, discuss how social states are inferred from the parsed speech together with low-level information from the vision system, and present a planning approach that models task, dialogue, and social actions in a simple bartending scenario. This approach allows us to build interesting plans, which have been evaluated in a real-world study, using a general purpose, off-the-shelf planner, as an alternative to more mainstream methods of interaction management.
Original languageEnglish
Title of host publicationProceedings of the AAAI 2012 Workshop on Grounding Language for Physical Systems
Pages32-38
Number of pages7
Publication statusPublished - 1 Jul 2012

Fingerprint

Dive into the research topics of 'Social State Recognition and Knowledge-Level Planning for Human-Robot Interaction in a Bartender Domain'. Together they form a unique fingerprint.

Cite this