EatSense: Human Centric, Action Recognition and Localization Dataset for Understanding Eating Behaviors and Quality of Motion Assessment

Muhammad Ahmed Raza*, Longfei Chen, Li Nanbo, Robert B Fisher

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Current datasets for computer vision-based action recognition and localization cover a wide range of classes and challenging scenarios. However, these datasets don’t cater to healthcare applications that involve longterm monitoring, tracking minor changes in movements over time for healthcare purposes, or completely modeling a specific human behavior that includes multiple sub-actions. Specifically, there are no existing datasets for research on either health monitoring on atomic-action-based eating behavior or for a full range of eating sub-actions that fully segment the main action. Addressing these gaps is valuable for extending research on the health monitoring of elderly people and is needed for creating richer and more complete descriptions of actions. This paper introduces a new benchmark dataset named EatSense that targets both the computer vision and healthcare communities and fills in the aforementioned gaps. EatSense is recorded while a person eats in an uncontrolled dining setting. The key features of EatSense are the introduction of challenging atomic actions for action recognition, the significantly diverse durations of actions that make it difficult for current temporal action localization frameworks to localize, the capability to model comprehensive eating behavior in terms of a sequence of action-based behaviors, and the simulation of minor variations in motion or performance. We conduct extensive experiments on EatSense with baseline deep learning-based approaches for benchmarking and hand-crafted feature-based approaches for explainable applications. We believe this dataset will benefit future researchers in building robust temporal action localization networks, behavior recognition, and performance assessment models for eating.
Original languageEnglish
Article number104762
Pages (from-to)1-14
JournalImage and vision computing
Volume137
DOIs
Publication statusPublished - 12 Jul 2023

Keywords / Materials (for Non-textual outputs)

  • EatSense
  • Eating Vision Dataset
  • Atomic-Action Recognition
  • Change in movement direction

Fingerprint

Dive into the research topics of 'EatSense: Human Centric, Action Recognition and Localization Dataset for Understanding Eating Behaviors and Quality of Motion Assessment'. Together they form a unique fingerprint.

Cite this