Abstract
Automated human activity recognition is an essential task for Human Robot Interaction (HRI). A successful activity recognition system enables an assistant robot to provide precise services. In this paper, we present a two-layered approach that can recognize sub-level activities and high-level activities successively. In the first layer, the low-level activities are recognized based on the RGB-D video. In the second layer, we use the recognized low-level activities as input features for estimating high-level activities. Our model is embedded with a latent node, so that it can capture a richer class of sub-level semantics compared with the traditional approach. Our model is evaluated on a challenging benchmark dataset. We show that the proposed approach outperforms the single-layered approach, suggesting that the hierarchical nature of the model is able to better explain the observed data. The results also show that our model outperforms the state-of-the-art approach in accuracy, precision and recall.
Original language | English |
---|---|
Title of host publication | 23rd IEEE International Symposium on Robot and Human Interactive Communication |
DOIs | |
Publication status | Published - 2014 |
Externally published | Yes |
Event | 2014 23rd IEEE International Symposium on Robot and Human Interactive Communication, RO-MAN 2014 - Edinburgh, United Kingdom Duration: 25 Aug 2014 → 29 Aug 2014 Conference number: 23 |
Conference
Conference | 2014 23rd IEEE International Symposium on Robot and Human Interactive Communication, RO-MAN 2014 |
---|---|
Abbreviated title | RO-MAN |
Country/Territory | United Kingdom |
City | Edinburgh |
Period | 25/08/14 → 29/08/14 |
Keywords
- n/a OA procedure