Please use this identifier to cite or link to this item:
Title: Understanding human-object interaction in RGB-D videos for human robot interaction
Authors: Fang, Zhiwen
Yuan, Junsong
Thalmann, Nadia Magnenat
Keywords: Engineering::Computer science and engineering
Issue Date: 2018
Source: Fang, Z., Yuan, J., & Thalmann, N. M. (2018). Understanding human-object interaction in RGB-D videos for human robot interaction. CGI 2018: Proceedings of Computer Graphics International 2018, 163-167. doi:10.1145/3208159.3208192
Abstract: Detecting small hand-held objects plays a critical role for human-robot interaction, because the hand-held objects often reveal the intention of the human, e.g., use a cell phone to make a call or use a cup to drink, thus helps the robots understand the human behavior and response accordingly. Existing solutions relying on wearable sensor to detect hand-held objects often comprise the user experiences thus may not be preferred. With the development of commodity RGB-D sensors, e.g., Microsoft Kinect II, RGB and depth information have been used for the understanding of human actions and recognizing objects. Motivated by the previous success, we propose to detect hand-held objects using RGB-D sensor. However, instead of performing object detection alone, we propose to leverage human body pose as the context to achieve robust hand-held object detection in RGB-D videos. Our system demonstrates a person can interact with a humanoid social robot with hand-held object such as a cell phone or a cup. Experimental evaluations validate the effectiveness of this proposed method.
DOI: 10.1145/3208159.3208192
Rights: © 2018 Association for Computing Machinery. All rights reserved.
Fulltext Permission: none
Fulltext Availability: No Fulltext
Appears in Collections:IMI Conference Papers

Citations 50

Updated on Jan 30, 2023

Page view(s)

Updated on Feb 2, 2023

Google ScholarTM




Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.