Please use this identifier to cite or link to this item:
|Title:||Understanding human-object interaction in RGB-D videos for human robot interaction||Authors:||Fang, Zhiwen
Thalmann, Nadia Magnenat
|Keywords:||Engineering::Computer science and engineering||Issue Date:||2018||Source:||Fang, Z., Yuan, J., & Thalmann, N. M. (2018). Understanding human-object interaction in RGB-D videos for human robot interaction. CGI 2018: Proceedings of Computer Graphics International 2018, 163-167. doi:10.1145/3208159.3208192||Abstract:||Detecting small hand-held objects plays a critical role for human-robot interaction, because the hand-held objects often reveal the intention of the human, e.g., use a cell phone to make a call or use a cup to drink, thus helps the robots understand the human behavior and response accordingly. Existing solutions relying on wearable sensor to detect hand-held objects often comprise the user experiences thus may not be preferred. With the development of commodity RGB-D sensors, e.g., Microsoft Kinect II, RGB and depth information have been used for the understanding of human actions and recognizing objects. Motivated by the previous success, we propose to detect hand-held objects using RGB-D sensor. However, instead of performing object detection alone, we propose to leverage human body pose as the context to achieve robust hand-held object detection in RGB-D videos. Our system demonstrates a person can interact with a humanoid social robot with hand-held object such as a cell phone or a cup. Experimental evaluations validate the effectiveness of this proposed method.||URI:||https://hdl.handle.net/10356/142068||DOI:||10.1145/3208159.3208192||Rights:||© 2018 Association for Computing Machinery. All rights reserved.||Fulltext Permission:||none||Fulltext Availability:||No Fulltext|
|Appears in Collections:||IMI Conference Papers|
checked on Sep 30, 2020
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.