Journal:IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
Abstract:Learning how humans manipulate objects requires machines to acquire knowledge from two perspectives: one for understanding object affordances and the other for learning human's interactions based on the affordances. Even though these two knowledge bases are crucial, we find that current databases lack a comprehensive awareness of them. In this work, we propose a multi-modal and rich-annotated knowledge repository, OakInk, for visual and cognitive understanding of hand-object interactions. We start to collect 1,800 common household objects and annotate their affordances to construct the first knowledge base: Oak. Given the affordance, we record rich human interactions with 100 selected objects in Oak. Finally, we transfer the interactions on the 100 recorded objects to their virtual counterparts through a novel method: Tink. The recorded and transferred hand-object interactions constitute the second knowledge base: Ink. As a result, OakInk contains 50,000 distinct affordance-aware and intent-oriented hand-object interactions. We benchmark OakInk on pose estimation and grasp generation tasks. Moreover, we propose two practical applications of OakInk: intent-based interaction generation and handover generation. Our dataset and source code are publicly available at www. oakink. net.
Co-author:Liu Liu,Xinyu Zhan,Kailin Li
First Author:Lixin Yang
Indexed by:Other
Correspondence Author:Cewu Lu
Page Number:20953-20962
Translation or Not:no
Date of Publication:2022-07-13
Associate professor
Supervisor of Master's Candidates
Date of Birth:1993-09-21
E-Mail:
Date of Employment:2022-09-08
School/Department:计算机科学与技术系
Education Level:Postgraduate (Postdoctoral)
Gender:Male
Degree:Doctoral degree
Status:Employed
Alma Mater:中国科学技术大学
Honors and Titles:
2022年工业信息学顶刊TII年度最佳论文 2022-10-19
2019年国际工业信息学会议最佳论文奖 2019-07-18
2020年上海市“超级博士后” 2020-12-01
The Last Update Time : ..