Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

OakInk: A Large-scale Knowledge Repository for Understanding Hand-Object Interaction

About

Learning how humans manipulate objects requires machines to acquire knowledge from two perspectives: one for understanding object affordances and the other for learning human's interactions based on the affordances. Even though these two knowledge bases are crucial, we find that current databases lack a comprehensive awareness of them. In this work, we propose a multi-modal and rich-annotated knowledge repository, OakInk, for visual and cognitive understanding of hand-object interactions. We start to collect 1,800 common household objects and annotate their affordances to construct the first knowledge base: Oak. Given the affordance, we record rich human interactions with 100 selected objects in Oak. Finally, we transfer the interactions on the 100 recorded objects to their virtual counterparts through a novel method: Tink. The recorded and transferred hand-object interactions constitute the second knowledge base: Ink. As a result, OakInk contains 50,000 distinct affordance-aware and intent-oriented hand-object interactions. We benchmark OakInk on pose estimation and grasp generation tasks. Moreover, we propose two practical applications of OakInk: intent-based interaction generation and handover generation. Our datasets and source code are publicly available at https://github.com/lixiny/OakInk.

Lixin Yang, Kailin Li, Xinyu Zhan, Fei Wu, Anran Xu, Liu Liu, Cewu Lu• 2022

Related benchmarks

TaskDatasetResultRank
Intent-Based Interaction GenerationOakInk Shape (test)
Penetration Depth (cm)1.57
4
Grasp Pose GenerationOakInk Shape (test)
Penetration Depth (cm)0.62
2
Hand Mesh RecoveryOakInk-Image (SP0)--
2
Hand-Object Pose EstimationOakInk-Image (SP0)--
2
Showing 4 of 4 rows

Other info

Code

Follow for update