Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

O2O-Afford: Annotation-Free Large-Scale Object-Object Affordance Learning

About

Contrary to the vast literature in modeling, perceiving, and understanding agent-object (e.g., human-object, hand-object, robot-object) interaction in computer vision and robotics, very few past works have studied the task of object-object interaction, which also plays an important role in robotic manipulation and planning tasks. There is a rich space of object-object interaction scenarios in our daily life, such as placing an object on a messy tabletop, fitting an object inside a drawer, pushing an object using a tool, etc. In this paper, we propose a unified affordance learning framework to learn object-object interaction for various tasks. By constructing four object-object interaction task environments using physical simulation (SAPIEN) and thousands of ShapeNet models with rich geometric diversity, we are able to conduct large-scale object-object affordance learning without the need for human annotations or demonstrations. At the core of technical contribution, we propose an object-kernel point convolution network to reason about detailed interaction between two objects. Experiments on large-scale synthetic data and real-world data prove the effectiveness of the proposed approach. Please refer to the project webpage for code, data, video, and more materials: https://cs.stanford.edu/~kaichun/o2oafford

Kaichun Mo, Yuzhe Qin, Fanbo Xiang, Hao Su, Leonidas Guibas• 2021

Related benchmarks

TaskDatasetResultRank
pullingSAPIEN (test)
Sample Manipulation Accuracy29.93
8
pushingSAPIEN (test)
Sample Manipulation Accuracy34.6
8
Pulling Affordance PredictionSAPIEN PartNet-Mobility & ShapeNet (test)
F-Score46.03
7
Pulling Affordance PredictionSAPIEN (PartNet-Mobility & ShapeNet) Novel (test)
F-Score43.42
7
Pushing Affordance PredictionSAPIEN PartNet-Mobility & ShapeNet (test)
F-Score65.18
7
Pushing Affordance PredictionSAPIEN (PartNet-Mobility & ShapeNet) Novel (test)
F-Score59.04
7
Object AffordanceEgo-Exo4D & GIMO (test)
AUC71.52
4
Showing 7 of 7 rows

Other info

Follow for update