Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Glance and Gaze: Inferring Action-aware Points for One-Stage Human-Object Interaction Detection

About

Modern human-object interaction (HOI) detection approaches can be divided into one-stage methods and twostage ones. One-stage models are more efficient due to their straightforward architectures, but the two-stage models are still advantageous in accuracy. Existing one-stage models usually begin by detecting predefined interaction areas or points, and then attend to these areas only for interaction prediction; therefore, they lack reasoning steps that dynamically search for discriminative cues. In this paper, we propose a novel one-stage method, namely Glance and Gaze Network (GGNet), which adaptively models a set of actionaware points (ActPoints) via glance and gaze steps. The glance step quickly determines whether each pixel in the feature maps is an interaction point. The gaze step leverages feature maps produced by the glance step to adaptively infer ActPoints around each pixel in a progressive manner. Features of the refined ActPoints are aggregated for interaction prediction. Moreover, we design an actionaware approach that effectively matches each detected interaction with its associated human-object pair, along with a novel hard negative attentive loss to improve the optimization of GGNet. All the above operations are conducted simultaneously and efficiently for all pixels in the feature maps. Finally, GGNet outperforms state-of-the-art methods by significant margins on both V-COCO and HICODET benchmarks. Code of GGNet is available at https: //github.com/SherlockHolmes221/GGNet.

Xubin Zhong, Xian Qu, Changxing Ding, Dacheng Tao• 2021

Related benchmarks

TaskDatasetResultRank
Human-Object Interaction DetectionHICO-DET (test)
mAP (full)29.2
493
Human-Object Interaction DetectionV-COCO (test)
AP (Role, Scenario 1)54.7
270
Human-Object Interaction DetectionHICO-DET
mAP (Full)27.36
233
Human-Object Interaction DetectionHICO-DET Known Object (test)
mAP (Full)33.5
112
Human-Object Interaction DetectionV-COCO 1.0 (test)
AP_role (#1)54.7
76
Human-Object Interaction DetectionHICO-DET 1 (test)
Full mAP33.5
33
Human-Object Interaction DetectionV-COCO
Box mAP (Scenario 1)54.7
32
Human-Object Interaction DetectionV-COCO 8 (test)
AP Role (S1)54.7
11
Showing 8 of 8 rows

Other info

Code

Follow for update