Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning Models for Actions and Person-Object Interactions with Transfer to Question Answering

About

This paper proposes deep convolutional network models that utilize local and global context to make human activity label predictions in still images, achieving state-of-the-art performance on two recent datasets with hundreds of labels each. We use multiple instance learning to handle the lack of supervision on the level of individual person instances, and weighted loss to handle unbalanced training data. Further, we show how specialized features trained on these datasets can be used to improve accuracy on the Visual Question Answering (VQA) task, in the form of multiple choice fill-in-the-blank questions (Visual Madlibs). Specifically, we tackle two types of questions on person activity and person-object relationship and show improvements over generic features trained on the ImageNet classification task.

Arun Mallya, Svetlana Lazebnik• 2016

Related benchmarks

TaskDatasetResultRank
Human-Object Interaction DetectionHICO-DET (test)--
493
Human-Object Interaction DetectionHICO-DET--
233
Activity RecognitionMPII (test)
mAP32.24
20
Multi-label HOI classificationHICO
mAP36.1
10
HOI ClassificationHICO (test)
mAP36.1
10
Activity PredictionHICO (test)
mAP36.1
9
Pair's RelationshipMadLibs Easy (test)
Accuracy78.5
7
Pair's RelationshipMadLibs Hard (test)
Accuracy56.17
7
Pair's RelationshipMadLibs Filtered Hard (test)
Accuracy62.06
7
Person's ActivityMadLibs Easy (test)
Accuracy87.57
7
Showing 10 of 12 rows

Other info

Follow for update