Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Translating Videos to Natural Language Using Deep Recurrent Neural Networks

About

Solving the visual symbol grounding problem has long been a goal of artificial intelligence. The field appears to be advancing closer to this goal with recent breakthroughs in deep learning for natural language grounding in static images. In this paper, we propose to translate videos directly to sentences using a unified deep neural network with both convolutional and recurrent structure. Described video datasets are scarce, and most existing methods have been applied to toy domains with a small vocabulary of possible words. By transferring knowledge from 1.2M+ images with category labels and 100,000+ images with captions, our method is able to create sentence descriptions of open-domain videos with large vocabularies. We compare our approach with recent work using language generation metrics, subject, verb, and object prediction accuracy, and a human evaluation.

Subhashini Venugopalan, Huijuan Xu, Jeff Donahue, Marcus Rohrbach, Raymond Mooney, Kate Saenko• 2014

Related benchmarks

TaskDatasetResultRank
Text-to-Video RetrievalDiDeMo (test)
R@111.9
376
Text-to-Video RetrievalDiDeMo
R@10.119
360
Text-to-Video RetrievalActivityNet
R@15
197
Video-to-Text retrievalDiDeMo
R@113.2
108
Video-to-Text retrievalActivityNet
R@17
99
Video-to-Text retrievalDiDeMo (test)
R@113.2
92
Paragraph-to-Video RetrievalActivityNet-captions (val1)
R@15
7
Video Retrieval (clip-caption)DiDeMo (test)
R@111.9
7
Text-to-Video RetrievalActivityNet Captions (val1 test)
R@15
6
Video-to-Paragraph RetrievalActivityNet-captions (val1)
R@17
6
Showing 10 of 11 rows

Other info

Follow for update