Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

FlowFormer: A Transformer Architecture for Optical Flow

About

We introduce optical Flow transFormer, dubbed as FlowFormer, a transformer-based neural network architecture for learning optical flow. FlowFormer tokenizes the 4D cost volume built from an image pair, encodes the cost tokens into a cost memory with alternate-group transformer (AGT) layers in a novel latent space, and decodes the cost memory via a recurrent transformer decoder with dynamic positional cost queries. On the Sintel benchmark, FlowFormer achieves 1.159 and 2.088 average end-point-error (AEPE) on the clean and final pass, a 16.5% and 15.5% error reduction from the best published result (1.388 and 2.47). Besides, FlowFormer also achieves strong generalization performance. Without being trained on Sintel, FlowFormer achieves 1.01 AEPE on the clean pass of Sintel training set, outperforming the best published result (1.29) by 21.7%.

Zhaoyang Huang, Xiaoyu Shi, Chao Zhang, Qiang Wang, Ka Chun Cheung, Hongwei Qin, Jifeng Dai, Hongsheng Li• 2022

Related benchmarks

TaskDatasetResultRank
Optical Flow EstimationKITTI 2015 (train)
Fl-epe0.53
446
Optical FlowSintel (train)
AEPE (Clean)0.48
200
Optical Flow EstimationSintel Final (test)
EPE2.09
133
Optical FlowSintel (test)
AEPE (Final)2.09
120
Optical Flow EstimationSintel clean (test)
EPE1.16
120
Optical FlowKITTI 2012 (train)--
115
Optical FlowKITTI 2015 (test)
Fl Error (All)4.68
109
Optical Flow EstimationKITTI 2015 (test)
Fl-all4.68
108
Optical FlowSintel Final (train)
EPE0.74
106
Optical FlowSintel Clean (train)
EPE0.48
98
Showing 10 of 30 rows

Other info

Follow for update