Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Aggregation Cross-Entropy for Sequence Recognition

About

In this paper, we propose a novel method, aggregation cross-entropy (ACE), for sequence recognition from a brand new perspective. The ACE loss function exhibits competitive performance to CTC and the attention mechanism, with much quicker implementation (as it involves only four fundamental formulas), faster inference\back-propagation (approximately O(1) in parallel), less storage requirement (no parameter and negligible runtime memory), and convenient employment (by replacing CTC with ACE). Furthermore, the proposed ACE loss function exhibits two noteworthy properties: (1) it can be directly applied for 2D prediction by flattening the 2D prediction into 1D prediction as the input and (2) it requires only characters and their numbers in the sequence annotation for supervision, which allows it to advance beyond sequence recognition, e.g., counting problem. The code is publicly available at https://github.com/summerlvsong/Aggregation-Cross-Entropy.

Zecheng Xie, Yaoxiong Huang, Yuanzhi Zhu, Lianwen Jin, Yuliang Liu, Lele Xie• 2019

Related benchmarks

TaskDatasetResultRank
Scene Text RecognitionSVT (test)
Word Accuracy82.6
289
Scene Text RecognitionIIIT5K (test)
Word Accuracy82.3
244
Scene Text RecognitionIC15 (test)
Word Accuracy68.9
210
Scene Text RecognitionIC13 (test)
Word Accuracy89.7
207
Scene Text RecognitionSVTP (test)
Word Accuracy70.1
153
Scene Text RecognitionCUTE
Accuracy82.6
92
Scene Text RecognitionIC15
Accuracy68.9
86
Scene Text RecognitionCUTE (test)
Accuracy82.6
59
Scene Text RecognitionSVTP
Accuracy70.1
52
Scene Text RecognitionCUTE80
Accuracy82.6
47
Showing 10 of 11 rows

Other info

Follow for update