Efficient OCR for Building a Diverse Digital History
About
Thousands of users consult digital archives daily, but the information they can access is unrepresentative of the diversity of documentary history. The sequence-to-sequence architecture typically used for optical character recognition (OCR) - which jointly learns a vision and language model - is poorly extensible to low-resource document collections, as learning a language-vision model requires extensive labeled sequences and compute. This study models OCR as a character level image retrieval problem, using a contrastively trained vision encoder. Because the model only learns characters' visual features, it is more sample efficient and extensible than existing architectures, enabling accurate OCR in settings where existing solutions fail. Crucially, the model opens new avenues for community engagement in making digital history more representative of documentary history.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Optical Character Recognition | Chronicle American LoCCA (eval) | Character Error Rate1.5 | 18 | |
| Optical Character Recognition | Ancient Greek | Character Error Rate4.3 | 15 | |
| Optical Character Recognition | Japanese Horizontal | Character Error Rate0.6 | 12 | |
| Optical Character Recognition | Vertical Japanese tables | Character Error Rate0.7 | 6 | |
| Optical Character Recognition | Vertical Japanese prose | Character Error Rate2.7 | 6 |