Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Sign2GPT: Leveraging Large Language Models for Gloss-Free Sign Language Translation

About

Automatic Sign Language Translation requires the integration of both computer vision and natural language processing to effectively bridge the communication gap between sign and spoken languages. However, the deficiency in large-scale training data to support sign language translation means we need to leverage resources from spoken language. We introduce, Sign2GPT, a novel framework for sign language translation that utilizes large-scale pretrained vision and language models via lightweight adapters for gloss-free sign language translation. The lightweight adapters are crucial for sign language translation, due to the constraints imposed by limited dataset sizes and the computational requirements when training with long sign videos. We also propose a novel pretraining strategy that directs our encoder to learn sign representations from automatically extracted pseudo-glosses without requiring gloss order information or annotations. We evaluate our approach on two public benchmark sign language translation datasets, namely RWTH-PHOENIX-Weather 2014T and CSL-Daily, and improve on state-of-the-art gloss-free translation performance with a significant margin.

Ryan Wong, Necati Cihan Camgoz, Richard Bowden• 2024

Related benchmarks

TaskDatasetResultRank
Sign Language TranslationPHOENIX-2014T (test)
BLEU-422.52
159
Sign Language TranslationCSL-Daily (test)
BLEU-415.4
99
Sign Language TranslationCSL-Daily (dev)
ROUGE21.75
80
Sign Language TranslationPHOENIX14T (test)
BLEU-422.52
50
Sign Language TranslationCSL-Daily v1 (test)
ROUGE42.36
25
Sign Language TranslationBOBSL SENT (test)
B40.9
23
Showing 6 of 6 rows

Other info

Follow for update