Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Fast and Accurate Deep Bidirectional Language Representations for Unsupervised Learning

About

Even though BERT achieves successful performance improvements in various supervised learning tasks, applying BERT for unsupervised tasks still holds a limitation that it requires repetitive inference for computing contextual language representations. To resolve the limitation, we propose a novel deep bidirectional language model called Transformer-based Text Autoencoder (T-TA). The T-TA computes contextual language representations without repetition and has benefits of the deep bidirectional architecture like BERT. In run-time experiments on CPU environments, the proposed T-TA performs over six times faster than the BERT-based model in the reranking task and twelve times faster in the semantic similarity task. Furthermore, the T-TA shows competitive or even better accuracies than those of BERT on the above tasks.

Joongbo Shin, Yoonhyung Lee, Seunghyun Yoon, Kyomin Jung• 2020

Related benchmarks

TaskDatasetResultRank
Automatic Speech RecognitionLibriSpeech (dev-other)
WER16.09
411
ASR rescoringLibriSpeech clean (test)
WER5.11
21
ASR rescoringLibriSpeech (test-other)
WER16.91
21
ASR rescoringLibriSpeech (dev-clean)
WER4.98
9
Showing 4 of 4 rows

Other info

Follow for update