VarMAE: Pre-training of Variational Masked Autoencoder for Domain-adaptive Language Understanding
About
Pre-trained language models have achieved promising performance on general benchmarks, but underperform when migrated to a specific domain. Recent works perform pre-training from scratch or continual pre-training on domain corpora. However, in many specific domains, the limited corpus can hardly support obtaining precise representations. To address this issue, we propose a novel Transformer-based language model named VarMAE for domain-adaptive language understanding. Under the masked autoencoding objective, we design a context uncertainty learning module to encode the token's context into a smooth latent distribution. The module can produce diverse and well-formed contextual representations. Experiments on science- and finance-domain NLU tasks demonstrate that VarMAE can be efficiently adapted to new domains with limited resources.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Citation Intent Classification | SciCite | Macro F186.32 | 14 | |
| Named Entity Recognition | Finance (test) | F1 Score70.15 | 14 | |
| Citation Intent Classification | ACL | F1 Score76.5 | 6 | |
| Named Entity Recognition | JNLPBA Science-domain (test) | Entity-level F174.43 | 4 | |
| Span Extraction | EBM-NLP Science-domain (test) | Token F176.01 | 4 | |
| Text Classification | ACL-ARC Science-domain (test) | Micro F176.5 | 4 | |
| Text Classification | SciCite Science-domain (test) | Micro-average F186.32 | 4 | |
| Text Classification | OIR Finance-domain (test) | Micro F168.77 | 4 | |
| Text Classification | MTC Finance-domain (test) | Micro F1 Score56.58 | 4 | |
| Text Matching | PSM Finance-domain (test) | Micro F1 Score53.68 | 4 |