Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

VarMAE: Pre-training of Variational Masked Autoencoder for Domain-adaptive Language Understanding

About

Pre-trained language models have achieved promising performance on general benchmarks, but underperform when migrated to a specific domain. Recent works perform pre-training from scratch or continual pre-training on domain corpora. However, in many specific domains, the limited corpus can hardly support obtaining precise representations. To address this issue, we propose a novel Transformer-based language model named VarMAE for domain-adaptive language understanding. Under the masked autoencoding objective, we design a context uncertainty learning module to encode the token's context into a smooth latent distribution. The module can produce diverse and well-formed contextual representations. Experiments on science- and finance-domain NLU tasks demonstrate that VarMAE can be efficiently adapted to new domains with limited resources.

Dou Hu, Xiaolong Hou, Xiyang Du, Mengyuan Zhou, Lianxin Jiang, Yang Mo, Xiaofeng Shi• 2022

Related benchmarks

TaskDatasetResultRank
Citation Intent ClassificationSciCite
Macro F186.32
14
Named Entity RecognitionFinance (test)
F1 Score70.15
14
Citation Intent ClassificationACL
F1 Score76.5
6
Named Entity RecognitionJNLPBA Science-domain (test)
Entity-level F174.43
4
Span ExtractionEBM-NLP Science-domain (test)
Token F176.01
4
Text ClassificationACL-ARC Science-domain (test)
Micro F176.5
4
Text ClassificationSciCite Science-domain (test)
Micro-average F186.32
4
Text ClassificationOIR Finance-domain (test)
Micro F168.77
4
Text ClassificationMTC Finance-domain (test)
Micro F1 Score56.58
4
Text MatchingPSM Finance-domain (test)
Micro F1 Score53.68
4
Showing 10 of 10 rows

Other info

Follow for update