Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

PERL: Pivot-based Domain Adaptation for Pre-trained Deep Contextualized Embedding Models

About

Pivot-based neural representation models have lead to significant progress in domain adaptation for NLP. However, previous works that follow this approach utilize only labeled data from the source domain and unlabeled data from the source and target domains, but neglect to incorporate massive unlabeled corpora that are not necessarily drawn from these domains. To alleviate this, we propose PERL: A representation learning model that extends contextualized word embedding models such as BERT with pivot-based fine-tuning. PERL outperforms strong baselines across 22 sentiment classification domain adaptation setups, improves in-domain model performance, yields effective reduced-size models and increases model stability.

Eyal Ben-David, Carmel Rabinovitz, Roi Reichart• 2020

Related benchmarks

TaskDatasetResultRank
Sentiment AnalysisAmazon Reviews (test)
Average Accuracy87.5
24
Sentiment ClassificationMulti-Domain Sentiment Dataset 2008 (test)
Accuracy (A->D)0.729
12
Showing 2 of 2 rows

Other info

Follow for update