ClusT3: Information Invariant Test-Time Training
About
Deep Learning models have shown remarkable performance in a broad range of vision tasks. However, they are often vulnerable against domain shifts at test-time. Test-time training (TTT) methods have been developed in an attempt to mitigate these vulnerabilities, where a secondary task is solved at training time simultaneously with the main task, to be later used as an self-supervised proxy task at test-time. In this work, we propose a novel unsupervised TTT technique based on the maximization of Mutual Information between multi-scale feature maps and a discrete latent representation, which can be integrated to the standard training as an auxiliary clustering task. Experimental results demonstrate competitive classification performance on different popular test-time adaptation benchmarks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Reading Comprehension | C3 | Accuracy49.32 | 73 | |
| Aspect-level Sentiment Analysis | COTE BD | F1 Score90.14 | 34 | |
| Semantic Similarity | LCQMC | Accuracy75.07 | 17 | |
| Natural Language Inference | OCNLI | Accuracy55.19 | 17 | |
| Sentiment Analysis | ChnSent | Accuracy90.43 | 17 | |
| Aspect-based Sentiment Analysis | COTE-MFW syntactically perturbed | F1 Score85.42 | 17 | |
| Reading Comprehension | DRCD | F1 Score83.82 | 17 | |
| Poetry Matching | CCPM | F1 Score72.81 | 17 | |
| Relation Extraction | FinRE | F1 Score72.6 | 17 | |
| Question Answering | CMRC syntactically perturbed 2018 | F1 Score74.77 | 17 |