Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Learning from Many and Adapting to the Unknown in Open-set Test Streams

About

Large Language Models (LLMs) generalize across tasks via reusable representations and flexible reasoning, yet remain brittle in real deployment under evolving tasks and continual distribution shift. A common approach is Test-Time Adaptation (TTA), existing ones of which updates models with hand-designed unsupervised objectives over the full parameter space and mostly overlook preserving shared source knowledge and the reliability of adaptation signals. Drawing on molecular signaling cascades of memory updating in Drosophila, we propose Synapse Consolidation (SyCo), a parameter-efficient LLM adaptation method that updates low-rank adapters through Rac1 and MAPK pathways under the guidance of a structured TTA objective driven by problem understanding, process understanding, and source-domain guardrail. Rac1 confines plasticity to a tail-gradient subspace that is less critical for source knowledge, enabling rapid specialization while preserving source representations. MAPK uses a tiered controller to suppress noisy updates and consolidate useful adaptations under non-stationary streams. To model real deployments with multiple sources and continually emerging tasks, we introduce Multi-source Open-set Adaptation (MOA) setting, where a model is trained on multiple labeled source tasks and then adapts on open, non-stationary unlabeled test streams that mix seen and unseen tasks with partial overlap in label and intent space. Across 18 NLP datasets and the MOA setting, SyCo consistently outperforms strong baselines, achieving 78.31\% on unseen-task adaptation and 85.37\% on unseen-data shifts.

Xiao Zhang, Juntao Lyu, Tianyu Hu, Qianchuan Zhao, Huimin Ma• 2026

Related benchmarks

TaskDatasetResultRank
Reading ComprehensionC3
Accuracy51
73
Aspect-level Sentiment AnalysisCOTE BD
F1 Score93.2
34
Natural Language InferenceCMNLI syntactically perturbed
Accuracy76.67
17
Poetry MatchingCCPM
F1 Score81.37
17
Question AnsweringCMRC syntactically perturbed 2018
F1 Score85.9
17
Reading ComprehensionSanWen syntactically perturbed
F1 Score92.47
17
Relation ExtractionFinRE
F1 Score79.5
17
Semantic SimilarityLCQMC
Accuracy83.6
17
Semantic SimilarityAFQMC syntactically perturbed
Accuracy94.27
17
Sentiment AnalysisAmazon syntactically perturbed
Accuracy72.38
17
Showing 10 of 17 rows

Other info

Follow for update