Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Test-Time Adaptation for Unsupervised Combinatorial Optimization

About

Unsupervised neural combinatorial optimization (NCO) enables learning powerful solvers without access to ground-truth solutions. Existing approaches fall into two disjoint paradigms: models trained for generalization across instances, and instance-specific models optimized independently at test time. While the former are efficient during inference, they lack effective instance-wise adaptability; the latter are flexible but fail to exploit learned inductive structure and are prone to poor local optima. This motivates the central question of our work: how can we leverage the inductive bias learned through generalization while unlocking the flexibility required for effective instance-wise adaptation? We first identify a challenge in bridging these two paradigms: generalization-focused models often constitute poor warm starts for instance-wise optimization, potentially underperforming even randomly initialized models when fine-tuned at test time. To resolve this incompatibility, we propose TACO, a model-agnostic test-time adaptation framework that unifies and extends the two existing paradigms for unsupervised NCO. TACO applies strategic warm-starting to partially relax trained parameters while preserving inductive bias, enabling rapid and effective unsupervised adaptation. Crucially, compared to naively fine-tuning a trained generalizable model or optimizing an instance-specific model from scratch, TACO achieves better solution quality while incurring negligible additional computational cost. Experiments on canonical CO problems, Minimum Vertex Cover and Maximum Clique, demonstrate the effectiveness and robustness of TACO across static, distribution-shifted, and dynamic combinatorial optimization problems, establishing it as a practical bridge between generalizable and instance-specific unsupervised NCO.

Yiqiao Liao, Farinaz Koushanfar, Parinaz Naghizadeh• 2026

Related benchmarks

TaskDatasetResultRank
Maximum CliqueRB200 MC instances (static)
Mean ApR98.465
38
Maximum CliqueTwitter MC instances (static)
Mean ApR0.9977
38
Maximum CliqueRB500 MC instances (static)
Mean ApR98.684
36
Maximum CliqueCOLLAB
Mean ApR1
30
Minimum Vertex CoverCOLLAB (test)
AR*1.0003
16
Minimum Vertex CoverRB500 (test)
Approximation Ratio1.0121
13
Maximum CliqueRB200 trained on Twitter
Mean ApR0.9917
10
Minimum Vertex CoverRB200 trained on Twitter
Mean ApR1.0196
10
Minimum Vertex CoverDynamic problems (Twitter Tennis UO and COVID-19 England)
Mean ApR0.9964
10
Maximum CliqueDynamic problems (Twitter Tennis UO and COVID-19 England)
Mean ApR100
10
Showing 10 of 16 rows

Other info

Follow for update