Test-Time Adaptation for Unsupervised Combinatorial Optimization
About
Unsupervised neural combinatorial optimization (NCO) enables learning powerful solvers without access to ground-truth solutions. Existing approaches fall into two disjoint paradigms: models trained for generalization across instances, and instance-specific models optimized independently at test time. While the former are efficient during inference, they lack effective instance-wise adaptability; the latter are flexible but fail to exploit learned inductive structure and are prone to poor local optima. This motivates the central question of our work: how can we leverage the inductive bias learned through generalization while unlocking the flexibility required for effective instance-wise adaptation? We first identify a challenge in bridging these two paradigms: generalization-focused models often constitute poor warm starts for instance-wise optimization, potentially underperforming even randomly initialized models when fine-tuned at test time. To resolve this incompatibility, we propose TACO, a model-agnostic test-time adaptation framework that unifies and extends the two existing paradigms for unsupervised NCO. TACO applies strategic warm-starting to partially relax trained parameters while preserving inductive bias, enabling rapid and effective unsupervised adaptation. Crucially, compared to naively fine-tuning a trained generalizable model or optimizing an instance-specific model from scratch, TACO achieves better solution quality while incurring negligible additional computational cost. Experiments on canonical CO problems, Minimum Vertex Cover and Maximum Clique, demonstrate the effectiveness and robustness of TACO across static, distribution-shifted, and dynamic combinatorial optimization problems, establishing it as a practical bridge between generalizable and instance-specific unsupervised NCO.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Maximum Clique | RB200 MC instances (static) | Mean ApR98.465 | 38 | |
| Maximum Clique | Twitter MC instances (static) | Mean ApR0.9977 | 38 | |
| Maximum Clique | RB500 MC instances (static) | Mean ApR98.684 | 36 | |
| Maximum Clique | COLLAB | Mean ApR1 | 30 | |
| Minimum Vertex Cover | COLLAB (test) | AR*1.0003 | 16 | |
| Minimum Vertex Cover | RB500 (test) | Approximation Ratio1.0121 | 13 | |
| Maximum Clique | RB200 trained on Twitter | Mean ApR0.9917 | 10 | |
| Minimum Vertex Cover | RB200 trained on Twitter | Mean ApR1.0196 | 10 | |
| Minimum Vertex Cover | Dynamic problems (Twitter Tennis UO and COVID-19 England) | Mean ApR0.9964 | 10 | |
| Maximum Clique | Dynamic problems (Twitter Tennis UO and COVID-19 England) | Mean ApR100 | 10 |