Decoding-based Regression
About
Language models have recently been shown capable of performing regression wherein numeric predictions are represented as decoded strings. In this work, we provide theoretical grounds for this capability and furthermore investigate the utility of causal sequence decoding models as numeric regression heads given any feature representation. We find that, despite being trained in the usual way - for next-token prediction via cross-entropy loss - decoder-based heads are as performant as standard pointwise heads when benchmarked over standard regression tasks, while being flexible enough to capture smooth numeric distributions, such as in the task of density estimation.
Xingyou Song, Dara Bahri• 2025
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Tabular Regression | TALENT 100 regression tasks | RMSE (Mean)0.5327 | 8 | |
| Code metric regression | APPS Leetcode (test) | RMSE0.493 | 6 | |
| Code metric regression | Triton Kernel Latency (test) | RMSE1.095 | 6 |
Showing 3 of 3 rows