Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Decoding-based Regression

About

Language models have recently been shown capable of performing regression wherein numeric predictions are represented as decoded strings. In this work, we provide theoretical grounds for this capability and furthermore investigate the utility of causal sequence decoding models as numeric regression heads given any feature representation. We find that, despite being trained in the usual way - for next-token prediction via cross-entropy loss - decoder-based heads are as performant as standard pointwise heads when benchmarked over standard regression tasks, while being flexible enough to capture smooth numeric distributions, such as in the task of density estimation.

Xingyou Song, Dara Bahri• 2025

Related benchmarks

TaskDatasetResultRank
Tabular RegressionTALENT 100 regression tasks
RMSE (Mean)0.5327
8
Code metric regressionAPPS Leetcode (test)
RMSE0.493
6
Code metric regressionTriton Kernel Latency (test)
RMSE1.095
6
Showing 3 of 3 rows

Other info

Follow for update