Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Fuel Gauge: Estimating Chain-of-Thought Length Ahead of Time in Large Multimodal Models

About

Reasoning Large Multi-modality Models (LMMs) have become the de facto choice for many applications. However, these models rely on a Chain-of-Thought (CoT) process that is lengthy and unpredictable at runtime, often resulting in inefficient use of computational resources (due to memory fragmentation) and sub-optimal accuracy (due to under- and over-thinking). We observe empirically that the CoT process follows a very simple form, whose behavior is independent of the specific generated samples. This suggests that the CoT length can be estimated ahead of time based on a hidden parameter representing the amount of "fuel" available to support the reasoning process. Based on this insight, we propose Fuel Gauge, the first method which extracts this hidden signal and predicts CoT length ahead of time. We demonstrate the utility on the Fuel Gauge on two downstream tasks: predictive KV cache allocation, which addresses memory fragmentation in LMM serving systems, and CoT length modulation, which mitigates under-thinking and over-thinking. Extensive experiments on LMMs across text-only, image-text, and video-text question answering benchmarks demonstrate the effectiveness, generalizability, and practical value of our Fuel Gauge. For example, on the GPQA-Diamond benchmark, our Fuel Gauge achieves less than half the CoT length prediction error compared to the baseline; this translates into a 13.37x reduction in the memory allocation frequency.

Yuedong Yang, Xiwen Wei, Mustafa Munir, Radu Marculescu• 2026

Related benchmarks

TaskDatasetResultRank
CoT Length PredictionMathVision m
rMAE0.2934
10
CoT Length PredictionLongVideoBench 15
rMAE0.4527
10
CoT Length PredictionLongVideoBench 60
rMAE0.4525
10
Fuel level estimationMathVision m
rMAE11.86
10
CoT Length PredictionGPQA Diamond
Relative MAE27.32
8
Fuel level estimationGPQA Diamond
rMAE13.22
8
KV Cache AllocationMathVision m
Allocation Count15.43
6
KV Cache AllocationLongVideoBench 60
Number of Allocations23.6
6
KV Cache AllocationLongVideoBench 15
Number of Allocations34.12
6
Predictive memory allocationGPQA Diamond
#Allocs39.87
4
Showing 10 of 12 rows

Other info

Follow for update