HiVid: LLM-Guided Video Saliency For Content-Aware VOD And Live Streaming
About
Content-aware streaming requires dynamic, chunk-level importance weights to optimize subjective quality of experience (QoE). However, direct human annotation is prohibitively expensive while vision-saliency models generalize poorly. We introduce HiVid, the first framework to leverage Large Language Models (LLMs) as a scalable human proxy to generate high-fidelity weights for both Video-on-Demand (VOD) and live streaming. We address 3 non-trivial challenges: (1) To extend LLMs' limited modality and circumvent token limits, we propose a perception module to assess frames in a local context window, autoregressively building a coherent understanding of the video. (2) For VOD with rating inconsistency across local windows, we propose a ranking module to perform global re-ranking with a novel LLM-guided merge-sort algorithm. (3) For live streaming which requires low-latency, online inference without future knowledge, we propose a prediction module to predict future weights with a multi-modal time series model, which comprises a content-aware attention and adaptive horizon to accommodate asynchronous LLM inference. Extensive experiments show HiVid improves weight prediction accuracy by up to 11.5\% for VOD and 26\% for live streaming over SOTA baselines. Real-world user study validates HiVid boosts streaming QoE correlation by 14.7\%.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Time Series Forecasting | 3 datasets averaged (test) | MAE0.08 | 22 | |
| Video highlight detection | Mr.HiSum | mAP (rho=50%)86 | 14 | |
| Video Saliency and Highlight Detection | TVSum (test) | PLCC0.5 | 9 | |
| Video Saliency and Highlight Detection | SumMe (test) | PLCC0.47 | 9 | |
| ABR MOS Correlation Prediction | Subjective MOS (test) | PLCC0.85 | 6 | |
| Mean Opinion Score (MOS) Correlation Analysis | User Study 30 participants | PLCC0.76 | 4 |