HiSAC: Hierarchical Sparse Activation Compression for Ultra-long Sequence Modeling in Recommenders
About
Modern recommender systems leverage ultra-long user behavior sequences to capture dynamic preferences, but end-to-end modeling is infeasible in production due to latency and memory constraints. While summarizing history via interest centers offers a practical alternative, existing methods struggle to (1) identify user-specific centers at appropriate granularity and (2) accurately assign behaviors, leading to quantization errors and loss of long-tail preferences. To alleviate these issues, we propose Hierarchical Sparse Activation Compression (HiSAC), an efficient framework for personalized sequence modeling. HiSAC encodes interactions into multi-level semantic IDs and constructs a global hierarchical codebook. A hierarchical voting mechanism sparsely activates personalized interest-agents as fine-grained preference centers. Guided by these agents, Soft-Routing Attention aggregates historical signals in semantic space, weighting by similarity to minimize quantization error and retain long-tail behaviors. Deployed on Taobao's "Guess What You Like" homepage, HiSAC achieves significant compression and cost reduction, with online A/B tests showing a consistent 1.65% CTR uplift -- demonstrating its scalability and real-world effectiveness.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Sequence Modeling | Industrial Dataset | AUC64.55 | 48 | |
| Sequence Modeling | Taobao-MM | AUC0.6733 | 12 | |
| Long-sequence modeling | industrial dataset 10k sequence length | AUC64.44 | 4 | |
| Recommendation | Taobao 'Guess What You Like' homepage feed A/B Test (online) | IPV1.93 | 1 |