Infinite Recommendation Networks: A Data-Centric Approach
About
We leverage the Neural Tangent Kernel and its equivalence to training infinitely-wide neural networks to devise $\infty$-AE: an autoencoder with infinitely-wide bottleneck layers. The outcome is a highly expressive yet simplistic recommendation model with a single hyper-parameter and a closed-form solution. Leveraging $\infty$-AE's simplicity, we also develop Distill-CF for synthesizing tiny, high-fidelity data summaries which distill the most important knowledge from the extremely large and sparse user-item interaction matrix for efficient and accurate subsequent data-usage like model training, inference, architecture search, etc. This takes a data-centric approach to recommendation, where we aim to improve the quality of logged user-feedback data for subsequent modeling, independent of the learning algorithm. We particularly utilize the concept of differentiable Gumbel-sampling to handle the inherent data heterogeneity, sparsity, and semi-structuredness, while being scalable to datasets with hundreds of millions of user-item interactions. Both of our proposed approaches significantly outperform their respective state-of-the-art and when used together, we observe 96-105% of $\infty$-AE's performance on the full dataset with as little as 0.1% of the original dataset size, leading us to explore the counter-intuitive question: Is more data what you need for better recommendation?
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Recommendation | MovieLens 1M (test) | -- | 34 | |
| Recommendation | Netflix (test) | NDCG@10036.59 | 30 | |
| Recommendation | Douban (test) | -- | 10 | |
| Recommendation | Amazon Magazine (test) | AUC85.84 | 9 | |
| Collaborative Filtering | ML 1M | HR@1031.15 | 9 | |
| Recommendation System Efficiency | ML 1M (overall) | Training Time (m)2.24 | 9 | |
| Collaborative Filtering | Yelp 2018 | HR@104.62 | 9 | |
| Collaborative Filtering | Gowalla | HR@1011.77 | 9 | |
| Collaborative Filtering | ML-10M | HR@100.3583 | 8 | |
| Recommendation System Efficiency | ML 10M (overall) | Training Time (h)388.4 | 8 |