Feature Hashing for Large Scale Multitask Learning
About
Empirical evidence suggests that hashing is an effective strategy for dimensionality reduction and practical nonparametric estimation. In this paper we provide exponential tail bounds for feature hashing and show that the interaction between random subspaces is negligible with high probability. We demonstrate the feasibility of this approach with experimental results for a new use case -- multitask learning with hundreds of thousands of tasks.
Kilian Weinberger, Anirban Dasgupta, Josh Attenberg, John Langford, Alex Smola• 2009
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Intra-creator embedding similarity | Video Dataset (Same Day) | Intra-creator Similarity66 | 2 | |
| Intra-creator embedding similarity | Video Dataset Same & Next Day | Intra-creator Similarity66 | 2 | |
| Intra-creator embedding similarity | Video Dataset Overall | Intra-creator Embedding Similarity62 | 2 |
Showing 3 of 3 rows