Add and Thin: Diffusion for Temporal Point Processes
About
Autoregressive neural networks within the temporal point process (TPP) framework have become the standard for modeling continuous-time event data. Even though these models can expressively capture event sequences in a one-step-ahead fashion, they are inherently limited for long-term forecasting applications due to the accumulation of errors caused by their sequential nature. To overcome these limitations, we derive ADD-THIN, a principled probabilistic denoising diffusion model for TPPs that operates on entire event sequences. Unlike existing diffusion approaches, ADD-THIN naturally handles data with discrete and continuous components. In experiments on synthetic and real-world datasets, our model matches the state-of-the-art TPP models in density estimation and strongly outperforms them in forecasting.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Event Sequence Forecasting | Reddit-S (test) | Wasserstein Distance0.04 | 13 | |
| Event Sequence Forecasting | PUBG (test) | Wasserstein Distance0.02 | 13 | |
| Event Sequence Forecasting | Yelp-2 (test) | Wasserstein Distance0.02 | 13 | |
| Event Sequence Forecasting | Reddit-C (test) | Wasserstein distance0.03 | 13 | |
| Event Sequence Forecasting | Twitter (test) | Wasserstein Distance0.01 | 13 | |
| Event Sequence Forecasting | Yelp1 (test) | Wasserstein Distance0.04 | 13 | |
| Event count prediction | Yelp-A (test) | MARE0.42 | 11 | |
| Event count prediction | Yelp M (test) | MARE0.46 | 11 | |
| Unconditional generation of event sequences | YelpA (test) | MMD0.045 | 11 | |
| Unconditional generation of event sequences | YelpM (test) | MMD3 | 11 |