Node-RF: Learning Generalized Continuous Space-Time Scene Dynamics with Neural ODE-based NeRFs
About
Predicting scene dynamics from visual observations is challenging. Existing methods capture dynamics only within observed boundaries failing to extrapolate far beyond the training sequence. Node-RF (Neural ODE-based NeRF) overcomes this limitation by integrating Neural Ordinary Differential Equations (NODEs) with dynamic Neural Radiance Fields (NeRFs), enabling a continuous-time, spatiotemporal representation that generalizes beyond observed trajectories at constant memory cost. From visual input, Node-RF learns an implicit scene state that evolves over time via an ODE solver, propagating feature embeddings via differential calculus. A NeRF-based renderer interprets calculated embeddings to synthesize arbitrary views for long-range extrapolation. Training on multiple motion sequences with shared dynamics allows for generalization to unseen conditions. Our experiments demonstrate that Node-RF can characterize abstract system behavior without explicit model to identify critical points for future predictions.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Video Extrapolation | Bouncing Balls Long-term 4x Extrapolation | X-CLIP Similarity0.1775 | 6 | |
| Dynamic Prediction | Bifurcating Hill synthetic (unseen sequence) | IoU48.5 | 4 | |
| Dynamic Scene Modeling | Pendulum Extrapolation | SSIM0.469 | 4 | |
| Dynamic Scene Modeling | Pendulum Interpolation | SSIM53.1 | 3 | |
| Dynamic Prediction | Oscillating Ball synthetic (unseen sequence) | IoU33.27 | 2 |