Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Inference-time Physics Alignment of Video Generative Models with Latent World Models

About

State-of-the-art video generative models produce promising visual content yet often violate basic physics principles, limiting their utility. While some attribute this deficiency to insufficient physics understanding from pre-training, we find that the shortfall in physics plausibility also stems from suboptimal inference strategies. We therefore introduce WMReward and treat improving physics plausibility of video generation as an inference-time alignment problem. In particular, we leverage the strong physics prior of a latent world model (here, VJEPA-2) as a reward to search and steer multiple candidate denoising trajectories, enabling scaling test-time compute for better generation performance. Empirically, our approach substantially improves physics plausibility across image-conditioned, multiframe-conditioned, and text-conditioned generation settings, with validation from human preference study. Notably, in the ICCV 2025 Perception Test PhysicsIQ Challenge, we achieve a final score of 62.64%, winning first place and outperforming the previous state of the art by 7.42%. Our work demonstrates the viability of using latent world models to improve physics plausibility of video generation, beyond this specific instantiation or parameterization.

Jianhao Yuan, Xiaofeng Zhang, Felix Friedrich, Nicolas Beltran-Velez, Melissa Hall, Reyhane Askari-Hemmat, Xiaochuang Han, Nicolas Ballas, Michal Drozdzal, Adriana Romero-Soriano• 2026

Related benchmarks

TaskDatasetResultRank
Video GenerationPhysics-IQ
Phys. IQ Score62
45
Text-to-Video GenerationVideoPhy--
20
Human Preference EvaluationPhysicsIQ 1.0 (test)
Physics Plausibility Win Rate54.9
4
Human Preference EvaluationVideoPhy 1.0 (test)
Physics Plausibility Win Rate59.3
4
Showing 4 of 4 rows

Other info

Follow for update