The Observer Effect in World Models: Invasive Adaptation Corrupts Latent Physics
About
Determining whether neural models internalize physical laws as world models, rather than exploiting statistical shortcuts, remains challenging, especially under out-of-distribution (OOD) shifts. Standard evaluations often test latent capability via downstream adaptation (e.g., fine-tuning or high-capacity probes), but such interventions can change the representations being measured and thus confound what was learned during self-supervised learning (SSL). We propose a non-invasive evaluation protocol, PhyIP. We test whether physical quantities are linearly decodable from frozen representations, motivated by the linear representation hypothesis. Across fluid dynamics and orbital mechanics, we find that when SSL achieves low error, latent structure becomes linearly accessible. PhyIP recovers internal energy and Newtonian inverse-square scaling on OOD tests (e.g., $\rho > 0.90$). In contrast, adaptation-based evaluations can collapse this structure ($\rho \approx 0.05$). These findings suggest that adaptation-based evaluation can obscure latent structures and that low-capacity probes offer a more accurate evaluation of physical world models.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Internal Energy Probing | 3D Red Supergiant N=3 OOD | MAPE18.2 | 9 | |
| Internal Energy Probing | 2D Turbulent Layer N=9 OOD | MAPE36.9 | 9 | |
| Internal Energy Probing | 3D Supernova N=27 OOD | MAPE92.1 | 9 | |
| Self-supervised Learning | 2D Turbulent Layer N=9 OOD | Epsilon OOD20 | 4 | |
| Self-supervised Learning | 3D Red Supergiant N=3 OOD | Epsilon OOD0.02 | 4 | |
| Self-supervised Learning | 3D Supernova N=27 OOD | Epsilon OOD0.3 | 4 |