Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

AnySplat: Feed-forward 3D Gaussian Splatting from Unconstrained Views

About

We introduce AnySplat, a feed forward network for novel view synthesis from uncalibrated image collections. In contrast to traditional neural rendering pipelines that demand known camera poses and per scene optimization, or recent feed forward methods that buckle under the computational weight of dense views, our model predicts everything in one shot. A single forward pass yields a set of 3D Gaussian primitives encoding both scene geometry and appearance, and the corresponding camera intrinsics and extrinsics for each input image. This unified design scales effortlessly to casually captured, multi view datasets without any pose annotations. In extensive zero shot evaluations, AnySplat matches the quality of pose aware baselines in both sparse and dense view scenarios while surpassing existing pose free approaches. Moreover, it greatly reduce rendering latency compared to optimization based neural fields, bringing real time novel view synthesis within reach for unconstrained capture settings.Project page: https://city-super.github.io/anysplat/

Lihan Jiang, Yucheng Mao, Linning Xu, Tao Lu, Kerui Ren, Yichen Jin, Xudong Xu, Mulin Yu, Jiangmiao Pang, Feng Zhao, Dahua Lin, Bo Dai• 2025

Related benchmarks

TaskDatasetResultRank
Novel View SynthesisRealEstate10K
PSNR23.88
173
Novel View SynthesisRE10K
SSIM61.5
142
Novel View SynthesisScanNet
PSNR25.559
130
Depth EstimationScanNet
AbsRel0.372
108
Novel View SynthesisDL3DV
PSNR18.84
84
Novel View SynthesisRe10K (test)
PSNR21.88
79
Novel View SynthesisScanNet++
PSNR22.337
67
Novel View SynthesisMip-NeRF360 (test)
PSNR17.329
62
Novel View SynthesisDL3DV (test)
PSNR18.27
61
Novel View SynthesisRE10K challenging views (test)
PSNR20.13
56
Showing 10 of 108 rows
...

Other info

Follow for update