Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Twinner: Shining Light on Digital Twins in a Few Snaps

About

We present the first large reconstruction model, Twinner, capable of recovering a scene's illumination as well as an object's geometry and material properties from only a few posed images. Twinner is based on the Large Reconstruction Model and innovates in three key ways: 1) We introduce a memory-efficient voxel-grid transformer whose memory scales only quadratically with the size of the voxel grid. 2) To deal with scarcity of high-quality ground-truth PBR-shaded models, we introduce a large fully-synthetic dataset of procedurally-generated PBR-textured objects lit with varied illumination. 3) To narrow the synthetic-to-real gap, we finetune the model on real life datasets by means of a differentiable physically-based shading model, eschewing the need for ground-truth illumination or material properties which are challenging to obtain in real life. We demonstrate the efficacy of our model on the real life StanfordORB benchmark where, given few input views, we achieve reconstruction quality significantly superior to existing feedforward reconstruction networks, and comparable to significantly slower per-scene optimization methods.

Jesus Zarzar, Tom Monnier, Roman Shapovalov, Andrea Vedaldi, David Novotny• 2025

Related benchmarks

TaskDatasetResultRank
Novel Scene RelightingStanford-ORB 1.0 (test)
PSNR-H24.55
26
Novel View SynthesisStanford-ORB 1.0 (test)
PSNR-H21.7
18
Geometry EstimationStanfordORB
Depth Error0.709
11
Illumination EstimationStanfordORB (full)
Angular Error8.09
7
Showing 4 of 4 rows

Other info

Follow for update