Perspective Fields for Single Image Camera Calibration
About
Geometric camera calibration is often required for applications that understand the perspective of the image. We propose perspective fields as a representation that models the local perspective properties of an image. Perspective Fields contain per-pixel information about the camera view, parameterized as an up vector and a latitude value. This representation has a number of advantages as it makes minimal assumptions about the camera model and is invariant or equivariant to common image editing operations like cropping, warping, and rotation. It is also more interpretable and aligned with human perception. We train a neural network to predict Perspective Fields and the predicted Perspective Fields can be converted to calibration parameters easily. We demonstrate the robustness of our approach under various scenarios compared with camera calibration-based methods and show example applications in image compositing.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Perspective Field prediction | Stanford2D3D (test) | Up Mean2.18 | 12 | |
| Perspective Field prediction | TartanAir (test) | Mean Angular Error (Up)2.81 | 12 | |
| Camera Calibration | iBIMS-1 | Mean Error10.6 | 8 | |
| Camera Calibration | ETH3D & iBims-1 Average | Mean Error12.1 | 8 | |
| Camera Calibration | ETH3D | Mean Error13.6 | 8 | |
| Camera Parameter Estimation | GSV uncentered (test) | Roll Mean Error1.37 | 6 | |
| Monocular Camera Calibration | GSV dataset (test) | Mean FoV Error (°)3.07 | 6 | |
| Object-centric prediction | Objectron 1.0 (crop) | Up Mean Error4.19 | 5 | |
| Object-centric prediction | Objectron 1.0 (isolated) | Up Mean Error4.45 | 5 | |
| Camera Parameter Estimation | GSV centered principal-point | Roll Error Mean (°)0.66 | 4 |