Bayesian Triplet Loss: Uncertainty Quantification in Image Retrieval
About
Uncertainty quantification in image retrieval is crucial for downstream decisions, yet it remains a challenging and largely unexplored problem. Current methods for estimating uncertainties are poorly calibrated, computationally expensive, or based on heuristics. We present a new method that views image embeddings as stochastic features rather than deterministic features. Our two main contributions are (1) a likelihood that matches the triplet constraint and that evaluates the probability of an anchor being closer to a positive than a negative; and (2) a prior over the feature space that justifies the conventional l2 normalization. To ensure computational efficiency, we derive a variational approximation of the posterior, called the Bayesian triplet loss, that produces state-of-the-art uncertainty estimates and matches the predictive performance of current state-of-the-art methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Visual Place Recognition Uncertainty Estimation | Nordland (test) | AUC-PR0.07 | 8 | |
| Visual Place Recognition Uncertainty Estimation | Pittsburgh (test) | AUC-PR44 | 8 | |
| Visual Place Recognition Uncertainty Estimation | San Francisco (test) | AUC-PR0.17 | 8 | |
| Visual Place Recognition Uncertainty Estimation | St Lucia (test) | AUC-PR34 | 8 | |
| Visual Place Recognition Uncertainty Estimation | Eynsham (test) | AUC-PR0.45 | 8 | |
| Visual Place Recognition Uncertainty Estimation | MSLS (test) | AUC-PR21 | 8 |