Guided Image-to-Image Translation with Bi-Directional Feature Transformation
About
We address the problem of guided image-to-image translation where we translate an input image into another while respecting the constraints provided by an external, user-provided guidance image. Various conditioning methods for leveraging the given guidance image have been explored, including input concatenation , feature concatenation, and conditional affine transformation of feature activations. All these conditioning mechanisms, however, are uni-directional, i.e., no information flow from the input image back to the guidance. To better utilize the constraints of the guidance image, we present a bi-directional feature transformation (bFT) scheme. We show that our bFT scheme outperforms other conditioning schemes and has comparable results to state-of-the-art methods on different tasks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Depth Super-Resolution | NYU v2 (test) | RMSE3.35 | 126 | |
| Person Image Generation | DeepFashion (test) | SSIM0.767 | 19 | |
| Depth Upsampling | NYU V2 | RMSE (x4)3.35 | 11 | |
| Person Image Synthesis | DeepFashion (test) | SSIM0.767 | 10 | |
| Pose Transfer | DeepFashion Full (test) | SSIM0.767 | 5 | |
| Pose Transfer | DeepFashion Modified (test) | SSIM0.771 | 4 | |
| Texture Transfer | Handbag Dataset (test) | LPIPS0.161 | 3 | |
| Texture Transfer | Clothes Dataset (test) | LPIPS0.067 | 3 | |
| Texture Transfer | Shoes Dataset (test) | LPIPS0.124 | 3 |