HOI-Swap: Swapping Objects in Videos with Hand-Object Interaction Awareness
About
We study the problem of precisely swapping objects in videos, with a focus on those interacted with by hands, given one user-provided reference object image. Despite the great advancements that diffusion models have made in video editing recently, these models often fall short in handling the intricacies of hand-object interactions (HOI), failing to produce realistic edits -- especially when object swapping results in object shape or functionality changes. To bridge this gap, we present HOI-Swap, a novel diffusion-based video editing framework trained in a self-supervised manner. Designed in two stages, the first stage focuses on object swapping in a single frame with HOI awareness; the model learns to adjust the interaction patterns, such as the hand grasp, based on changes in the object's properties. The second stage extends the single-frame edit across the entire sequence; we achieve controllable motion alignment with the original video by: (1) warping a new sequence from the stage-I edited frame based on sampled motion points and (2) conditioning video generation on the warped sequence. Comprehensive qualitative and quantitative evaluations demonstrate that HOI-Swap significantly outperforms existing methods, delivering high-quality video edits with realistic HOIs.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Cross-Object Reenactment | Custom Dataset (Ours) (test) | Hand Fidelity99.4 | 6 | |
| Video Reenactment | User Study | HOI Consistency76 | 6 | |
| Self-Reenactment | Custom Dataset (test) | PSNR31.634 | 6 | |
| Cross-Object Reenactment | HOI4D (test) | Hand Fidelity99.4 | 4 | |
| Image Editing | HOI4D and Ego-Exo4D Image (test) | Contact Agreement87.9 | 4 | |
| Video Editing | HOI4D, Ego-Exo4D, EPIC-Kitchens, and TCN Pouring Video Evaluation Set (test) | Subject Consistency0.924 | 4 | |
| Self-Reenactment | HOI4D (test) | PSNR31.528 | 4 |