UniGarmentManip: A Unified Framework for Category-Level Garment Manipulation via Dense Visual Correspondence
About
Garment manipulation (e.g., unfolding, folding and hanging clothes) is essential for future robots to accomplish home-assistant tasks, while highly challenging due to the diversity of garment configurations, geometries and deformations. Although able to manipulate similar shaped garments in a certain task, previous works mostly have to design different policies for different tasks, could not generalize to garments with diverse geometries, and often rely heavily on human-annotated data. In this paper, we leverage the property that, garments in a certain category have similar structures, and then learn the topological dense (point-level) visual correspondence among garments in the category level with different deformations in the self-supervised manner. The topological correspondence can be easily adapted to the functional correspondence to guide the manipulation policies for various downstream tasks, within only one or few-shot demonstrations. Experiments over garments in 3 different categories on 3 representative tasks in diverse scenarios, using one or two arms, taking one or more steps, inputting flat or messy garments, demonstrate the effectiveness of our proposed method. Project page: https://warshallrho.github.io/unigarmentmanip.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Folding | Garments Dress | Fold-FLAT84 | 3 | |
| Folding | Garments (Trouser) | Fold-FLAT83.3 | 3 | |
| Hanging | Garments Hang-RAND (test) | Top Success Rate81.9 | 3 | |
| Hanging | Garments (Hang-FLING) (test) | Top Success Rate83.8 | 3 | |
| Unfolding | CLOTH3D Unfold-RAND (test) | Success Rate (Tops)83.6 | 3 | |
| Unfolding | CLOTH3D Unfold-DROP (test) | Success Rate (Tops)85.3 | 3 | |
| Folding | Garments (Tops) | Fold-FLAT0.835 | 3 |