🤖 AI Summary
Single-image-driven 3D geometric reconstruction of loose clothing remains highly challenging. This paper introduces the first diffusion-based mapping framework grounded in pattern coordinates, jointly modeling image pixels, UV coordinates, and 3D mesh geometry. We propose an Implicit Segmentation Pattern (ISP) representation in UV space to encode garment structure and integrate it with a generative diffusion model to learn geometric priors. A cross-domain differentiable mapping network is designed to enable end-to-end optimization from image → UV segmentation pattern → 3D mesh. Crucially, our method is trained exclusively on synthetic cloth data yet generalizes robustly to real-world images. Quantitative and qualitative evaluations demonstrate significant improvements over state-of-the-art methods on both tight-fitting and loose garments. Reconstructed geometries are physically plausible, rich in fine-scale detail, and support downstream applications including pose retargeting and texture editing.
📝 Abstract
Reconstructing 3D clothed humans from images is fundamental to applications like virtual try-on, avatar creation, and mixed reality. While recent advances have enhanced human body recovery, accurate reconstruction of garment geometry -- especially for loose-fitting clothing -- remains an open challenge. We present a novel method for high-fidelity 3D garment reconstruction from single images that bridges 2D and 3D representations. Our approach combines Implicit Sewing Patterns (ISP) with a generative diffusion model to learn rich garment shape priors in a 2D UV space. A key innovation is our mapping model that establishes correspondences between 2D image pixels, UV pattern coordinates, and 3D geometry, enabling joint optimization of both 3D garment meshes and the corresponding 2D patterns by aligning learned priors with image observations. Despite training exclusively on synthetically simulated cloth data, our method generalizes effectively to real-world images, outperforming existing approaches on both tight- and loose-fitting garments. The reconstructed garments maintain physical plausibility while capturing fine geometric details, enabling downstream applications including garment retargeting and texture manipulation.