🤖 AI Summary
This work proposes Neural Clothing Tryer, a novel framework that addresses the limited flexibility of existing virtual try-on systems in controlling human appearance, pose, and attributes. For the first time, it integrates semantic descriptions and vision-language aligned features into the virtual try-on task. By combining a controllable diffusion model, a semantic enhancement module, and multimodal conditional inputs, the method achieves high-fidelity preservation of garment details alongside fine-grained human editing capabilities. Experimental results demonstrate that the proposed framework significantly outperforms state-of-the-art approaches on public benchmarks, enabling users to customize digital avatars and accurately overlay specified garments, thereby delivering a high-quality, customizable, and multi-attribute virtual try-on experience.