back to home

levihsu / OOTDiffusion

[AAAI 2025] Official implementation of "OOTDiffusion: Outfitting Fusion based Latent Diffusion for Controllable Virtual Try-on"

6,516 stars
946 forks
87 issues
PythonCudaC++

AI Architecture Analysis

This repository is indexed by RepoMind. By analyzing levihsu/OOTDiffusion in our AI interface, you can instantly generate complete architecture diagrams, visualize control flows, and perform automated security audits across the entire codebase.

Our Agentic Context Augmented Generation (Agentic CAG) engine loads full source files into context, avoiding the fragmentation of traditional RAG systems. Ask questions about the architecture, dependencies, or specific features to see it in action.

Embed this Badge

Showcase RepoMind's analysis directly in your repository's README.

[![Analyzed by RepoMind](https://img.shields.io/badge/Analyzed%20by-RepoMind-4F46E5?style=for-the-badge)](https://repomind-ai.vercel.app/repo/levihsu/OOTDiffusion)
Preview:Analyzed by RepoMind

Repository Summary (README)

Preview

OOTDiffusion

This repository is the official implementation of OOTDiffusion

🤗 Try out OOTDiffusion

(Thanks to ZeroGPU for providing A100 GPUs)

<!-- Or [try our own demo](https://ootd.ibot.cn/) on RTX 4090 GPUs -->

OOTDiffusion: Outfitting Fusion based Latent Diffusion for Controllable Virtual Try-on [arXiv paper]<br> Yuhao Xu, Tao Gu, Weifeng Chen, Chengcai Chen<br> Xiao-i Research

Our model checkpoints trained on VITON-HD (half-body) and Dress Code (full-body) have been released

  • 🤗 Hugging Face link for checkpoints (ootd, humanparsing, and openpose)
  • 📢📢 We support ONNX for humanparsing now. Most environmental issues should have been addressed : )
  • Please also download clip-vit-large-patch14 into checkpoints folder
  • We've only tested our code and models on Linux (Ubuntu 22.04)

demo  workflow 

Installation

  1. Clone the repository
git clone https://github.com/levihsu/OOTDiffusion
  1. Create a conda environment and install the required packages
conda create -n ootd python==3.10
conda activate ootd
pip install torch==2.0.1 torchvision==0.15.2 torchaudio==2.0.2
pip install -r requirements.txt

Inference

  1. Half-body model
cd OOTDiffusion/run
python run_ootd.py --model_path <model-image-path> --cloth_path <cloth-image-path> --scale 2.0 --sample 4
  1. Full-body model

Garment category must be paired: 0 = upperbody; 1 = lowerbody; 2 = dress

cd OOTDiffusion/run
python run_ootd.py --model_path <model-image-path> --cloth_path <cloth-image-path> --model_type dc --category 2 --scale 2.0 --sample 4

Citation

@article{xu2024ootdiffusion,
  title={OOTDiffusion: Outfitting Fusion based Latent Diffusion for Controllable Virtual Try-on},
  author={Xu, Yuhao and Gu, Tao and Chen, Weifeng and Chen, Chengcai},
  journal={arXiv preprint arXiv:2403.01779},
  year={2024}
}

Star History

Star History Chart

TODO List

  • Paper
  • Gradio demo
  • Inference code
  • Model weights
  • Training code