Skip to content

HiDream-ai/VTON-VLLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

VTON-VLLM: Aligning Virtual Try-On Models with Human Preferences (NeurIPS 2025)

Siqi Wan1, Jingwen Chen2, Qi Cai2, Yingwei Pan2, Ting Yao2, Tao Mei2
1University of Science and Technology of China; 2HiDream.ai Inc

This is the official repository for the NeurIPS 2025 paper "VTON-VLLM: Aligning Virtual Try-On Models with Human Preferences"

Overview

We novelly propose a vision large language model, namely VTON-VLLM, functions as a unified “fashion expert” and is capable of both evaluating and steering VTON synthesis towards human preferences. VTON-VLLM upgrades VTON model through two pivotal ways: (1) providing fine-grained supervisory signals during the training of a plug-and-play VTON refinement model, and (2) enabling adaptive and preference-aware test-time scaling at inference. To benchmark VTON models more holistically, we introduce VITON-Bench, a challenging test suite of complex try-on scenarios, and human-preference–aware metrics.

Installation

Create a conda environment & Install requirments

conda create -n VTON-VLLM python==3.9.0
conda activate VTON-VLLM
cd VTON-VLLM-main 
pip install -r requirements.txt

VTON-VLLM

You can directly download the VTON-VLLM or follow the instructions in preprocessing.md to extract the Semantic Point Feature yourself.

VTON Refinement Model

Inference

Please download the pre-trained model from Link.

sh src/inference.sh

Train

sh src/train_VTON_refinement_model.sh

Human-Preference–Aware Metrics

sh metrics/vllm_metrics.py

Acknowledgement

Thanks the contribution of LLaMA-Factory and CAT-VTON.

About

official repository for the NeurIPS 2025 paper "VTON-VLLM: Aligning Virtual Try-On Models with Human Preferences"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages