Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
24 commits
Select commit Hold shift + click to select a range
26a4e8d
add time cost log for different stages
SamitHuang Mar 6, 2026
b3b70a8
reduce hop3 overhead
SamitHuang Mar 6, 2026
104d71c
perf: reduce IPC overhead for single-stage diffusion serving
SamitHuang Mar 6, 2026
bf2ddb0
perf: reduce IPC overhead for single-stage diffusion serving (~6.5s, …
SamitHuang Mar 6, 2026
735b2ca
Merge branch 'main' into main
SamitHuang Mar 6, 2026
dd4468c
fix conflicts
SamitHuang Mar 6, 2026
ff62a1e
rm redundancy
SamitHuang Mar 9, 2026
870963e
Merge branch 'main' into main
SamitHuang Mar 9, 2026
5414a42
rm logs
SamitHuang Mar 9, 2026
e3dec54
fix inline
SamitHuang Mar 9, 2026
2cd9f9f
fix ci
SamitHuang Mar 9, 2026
172040a
fix ci
SamitHuang Mar 9, 2026
0a86fc5
fix log
SamitHuang Mar 9, 2026
9b9c597
fix
SamitHuang Mar 9, 2026
bda0f2d
fix log
SamitHuang Mar 9, 2026
3452ad3
Merge branch 'main' of https://github.com/samithuang/vllm-omni
SamitHuang Mar 9, 2026
9ab7c55
Merge remote-tracking branch 'upstream/main'
SamitHuang Mar 12, 2026
c32a78a
Merge remote-tracking branch 'upstream/main'
SamitHuang Mar 12, 2026
91afe27
Merge branch 'main' of https://github.com/samithuang/vllm-omni
SamitHuang Mar 13, 2026
9b783a6
Merge branch 'main' of https://github.com/samithuang/vllm-omni
SamitHuang Mar 16, 2026
5c71c89
[Docs] Add Wan2.1-T2V-1.3B as supported video generation model
SamitHuang Mar 16, 2026
17e21dd
Add Wan2.1 model identifier for 14B Diffusers
SamitHuang Mar 19, 2026
8e6a284
docs: add Wan2.1-T2V to diffusion_acceleration and supported_models
SamitHuang Mar 20, 2026
a7684d4
Merge branch 'main' into docs/add-wan21-support
SamitHuang Mar 20, 2026
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions docs/models/supported_models.md
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ th {
| `QwenImageLayeredPipeline` | Qwen-Image-Layered | `Qwen/Qwen-Image-Layered` |
| `GlmImagePipeline` | GLM-Image | `zai-org/GLM-Image` |
|`ZImagePipeline` | Z-Image | `Tongyi-MAI/Z-Image-Turbo` |
| `WanPipeline` | Wan2.2-T2V, Wan2.2-TI2V | `Wan-AI/Wan2.2-T2V-A14B-Diffusers`, `Wan-AI/Wan2.2-TI2V-5B-Diffusers` |
| `WanPipeline` | Wan2.1-T2V, Wan2.2-T2V, Wan2.2-TI2V | `Wan-AI/Wan2.1-T2V-1.3B-Diffusers`, `Wan-AI/Wan2.1-T2V-14B-Diffusers`, `Wan-AI/Wan2.2-T2V-A14B-Diffusers`, `Wan-AI/Wan2.2-TI2V-5B-Diffusers` |
| `WanImageToVideoPipeline` | Wan2.2-I2V | `Wan-AI/Wan2.2-I2V-A14B-Diffusers` |
| `LTX2Pipeline` | LTX-2-T2V | `Lightricks/LTX-2` |
| `LTX2ImageToVideoPipeline` | LTX-2-I2V | `Lightricks/LTX-2` |
Expand Down Expand Up @@ -76,7 +76,7 @@ th {
| `QwenImageLayeredPipeline` | Qwen-Image-Layered | `Qwen/Qwen-Image-Layered` |
| `QwenImageEditPlusPipeline` | Qwen-Image-Edit-2511 | `Qwen/Qwen-Image-Edit-2511` |
|`ZImagePipeline` | Z-Image | `Tongyi-MAI/Z-Image-Turbo` |
| `WanPipeline` | Wan2.2-T2V, Wan2.2-TI2V | `Wan-AI/Wan2.2-T2V-A14B-Diffusers`, `Wan-AI/Wan2.2-TI2V-5B-Diffusers` |
| `WanPipeline` | Wan2.1-T2V, Wan2.2-T2V, Wan2.2-TI2V | `Wan-AI/Wan2.1-T2V-1.3B-Diffusers`, `Wan-AI/Wan2.1-T2V-14B-Diffusers`, `Wan-AI/Wan2.2-T2V-A14B-Diffusers`, `Wan-AI/Wan2.2-TI2V-5B-Diffusers` |
| `WanImageToVideoPipeline` | Wan2.2-I2V | `Wan-AI/Wan2.2-I2V-A14B-Diffusers` |
|`LongcatImagePipeline` | LongCat-Image | `meituan-longcat/LongCat-Image` |
|`LongCatImageEditPipeline` | LongCat-Image-Edit | `meituan-longcat/LongCat-Image-Edit` |
Expand Down
2 changes: 2 additions & 0 deletions docs/user_guide/diffusion/parallelism_acceleration.md
Original file line number Diff line number Diff line change
Expand Up @@ -58,6 +58,8 @@ The following table shows which models are currently supported by parallelism me

| Model | Model Identifier | Ulysses-SP | Ring-Attention | Tensor-Parallel | HSDP | VAE-Patch-Parallel |
|-------|------------------|:----------:|:--------------:|:---------------:|:----:| :----:|
| **Wan2.1** | `Wan-AI/Wan2.1-T2V-1.3B-Diffusers` | ✅ | ✅ | ✅ | ✅ | ✅ |
| **Wan2.1** | `Wan-AI/Wan2.1-T2V-14B-Diffusers` | ✅ | ✅ | ✅ | ✅ | ✅ |
| **Wan2.2** | `Wan-AI/Wan2.2-T2V-A14B-Diffusers` | ✅ | ✅ | ✅ | ✅ | ✅ |
| **LTX-2** | `Lightricks/LTX-2` | ✅ | ✅ | ✅ | ❌ | ❌ |

Expand Down
2 changes: 2 additions & 0 deletions docs/user_guide/diffusion_acceleration.md
Original file line number Diff line number Diff line change
Expand Up @@ -76,6 +76,8 @@ The following table shows which models are currently supported by each accelerat

| Model | Model Identifier | TeaCache | Cache-DiT | Ulysses-SP | Ring-Attention | CFG-Parallel | HSDP | VAE-Patch-Parallel |
|-------|------------------|:--------:|:---------:|:----------:|:--------------:|:------------:|:----:|:----:|
| **Wan2.1-T2V** | `Wan-AI/Wan2.1-T2V-1.3B-Diffusers` | ❌ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
| **Wan2.1-T2V** | `Wan-AI/Wan2.1-T2V-14B-Diffusers` | ❌ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
| **Wan2.2** | `Wan-AI/Wan2.2-T2V-A14B-Diffusers` | ❌ | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
| **LTX-2** | `Lightricks/LTX-2` | ❌ | ✅ | ✅ | ✅ | ✅ | ❌ | ❌ |
| **DreamID-Omni** | `XuGuo699/DreamID-Omni` | ❌ | ❌ | ❌ | ❌ | ✅ | ❌ | ❌ |
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,10 @@
Source <https://github.com/vllm-project/vllm-omni/tree/main/examples/offline_inference/text_to_video>.


The `Wan-AI/Wan2.2-T2V-A14B-Diffusers` pipeline generates short videos from text prompts.
The Wan T2V pipeline generates short videos from text prompts. Supported models include:

- `Wan-AI/Wan2.2-T2V-A14B-Diffusers` (14B, requires ~48GB VRAM)
- `Wan-AI/Wan2.1-T2V-1.3B-Diffusers` (1.3B, requires ~16GB VRAM, use `--flow-shift 3.0 --boundary-ratio 0.0`)

## Local CLI Usage

Expand Down
Loading