Text-to-Video
Diffusers
Safetensors
WanDMDPipeline
BrianChen1129 commited on
Commit
a4c14c2
·
verified ·
1 Parent(s): cd4a0b3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -24,8 +24,9 @@ base_model:
24
 
25
 
26
  ## Introduction
 
27
 
28
- This model is jointly finetuned with [DMD](https://arxiv.org/pdf/2405.14867) and [VSA](https://arxiv.org/pdf/2505.13389), based on [Wan-AI/Wan2.1-T2V-1.3B-Diffusers](https://huggingface.co/Wan-AI/Wan2.1-T2V-1.3B-Diffusers). It supports efficient 3-step inference and generates high-quality videos at **61×448×832** resolution. We adopt the [FastVideo 480P Synthetic Wan dataset](https://huggingface.co/datasets/FastVideo/Wan-Syn_77x448x832_600k), consisting of 600k synthetic latents.
29
 
30
  ---
31
 
 
24
 
25
 
26
  ## Introduction
27
+ We're excited to introduce the FastWan2.1 series—a new line of models finetuned with our novel **Sparse-distill** strategy. This approach jointly integrates DMD and VSA in a single training process, combining the benefits of distillation to shorten diffusion steps and sparse attention to reduce attention computations, enabling even faster video generation.
28
 
29
+ FastWan2.1-T2V-1.3B-Diffusers is built upon Wan-AI/Wan2.1-T2V-1.3B-Diffusers. It supports efficient 3-step inference and produces high-quality videos at 61×448×832 resolution. For training, we use the FastVideo 480P Synthetic Wan dataset, which contains 600k synthetic latents.
30
 
31
  ---
32