Text-to-Video
Diffusers
Safetensors
WanDMDPipeline
PY007 commited on
Commit
51fd951
·
verified ·
1 Parent(s): 581758e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -0
README.md CHANGED
@@ -38,6 +38,36 @@ FastWan2.1-T2V-1.3B-Diffusers is built upon Wan-AI/Wan2.1-T2V-1.3B-Diffusers. It
38
  - [1 Node/GPU debugging finetuning script](https://github.com/hao-ai-lab/FastVideo/blob/main/scripts/distill/v1_distill_dmd_wan_VSA.sh)
39
  - [Slurm training example script](https://github.com/hao-ai-lab/FastVideo/blob/main/examples/distill/Wan2.1-T2V/Wan-Syn-Data-480P/distill_dmd_VSA_t2v_1.3B.slurm)
40
  - [Inference script](https://github.com/hao-ai-lab/FastVideo/blob/main/scripts/inference/v1_inference_wan_dmd.sh)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
  - Try it out on **FastVideo** — we support a wide range of GPUs from **H100** to **4090**, and also support **Mac** users!
42
 
43
  ### Training Infrastructure
 
38
  - [1 Node/GPU debugging finetuning script](https://github.com/hao-ai-lab/FastVideo/blob/main/scripts/distill/v1_distill_dmd_wan_VSA.sh)
39
  - [Slurm training example script](https://github.com/hao-ai-lab/FastVideo/blob/main/examples/distill/Wan2.1-T2V/Wan-Syn-Data-480P/distill_dmd_VSA_t2v_1.3B.slurm)
40
  - [Inference script](https://github.com/hao-ai-lab/FastVideo/blob/main/scripts/inference/v1_inference_wan_dmd.sh)
41
+ ```python
42
+ # install FastVideo and VSA first
43
+ git clone https://github.com/hao-ai-lab/FastVideo
44
+ pip install -e .
45
+ cd csrc/attn
46
+ git submodule update --init --recursive
47
+ python setup_vsa.py install
48
+
49
+ num_gpus=1
50
+ export FASTVIDEO_ATTENTION_BACKEND=VIDEO_SPARSE_ATTN
51
+ export MODEL_BASE=FastVideo/FastWan2.1-T2V-1.3B-Diffusers
52
+ # export MODEL_BASE=hunyuanvideo-community/HunyuanVideo
53
+ # You can either use --prompt or --prompt-txt, but not both.
54
+ fastvideo generate \
55
+ --model-path $MODEL_BASE \
56
+ --sp-size $num_gpus \
57
+ --tp-size 1 \
58
+ --num-gpus $num_gpus \
59
+ --height 480 \
60
+ --width 848 \
61
+ --num-frames 81 \
62
+ --num-inference-steps 3 \
63
+ --fps 16 \
64
+ --prompt-txt assets/prompt.txt \
65
+ --negative-prompt "Bright tones, overexposed, static, blurred details, subtitles, style, works, paintings, images, static, overall gray, worst quality, low quality, JPEG compression residue, ugly, incomplete, extra fingers, poorly drawn hands, poorly drawn faces, deformed, disfigured, misshapen limbs, fused fingers, still picture, messy background, three legs, many people in the background, walking backwards" \
66
+ --seed 1024 \
67
+ --output-path outputs_video_dmd/ \
68
+ --VSA-sparsity 0.8 \
69
+ --dmd-denoising-steps "1000,757,522"
70
+ ```
71
  - Try it out on **FastVideo** — we support a wide range of GPUs from **H100** to **4090**, and also support **Mac** users!
72
 
73
  ### Training Infrastructure