Update README.md
Browse files
README.md
CHANGED
@@ -43,7 +43,7 @@ This model is jointly finetuned with [DMD](https://arxiv.org/pdf/2405.14867) and
|
|
43 |
Training was conducted on **4 nodes with 32 H200 GPUs** in total, using a `global batch size = 64`.
|
44 |
We enable `gradient checkpointing`, set `gradient_accumulation_steps=2`, and use `learning rate = 1e-5`.
|
45 |
We set **VSA attention sparsity** to 0.8, and training runs for **4000 steps (~12 hours)**
|
46 |
-
The detailed training example script is available [here](https://github.com/hao-ai-lab/FastVideo/blob/main/examples/distill/Wan-Syn-480P/distill_dmd_VSA_t2v.slurm).
|
47 |
|
48 |
|
49 |
|
|
|
43 |
Training was conducted on **4 nodes with 32 H200 GPUs** in total, using a `global batch size = 64`.
|
44 |
We enable `gradient checkpointing`, set `gradient_accumulation_steps=2`, and use `learning rate = 1e-5`.
|
45 |
We set **VSA attention sparsity** to 0.8, and training runs for **4000 steps (~12 hours)**
|
46 |
+
The detailed **training example script** is available [here](https://github.com/hao-ai-lab/FastVideo/blob/main/examples/distill/Wan-Syn-480P/distill_dmd_VSA_t2v.slurm).
|
47 |
|
48 |
|
49 |
|