Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ base_model:
|
|
15 |
|
16 |
# Qwen2.5-VL-72B-Instruct-Pointer-AWQ
|
17 |
|
18 |
-
|
19 |
|
20 |
# Qwen2.5-VL-72B-Instruct
|
21 |
<a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;">
|
|
|
15 |
|
16 |
# Qwen2.5-VL-72B-Instruct-Pointer-AWQ
|
17 |
|
18 |
+
Since the official `Qwen/Qwen2.5-VL-72B-Instruct-AWQ` doesn't work with tensor parallel on vllm yet, this model fixed the issue and supports `--tensor-parallel` with 2, 4 or 8 GPUs. Use `vllm==0.7.3`.
|
19 |
|
20 |
# Qwen2.5-VL-72B-Instruct
|
21 |
<a href="https://chat.qwenlm.ai/" target="_blank" style="margin: 2px;">
|