Update README.md
Browse files
README.md
CHANGED
@@ -233,7 +233,7 @@ import torch
|
|
233 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
234 |
from auto_round import AutoRound
|
235 |
|
236 |
-
model_name = "/
|
237 |
|
238 |
model = AutoModelForCausalLM.from_pretrained(model_name,
|
239 |
device_map="cpu", torch_dtype="auto")
|
@@ -246,7 +246,7 @@ for n, m in model.named_modules():
|
|
246 |
layer_config[n] = {"bits": 4}
|
247 |
|
248 |
autoround = AutoRound(model, tokenizer, iters=0, layer_config=layer_config, nsamples=512, dataset="github-code-clean")
|
249 |
-
autoround.quantize_and_save("
|
250 |
|
251 |
```
|
252 |
|
|
|
233 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
234 |
from auto_round import AutoRound
|
235 |
|
236 |
+
model_name = "Qwen/Qwen3-Coder-480B-A35B-Instruct"
|
237 |
|
238 |
model = AutoModelForCausalLM.from_pretrained(model_name,
|
239 |
device_map="cpu", torch_dtype="auto")
|
|
|
246 |
layer_config[n] = {"bits": 4}
|
247 |
|
248 |
autoround = AutoRound(model, tokenizer, iters=0, layer_config=layer_config, nsamples=512, dataset="github-code-clean")
|
249 |
+
autoround.quantize_and_save("./Qwen3-Coder-480B-A35B-Instruct-q2ks", format="gguf:q2_k_s")
|
250 |
|
251 |
```
|
252 |
|