GGUF
conversational
wenhuach commited on
Commit
0fc06d4
·
verified ·
1 Parent(s): 0c21629

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -233,7 +233,7 @@ import torch
233
  from transformers import AutoModelForCausalLM, AutoTokenizer
234
  from auto_round import AutoRound
235
 
236
- model_name = "/data2/Qwen3-Coder-480B-A35B-Instruct"
237
 
238
  model = AutoModelForCausalLM.from_pretrained(model_name,
239
  device_map="cpu", torch_dtype="auto")
@@ -246,7 +246,7 @@ for n, m in model.named_modules():
246
  layer_config[n] = {"bits": 4}
247
 
248
  autoround = AutoRound(model, tokenizer, iters=0, layer_config=layer_config, nsamples=512, dataset="github-code-clean")
249
- autoround.quantize_and_save("/dataset/Qwen3-Coder-480B-A35B-Instruct-q2ks", format="gguf:q2_k_s")
250
 
251
  ```
252
 
 
233
  from transformers import AutoModelForCausalLM, AutoTokenizer
234
  from auto_round import AutoRound
235
 
236
+ model_name = "Qwen/Qwen3-Coder-480B-A35B-Instruct"
237
 
238
  model = AutoModelForCausalLM.from_pretrained(model_name,
239
  device_map="cpu", torch_dtype="auto")
 
246
  layer_config[n] = {"bits": 4}
247
 
248
  autoround = AutoRound(model, tokenizer, iters=0, layer_config=layer_config, nsamples=512, dataset="github-code-clean")
249
+ autoround.quantize_and_save("./Qwen3-Coder-480B-A35B-Instruct-q2ks", format="gguf:q2_k_s")
250
 
251
  ```
252