license: apache-2.0 | |
base_model: | |
- nvidia/OpenReasoning-Nemotron-7B | |
datasets: | |
- mit-han-lab/pile-val-backup | |
# OpenReasoning-Nemotron-7B-AWQ | |
## Method | |
Quantised using [vllm-project/llm-compressor](https://github.com/vllm-project/llm-compressor.git) and the following configs: | |
``` | |
recipe = [ | |
AWQModifier(ignore=["lm_head"], scheme="W4A16_ASYM", targets=["Linear"]), | |
] | |
``` |