--- license: apache-2.0 base_model: - nvidia/OpenReasoning-Nemotron-7B datasets: - mit-han-lab/pile-val-backup --- # OpenReasoning-Nemotron-7B-AWQ ## Method Quantised using [vllm-project/llm-compressor](https://github.com/vllm-project/llm-compressor.git) and the following configs: ``` recipe = [ AWQModifier(ignore=["lm_head"], scheme="W4A16_ASYM", targets=["Linear"]), ] ```