File size: 731 Bytes
1769f91
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
---
base_model: Qwen/Qwen3-0.6B
---

[EXL3](https://github.com/turboderp-org/exllamav3) quantization of [Qwen3-0.6B](https://huggingface.co/Qwen/Qwen3-0.6B), 8 bits per weight, including output layers.

### HumanEval (argmax)

| Model                                                                                      | Q4   | Q6    | Q8    | FP16  |
| ------------------------------------------------------------------------------------------ | ---- | ----- | ----- | ----- |
| [Qwen3-0.6B-exl3-8bpw-h8](https://huggingface.co/isogen/Qwen3-0.6B-exl3-8bpw-h8)           | 0.0% | 38.4% | 40.9% | 40.2% |
| [Qwen3-0.6B-Base-exl3-8bpw-h8](https://huggingface.co/isogen/Qwen3-0.6B-Base-exl3-8bpw-h8) | 0.0% | 36.0% | 37.2% | 36.6% |