Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
base_model:
|
| 4 |
+
- openai/gpt-oss-20b
|
| 5 |
+
tags:
|
| 6 |
+
- mlx
|
| 7 |
+
- gpt
|
| 8 |
+
- openai
|
| 9 |
+
- chatGPT
|
| 10 |
+
---
|
| 11 |
+
|
| 12 |
+
# NexaAI/gpt-oss-20b-MLX-4bit
|
| 13 |
+
|
| 14 |
+
## Quickstart
|
| 15 |
+
|
| 16 |
+
Run them directly with [nexa-sdk](https://github.com/NexaAI/nexa-sdk) installed
|
| 17 |
+
In nexa-sdk CLI:
|
| 18 |
+
|
| 19 |
+
```bash
|
| 20 |
+
NexaAI/gpt-oss-20b-MLX-4bit
|
| 21 |
+
```
|
| 22 |
+
|
| 23 |
+
## Overview
|
| 24 |
+
This is a 4-bit quantized version of the OpenAI GPT OSS 20B model, optimized for Apple Silicon using the MLX framework. The model was successfully converted from the original `gpt_oss` architecture to MLX format using the development version of `mlx-lm` (v0.26.3).
|
| 25 |
+
|
| 26 |
+
>[!TIP] Tip: For best results with tool calling and reasoning, update your LMStudio to the latest version (0.3.22)
|
| 27 |
+
|
| 28 |
+
|
| 29 |
+
## Reference
|
| 30 |
+
**Original model card**: [InferenceIllusionist/gpt-oss-20b-MLX-4bit](https://huggingface.co/InferenceIllusionist/gpt-oss-20b-MLX-4bit)
|