GPT-OSS-20B fine-tuned on adamo1139/HESOYAM_v0.4 dataset, 1 epoch, chatml format that erases reasoning. 1024 rank, 128 alpha QLoRA made with Unsloth. It will be undergoing further preference alignment once some issues preventing me from doing it right now will be patched out.

Total batch size 16, learning rate 0.0002 with cosine schedule, with sample packing enabled. Training took about 8 hours on single 3090 Ti.

Loss curve looks a bit underwhelming.

image/png

I tried merging this lora with the huizimao/gpt-oss-20b-uncensored-mxfp4 but that wasn't producing great effects.

No reasoning is present, and model definitely learns something from the dataset, but it feels pretty dumb, so it could be a wrong path.

Downloads last month
896
GGUF
Model size
20.9B params
Architecture
gpt-oss
Hardware compatibility
Log In to view the estimation

4-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for adamo1139/GPT-OSS-20B-HESOYAM-1108-WIP-CHATML-GGUF

Base model

openai/gpt-oss-20b
Quantized
(73)
this model

Dataset used to train adamo1139/GPT-OSS-20B-HESOYAM-1108-WIP-CHATML-GGUF