Pruned Qwen (Epoch 1)

This is ToastyPigeon/qwen2.5-32b-unnamed-test-model pruned down from 32b -> 27b.

Using PruneMe to find layers to remove resulted in the removal of layers [25, 29) and [36, 43) for a reduction from 64 -> 52 layers.

Trained on 1 epoch of mixed data from the datasets that went into the pre-pruned model (I'll document that later), totaling about ~10M tokens so far of retraining.

Coherent but a little dumb. Will likely run a second epoch on this.

Downloads last month
6
Safetensors
Model size
26.9B params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for Columbidae/Qwen-27B-Pruned-Retrained

Base model

Qwen/Qwen2.5-32B
Finetuned
(126)
this model
Quantizations
1 model