Mistral-NeMo-Minitron-8B-ARChitects-Full-bnb-4bit

Model Overview

Mistral-NeMo-Minitron-8B-ARChitects-Full-bnb-4bit is a retrained variant of Nvidia Mistral-NeMo-Minitron-8B-Base, finetuned specifically to solve ARC-AGI tasks. In order to save GPU memory, the embedding and vocabulary size have been reduced to only 77 tokens. The model achieved a score of 53.5 on the ARC-AGI private evaluation set during the Kaggle ARC Prize 2024 Competition. Note that the ARC-AGI public evaluation set was used as training data for this model. Please refer to our paper for more details. For more models tuned for ARC-AGI, check out our model collection.

Finetuning Datasets

This model was finetuned on the following datasets:

License

This model is released under the NVIDIA Open Model License Agreement.

Usage

This model can be used with the transformers or unsloth packages. For more information on preprocessing the ARC Prize tasks to generate prompts for the model, please refer to our Paper and our github repositiory.

References

Downloads last month
5,112
Safetensors
Model size
3.79B params
Tensor type
F32
·
U8
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Collection including da-fr/Mistral-NeMo-Minitron-8B-ARChitects-Full-bnb-4bit