Model Card for Model ID

The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. This model uses peft finetuning with NEFTune for robustness.

Model Details

Model Description

This model is a finetuned model of the mistralai/Mistral-7B-v0.1.

Training Details

Training Data

This model is finetuned with kaist-ai/CoT-Collection.

Training Procedure

This model trained with SFT trainer and NEFTune method. (According to the paper, NEFTune adds noise to the embedding vectors during training)

Training Hyperparameters

  • lora alpha: 16
  • lora r: 64
  • lora dropout: 0.05
  • max sequence length: 4096
  • learning rate: 2e-4
  • max_grad_norm: 0.3
  • weight_decay: 0.001
  • gradient checkpoint: True
  • optim: paged_adamw_32bit
  • use_bf16: True
  • use_4bit: True
  • use_nested_quant: False
  • bnb_4bit_compute_dtype: float16
  • bnb_4bit_quant_type: nf4
Downloads last month
74
Safetensors
Model size
3.86B params
Tensor type
F32
FP16
U8
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.