Model Card for cloudyu/Mixtral-8x7B-Instruct-v0.1-DPO

Metrics improved by Truthful DPO traingin after 100 steps Metrsc improment

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 73.44
AI2 Reasoning Challenge (25-Shot) 69.80
HellaSwag (10-Shot) 87.83
MMLU (5-Shot) 71.05
TruthfulQA (0-shot) 69.18
Winogrande (5-shot) 81.37
GSM8k (5-shot) 61.41
Downloads last month
805
Safetensors
Model size
47B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for cloudyu/Mixtral-8x7B-Instruct-v0.1-DPO

Quantizations
1 model

Evaluation results