Qwen2.5-32B-Instruct-CFT

Introduction

Qwen2.5-32B-Instruct-CFT is a 32B parameter model fine-tuned using our novel Critique Fine-Tuning (CFT) approach. Built upon the Qwen2.5-32B-Instruct base model, this variant is trained to critique and analyze responses rather than simply imitate them, leading to enhanced reasoning capabilities.

Key Features

  • Built on the powerful Qwen2.5-32B-Instruct foundation
  • Trained using Critique Fine-Tuning (CFT) methodology
  • Highly efficient training with minimal data requirements
  • Inherits the strong instruction-following capabilities of the base model

Training Details

Training Data

  • Dataset: WebInstruct-CFT-4K
  • Training format: (input=[query; noisy response], output=critique)
  • Teacher model: GPT-4o for generating critiques

Training Infrastructure

  • Framework: LLaMA-Factory
  • Hardware: 8x NVIDIA H100 GPUs
  • Training time: ~1.5 hours with DeepSpeed Zero-3

For more details about the model architecture, methodology, and comprehensive evaluation results, please visit our project webpage.

Downloads last month
185
Safetensors
Model size
32.8B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for TIGER-Lab/Qwen2.5-32B-Instruct-CFT

Base model

Qwen/Qwen2.5-32B
Finetuned
(126)
this model
Quantizations
3 models

Dataset used to train TIGER-Lab/Qwen2.5-32B-Instruct-CFT

Spaces using TIGER-Lab/Qwen2.5-32B-Instruct-CFT 3

Collection including TIGER-Lab/Qwen2.5-32B-Instruct-CFT