Qwen2.5-32B-Instruct-CFT

Introduction

Qwen2.5-32B-Instruct-CFT is a 32B parameter model fine-tuned using our novel Critique Fine-Tuning (CFT) approach. Built upon the Qwen2.5-32B-Instruct base model, this variant is trained to critique and analyze responses rather than simply imitate them, leading to enhanced reasoning capabilities.

Key Features

  • Built on the powerful Qwen2.5-32B-Instruct foundation
  • Trained using Critique Fine-Tuning (CFT) methodology
  • Highly efficient training with minimal data requirements
  • Inherits the strong instruction-following capabilities of the base model

Training Details

Training Data

  • Dataset: WebInstruct-CFT-4K
  • Training format: (input=[query; noisy response], output=critique)
  • Teacher model: GPT-4o for generating critiques

Training Infrastructure

  • Framework: LLaMA-Factory
  • Hardware: 8x NVIDIA H100 GPUs
  • Training time: ~1.5 hours with DeepSpeed Zero-3

For more details about the model architecture, methodology, and comprehensive evaluation results, please visit our project webpage.

Downloads last month
0
Safetensors
Model size
32.8B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for TIGER-Lab/Qwen2.5-32B-Instruct-CFT

Base model

Qwen/Qwen2.5-32B
Finetuned
(109)
this model

Dataset used to train TIGER-Lab/Qwen2.5-32B-Instruct-CFT

Collection including TIGER-Lab/Qwen2.5-32B-Instruct-CFT