File size: 1,824 Bytes
d0dd60b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
---
license: mit
base_model:
- Qwen/Qwen2.5-7B-Instruct-1M
---

### Model Card: Graph-R1 Series

This model card covers the Graph-R1 series of models, including the final released versions and variants used in ablation studies. All information is based on the provided research paper.

#### **Model Details**

* **Model Developer**: HKUST-DSAIL
* **Model Series**: Graph-R1
* **Model Variants**:
    * **Graph-R1-7B**: Fine-tuned from Qwen2.5-7B-Instruct-1M.
    * **Graph-R1-1.5B**: Fine-tuned from Qwen2.5-1.5B.
    * **Ablation Models**: Multiple variants based on different training configurations (e.g., data volume, training stages, reward functions, curriculum learning strategies).
* **Model Type**: Small reasoning language model, specialized in solving complex NP graph-theoretic problems.
* **Architecture**:
    * **Base Model**: Qwen2.5
    * **Training Framework**:
        1.  **Cold-start Supervised Fine-Tuning (SFT)**: Fine-tuned using long Chain-of-Thought (Long-CoT) data extracted from the QwQ-32B model to inject graph reasoning knowledge.
        2.  **Reasoning Optimization via Reinforcement Learning (RL)**: Employs a Group Relative Policy Optimization (GRPO)-based RL framework, combined with a curriculum learning strategy. 
* **Model Date**: 2025/04

#### **Intended Use**

* **Primary Use Cases**:
    * Solving complex graph-theoretic computational problems at the NP-Complete level, such as the Traveling Salesman Problem (TSP), Graph Edit Distance (GED), and Maximum Clique Problem (MCP).
    * Serving as a compact, resource-efficient reasoning model for academic research and practical applications. 
* **Potential Cross-Domain Applications**:
    * The model demonstrates transferability to other complex reasoning tasks, including mathematics, programming, STEM, and logical reasoning.