Update pipeline tag and add project page link
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,21 +1,20 @@
|
|
1 |
---
|
2 |
-
|
|
|
3 |
datasets:
|
4 |
- PKU-ML/Erdos
|
5 |
language:
|
6 |
- en
|
|
|
|
|
7 |
metrics:
|
8 |
- accuracy
|
9 |
-
|
10 |
-
- Qwen/Qwen2.5-3B-Instruct
|
11 |
-
pipeline_tag: text-generation
|
12 |
tags:
|
13 |
- graph
|
14 |
- chat
|
15 |
-
library_name: transformers
|
16 |
---
|
17 |
|
18 |
-
|
19 |
# G1-3B
|
20 |
|
21 |
## Introduction
|
@@ -25,19 +24,20 @@ We apply Group Relative Policy Optimization (GRPO) for reinforcement learning wi
|
|
25 |
|
26 |
G1 brings the following improvements:
|
27 |
|
28 |
-
-
|
29 |
-
-
|
30 |
-
-
|
31 |
|
32 |
|
33 |
**This repo contains the G1-3B model**, which has the following features:
|
34 |
-
-
|
35 |
-
-
|
36 |
-
-
|
37 |
-
-
|
38 |
-
-
|
39 |
|
40 |
For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499) and [GitHub](https://github.com/PKU-ML/G1/tree/main).
|
|
|
41 |
|
42 |
|
43 |
## Requirements
|
@@ -72,10 +72,18 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
72 |
)
|
73 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
74 |
|
75 |
-
prompt = "The task is to determine the degree centrality of a node in the graph
|
76 |
-
|
77 |
-
|
78 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
79 |
"You need to format your answer as a float number."
|
80 |
messages = [
|
81 |
{"role": "user", "content": INSTRUCTION_TEMPLATE.format(instruction=prompt)}
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- Qwen/Qwen2.5-3B-Instruct
|
4 |
datasets:
|
5 |
- PKU-ML/Erdos
|
6 |
language:
|
7 |
- en
|
8 |
+
library_name: transformers
|
9 |
+
license: apache-2.0
|
10 |
metrics:
|
11 |
- accuracy
|
12 |
+
pipeline_tag: graph-ml
|
|
|
|
|
13 |
tags:
|
14 |
- graph
|
15 |
- chat
|
|
|
16 |
---
|
17 |
|
|
|
18 |
# G1-3B
|
19 |
|
20 |
## Introduction
|
|
|
24 |
|
25 |
G1 brings the following improvements:
|
26 |
|
27 |
+
- **Significant improvement on graph reasoning**: G1 models achieve up to 46% improvement over baselines on Erdős, with the 7B variant matching OpenAI’s o3-mini and the 3B model surpassing Qwen2.5-72B-Instruct by notable margins.
|
28 |
+
- **Strong Generalization to unseen graph tasks**: G1 exhibits zero-shot generalization on unseen graph tasks, improving performance on *other graph reasoning benchmarks* (GraphWiz, GraphArena) and *real-world graphs* (Cora, PubMed).
|
29 |
+
- **NO Compromise on general reasoning**: Crucially, G1 preserves general reasoning ability (GSM8K, MATH, MMLU-Pro), proving its versatility.
|
30 |
|
31 |
|
32 |
**This repo contains the G1-3B model**, which has the following features:
|
33 |
+
- Type: Causal Language Models
|
34 |
+
- Training Stage: SFT & RL
|
35 |
+
- Architecture: the same with Qwen2.5-Instruct
|
36 |
+
- Number of Parameters: 3.09B
|
37 |
+
- Context Length: Full 32,768 tokens and generation 8192 tokens
|
38 |
|
39 |
For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499) and [GitHub](https://github.com/PKU-ML/G1/tree/main).
|
40 |
+
Project page: [Hugging Face Collection](https://huggingface.co/collections/PKU-ML/g1-683d659e992794fc99618cf2)
|
41 |
|
42 |
|
43 |
## Requirements
|
|
|
72 |
)
|
73 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
74 |
|
75 |
+
prompt = "The task is to determine the degree centrality of a node in the graph.
|
76 |
+
|
77 |
+
"\
|
78 |
+
"Degree centrality for a node is the fraction of nodes it is connected to.
|
79 |
+
|
80 |
+
"\
|
81 |
+
"Here is an undirected graph containing nodes from 1 to 15. The edges are: (1, 15), (15, 11), (2, 3), (2, 6), (3, 6), (3, 7), (6, 7), (6, 8), (7, 8), (7, 14), (4, 10), (10, 5), (10, 12), (8, 14), (8, 9), (12, 11), (12, 13).
|
82 |
+
|
83 |
+
"\
|
84 |
+
"Question: What is the degree centrality of node 2 in the graph?
|
85 |
+
|
86 |
+
"\
|
87 |
"You need to format your answer as a float number."
|
88 |
messages = [
|
89 |
{"role": "user", "content": INSTRUCTION_TEMPLATE.format(instruction=prompt)}
|