Update pipeline tag and add project page link

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +26 -18
README.md CHANGED
@@ -1,21 +1,20 @@
1
  ---
2
- license: apache-2.0
 
3
  datasets:
4
  - PKU-ML/Erdos
5
  language:
6
  - en
 
 
7
  metrics:
8
  - accuracy
9
- base_model:
10
- - Qwen/Qwen2.5-3B-Instruct
11
- pipeline_tag: text-generation
12
  tags:
13
  - graph
14
  - chat
15
- library_name: transformers
16
  ---
17
 
18
-
19
  # G1-3B
20
 
21
  ## Introduction
@@ -25,19 +24,20 @@ We apply Group Relative Policy Optimization (GRPO) for reinforcement learning wi
25
 
26
  G1 brings the following improvements:
27
 
28
- - **Significant improvement on graph reasoning**: G1 models achieve up to 46% improvement over baselines on Erdős, with the 7B variant matching OpenAI’s o3-mini and the 3B model surpassing Qwen2.5-72B-Instruct by notable margins.
29
- - **Strong Generalization to unseen graph tasks**: G1 exhibits zero-shot generalization on unseen graph tasks, improving performance on *other graph reasoning benchmarks* (GraphWiz, GraphArena) and *real-world graphs* (Cora, PubMed).
30
- - **NO Compromise on general reasoning**: Crucially, G1 preserves general reasoning ability (GSM8K, MATH, MMLU-Pro), proving its versatility.
31
 
32
 
33
  **This repo contains the G1-3B model**, which has the following features:
34
- - Type: Causal Language Models
35
- - Training Stage: SFT & RL
36
- - Architecture: the same with Qwen2.5-Instruct
37
- - Number of Parameters: 3.09B
38
- - Context Length: Full 32,768 tokens and generation 8192 tokens
39
 
40
  For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499) and [GitHub](https://github.com/PKU-ML/G1/tree/main).
 
41
 
42
 
43
  ## Requirements
@@ -72,10 +72,18 @@ model = AutoModelForCausalLM.from_pretrained(
72
  )
73
  tokenizer = AutoTokenizer.from_pretrained(model_name)
74
 
75
- prompt = "The task is to determine the degree centrality of a node in the graph.\n\n"\
76
- "Degree centrality for a node is the fraction of nodes it is connected to.\n\n"\
77
- "Here is an undirected graph containing nodes from 1 to 15. The edges are: (1, 15), (15, 11), (2, 3), (2, 6), (3, 6), (3, 7), (6, 7), (6, 8), (7, 8), (7, 14), (4, 10), (10, 5), (10, 12), (8, 14), (8, 9), (12, 11), (12, 13).\n\n"\
78
- "Question: What is the degree centrality of node 2 in the graph?\n\n"\
 
 
 
 
 
 
 
 
79
  "You need to format your answer as a float number."
80
  messages = [
81
  {"role": "user", "content": INSTRUCTION_TEMPLATE.format(instruction=prompt)}
 
1
  ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-3B-Instruct
4
  datasets:
5
  - PKU-ML/Erdos
6
  language:
7
  - en
8
+ library_name: transformers
9
+ license: apache-2.0
10
  metrics:
11
  - accuracy
12
+ pipeline_tag: graph-ml
 
 
13
  tags:
14
  - graph
15
  - chat
 
16
  ---
17
 
 
18
  # G1-3B
19
 
20
  ## Introduction
 
24
 
25
  G1 brings the following improvements:
26
 
27
+ - **Significant improvement on graph reasoning**: G1 models achieve up to 46% improvement over baselines on Erdős, with the 7B variant matching OpenAI’s o3-mini and the 3B model surpassing Qwen2.5-72B-Instruct by notable margins.
28
+ - **Strong Generalization to unseen graph tasks**: G1 exhibits zero-shot generalization on unseen graph tasks, improving performance on *other graph reasoning benchmarks* (GraphWiz, GraphArena) and *real-world graphs* (Cora, PubMed).
29
+ - **NO Compromise on general reasoning**: Crucially, G1 preserves general reasoning ability (GSM8K, MATH, MMLU-Pro), proving its versatility.
30
 
31
 
32
  **This repo contains the G1-3B model**, which has the following features:
33
+ - Type: Causal Language Models
34
+ - Training Stage: SFT & RL
35
+ - Architecture: the same with Qwen2.5-Instruct
36
+ - Number of Parameters: 3.09B
37
+ - Context Length: Full 32,768 tokens and generation 8192 tokens
38
 
39
  For more details, please refer to our [paper](https://arxiv.org/pdf/2505.18499) and [GitHub](https://github.com/PKU-ML/G1/tree/main).
40
+ Project page: [Hugging Face Collection](https://huggingface.co/collections/PKU-ML/g1-683d659e992794fc99618cf2)
41
 
42
 
43
  ## Requirements
 
72
  )
73
  tokenizer = AutoTokenizer.from_pretrained(model_name)
74
 
75
+ prompt = "The task is to determine the degree centrality of a node in the graph.
76
+
77
+ "\
78
+ "Degree centrality for a node is the fraction of nodes it is connected to.
79
+
80
+ "\
81
+ "Here is an undirected graph containing nodes from 1 to 15. The edges are: (1, 15), (15, 11), (2, 3), (2, 6), (3, 6), (3, 7), (6, 7), (6, 8), (7, 8), (7, 14), (4, 10), (10, 5), (10, 12), (8, 14), (8, 9), (12, 11), (12, 13).
82
+
83
+ "\
84
+ "Question: What is the degree centrality of node 2 in the graph?
85
+
86
+ "\
87
  "You need to format your answer as a float number."
88
  messages = [
89
  {"role": "user", "content": INSTRUCTION_TEMPLATE.format(instruction=prompt)}