ryanmarten commited on
Commit
1a467c4
Β·
verified Β·
1 Parent(s): b70fa61

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +36 -14
README.md CHANGED
@@ -9,29 +9,43 @@ tags:
9
  model-index:
10
  - name: DCFT-Stratos-Verified-114k-7B-4gpus
11
  results: []
 
 
12
  ---
13
 
14
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
- should probably proofread and complete it, then remove this comment. -->
 
16
 
17
- # DCFT-Stratos-Verified-114k-7B-4gpus
18
 
19
- This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on the mlfoundations-dev/stratos_verified_mix dataset.
 
20
 
21
- ## Model description
 
22
 
23
- More information needed
24
 
25
- ## Intended uses & limitations
26
 
27
- More information needed
 
 
 
 
 
 
 
28
 
29
- ## Training and evaluation data
 
 
30
 
31
- More information needed
32
 
33
  ## Training procedure
34
 
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
@@ -49,13 +63,21 @@ The following hyperparameters were used during training:
49
  - lr_scheduler_warmup_ratio: 0.1
50
  - num_epochs: 3.0
51
 
52
- ### Training results
53
-
54
-
55
-
56
  ### Framework versions
57
 
58
  - Transformers 4.46.1
59
  - Pytorch 2.3.0
60
  - Datasets 3.1.0
61
  - Tokenizers 0.20.3
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  model-index:
10
  - name: DCFT-Stratos-Verified-114k-7B-4gpus
11
  results: []
12
+ datasets:
13
+ - open-thoughts/open-thoughts-114k
14
  ---
15
 
16
+ <p align="center">
17
+ <img src="https://huggingface.co/datasets/open-thoughts/open-thoughts-114k/resolve/main/open_thoughts.png" width="50%">
18
+ </p>
19
 
20
+ # OpenThinker-7B
21
 
22
+ This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on the
23
+ [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k) dataset.
24
 
25
+ The dataset is derived by distilling DeepSeek-R1 using the [data pipeline available on github](https://github.com/open-thoughts/open-thoughts).
26
+ More info about the dataset can be found on the dataset card at [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
27
 
28
+ This model improves upon the [Bespoke-Stratos-7B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-7B), which used 17k examples ([Bespoke-Stratos-17k dataset](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)).
29
 
 
30
 
31
+ | | AIME24 | MATH500 | GPQA-D | LCB Easy | LCB Med | LCB Hard | Open Weights | Open Data | Open Code |
32
+ |--|--------|---------|--------|----------|---------|----------|--------------|-----------| --------- |
33
+ |OpenThinker-7B|26.7|86|40.4|74.2|31.1|3.3|βœ”|[βœ”](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k)|[βœ”](https://github.com/open-thoughts/open-thoughts) |
34
+ |Bespoke-Stratos-7B|20.0|82.0|37.8|71.4|25.2|1.6|βœ”|[βœ”](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)|[βœ”](https://github.com/bespokelabsai/curator/tree/main/examples/bespoke-stratos-data-generation)|
35
+ |DeepSeek-R1-Distill-Qwen-7B|43.3|89.4|44.9|81.3|42.2|2.4|βœ”|❌|❌|
36
+ |gpt-4o-0513|9.3|74.6|49.9|-|-|-|❌|❌|❌|
37
+ |o1-mini|63.6|90.0|60.0|-|-|-|❌|❌|❌|
38
+
39
 
40
+ ## Intended uses & limitations
41
+
42
+ Apache 2.0 License
43
 
 
44
 
45
  ## Training procedure
46
 
47
+ We used four 8xH100 nodes to train the model for 20 hours.
48
+
49
  ### Training hyperparameters
50
 
51
  The following hyperparameters were used during training:
 
63
  - lr_scheduler_warmup_ratio: 0.1
64
  - num_epochs: 3.0
65
 
 
 
 
 
66
  ### Framework versions
67
 
68
  - Transformers 4.46.1
69
  - Pytorch 2.3.0
70
  - Datasets 3.1.0
71
  - Tokenizers 0.20.3
72
+
73
+ More info can be found in our repository: [https://github.com/open-thoughts/open-thoughts](https://github.com/open-thoughts/open-thoughts).
74
+
75
+ # Links
76
+ - πŸ“Š [Open Thoughts Launch Blog Post](https://www.open-thoughts.ai/blog/launch)
77
+ - πŸ“Š [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
78
+ - 🧠 [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k)
79
+ - πŸ€– [OpenThinker-7B model](https://huggingface.co/open-thoughts/OpenThinker-7B) - this model.
80
+ - πŸ“Š [Bespoke-Stratos Blog Post](https://www.bespokelabs.ai/blog/bespoke-stratos-the-unreasonable-effectiveness-of-reasoning-distillation)
81
+ - 🧠 [Bespoke-Stratos-17k dataset](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)
82
+ - πŸ€– [Bespoke-Stratos-32B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-32B)
83
+ - πŸ€– [Bespoke-Stratos-7B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-7B)