lbourdois commited on
Commit
11ce3ec
Β·
verified Β·
1 Parent(s): 5ad2ec3

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +109 -95
README.md CHANGED
@@ -1,95 +1,109 @@
1
- ---
2
- library_name: transformers
3
- license: apache-2.0
4
- base_model: Qwen/Qwen2.5-7B-Instruct
5
- tags:
6
- - llama-factory
7
- - full
8
- - generated_from_trainer
9
- model-index:
10
- - name: OpenThinker2-7B
11
- results: []
12
- datasets:
13
- - open-thoughts/OpenThoughts2-1M
14
- ---
15
-
16
- <p align="center">
17
- <img src="https://huggingface.co/datasets/open-thoughts/open-thoughts-114k/resolve/main/open_thoughts.png" width="50%">
18
- </p>
19
-
20
- # OpenThinker2-7B
21
-
22
- This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on the
23
- [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
24
-
25
- The [OpenThinker2-7B](https://huggingface.co/open-thoughts/OpenThinker2-7B) model is the top 7B open-data reasoning model. It delivers performance comparable to state of the art 7B models like [DeepSeek-R1-Distill-7B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B) across a suite of tasks.
26
- This model improves upon our previous [OpenThinker-7B](https://huggingface.co/open-thoughts/OpenThinker-7B) model, which was trained on 114k examples from [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
27
- The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
28
-
29
- | Model | Data | AIME24 | AIME25 | AMC23 | MATH500 | GPQA-D | LCBv2 |
30
- | --------------------------------------------------------------------------------------------- | ---- | ------ | ------ | ----- | ------- | ------ | ----------- |
31
- | [OpenThinker2-7B](https://huggingface.co/open-thoughts/OpenThinker2-7B) | βœ… | 50.0 | 33.3 | 89.5 | 88.4 | 49.3 | 55.6 |
32
- | [OpenThinker-7B](https://huggingface.co/open-thoughts/OpenThinker-7B) | βœ… | 31.3 | 23.3 | 74.5 | 83.2 | 42.9 | 38.0 |
33
- | [DeepSeek-R1-Distill-Qwen-7B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B) | ❌ | 57.3 | 33.3 | 92.0 | 89.6 | 47.3 | 48.4 |
34
- | [OlympicCoder-7B](https://huggingface.co/open-r1/OlympicCoder-7B) | βœ… | 20.7 | 15.3 | 63.0 | 74.8 | 25.3 | 55.4 |
35
- | [OpenR1-Qwen-7B](https://huggingface.co/open-r1/OpenR1-Qwen-7B) | βœ… | 48.7 | 34.7 | 88.5 | 87.8 | 21.2 | 9.5<br><br> |
36
-
37
- ## Data
38
-
39
- This model was trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
40
-
41
- The [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset was constructed by augmenting [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k) with existing datasets like [OpenR1](https://huggingface.co/open-r1), as well as additional math and code reasoning data.
42
- We generate the additional math and code data by ablating over 26 different question generation methodologies and sampling from the highest performing ones.
43
-
44
- See the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset page or our [blog post](https://www.open-thoughts.ai/blog/thinkagain) for additional information.
45
-
46
-
47
- ## Intended uses & limitations
48
-
49
- Apache 2.0 License
50
-
51
-
52
- ## Training procedure
53
-
54
- We used 32 8xA100 nodes to train the model for 36 hours.
55
-
56
- ### Training hyperparameters
57
-
58
- The following hyperparameters were used during training:
59
- - learning_rate: 8e-05
60
- - seed: 42
61
- - distributed_type: multi-GPU
62
- - num_devices: 256
63
- - gradient_accumulation_steps: 2
64
- - total_train_batch_size: 512
65
- - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
66
- - lr_scheduler_type: cosine
67
- - lr_scheduler_warmup_ratio: 0.1
68
- - num_epochs: 5.0
69
-
70
- ### Framework versions
71
-
72
- - Transformers 4.46.1
73
- - Pytorch 2.3.0
74
- - Datasets 3.1.0
75
- - Tokenizers 0.20.3
76
-
77
- More info can be found in our repository: [https://github.com/open-thoughts/open-thoughts](https://github.com/open-thoughts/open-thoughts).
78
-
79
- # Citation
80
- ```
81
- @misc{openthoughts,
82
- author = {Team, OpenThoughts},
83
- month = jan,
84
- title = {{Open Thoughts}},
85
- howpublished = {https://open-thoughts.ai},
86
- year = {2025}
87
- }
88
- ```
89
-
90
- # Links
91
- - πŸ“Š [OpenThoughts2 and OpenThinker2 Blog Post](https://www.open-thoughts.ai/blog/thinkagain)
92
- - πŸ’» [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
93
- - 🧠 [OpenThoughts2-1M dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M)
94
- - πŸ€– [OpenThinker2-7B model](https://huggingface.co/open-thoughts/OpenThinker2-7B) - this model.
95
- - πŸ€– [OpenThinker2-32B model](https://huggingface.co/open-thoughts/OpenThinker2-32B)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: Qwen/Qwen2.5-7B-Instruct
5
+ tags:
6
+ - llama-factory
7
+ - full
8
+ - generated_from_trainer
9
+ datasets:
10
+ - open-thoughts/OpenThoughts2-1M
11
+ language:
12
+ - zho
13
+ - eng
14
+ - fra
15
+ - spa
16
+ - por
17
+ - deu
18
+ - ita
19
+ - rus
20
+ - jpn
21
+ - kor
22
+ - vie
23
+ - tha
24
+ - ara
25
+ model-index:
26
+ - name: OpenThinker2-7B
27
+ results: []
28
+ ---
29
+
30
+ <p align="center">
31
+ <img src="https://huggingface.co/datasets/open-thoughts/open-thoughts-114k/resolve/main/open_thoughts.png" width="50%">
32
+ </p>
33
+
34
+ # OpenThinker2-7B
35
+
36
+ This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on the
37
+ [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
38
+
39
+ The [OpenThinker2-7B](https://huggingface.co/open-thoughts/OpenThinker2-7B) model is the top 7B open-data reasoning model. It delivers performance comparable to state of the art 7B models like [DeepSeek-R1-Distill-7B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B) across a suite of tasks.
40
+ This model improves upon our previous [OpenThinker-7B](https://huggingface.co/open-thoughts/OpenThinker-7B) model, which was trained on 114k examples from [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
41
+ The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
42
+
43
+ | Model | Data | AIME24 | AIME25 | AMC23 | MATH500 | GPQA-D | LCBv2 |
44
+ | --------------------------------------------------------------------------------------------- | ---- | ------ | ------ | ----- | ------- | ------ | ----------- |
45
+ | [OpenThinker2-7B](https://huggingface.co/open-thoughts/OpenThinker2-7B) | βœ… | 50.0 | 33.3 | 89.5 | 88.4 | 49.3 | 55.6 |
46
+ | [OpenThinker-7B](https://huggingface.co/open-thoughts/OpenThinker-7B) | βœ… | 31.3 | 23.3 | 74.5 | 83.2 | 42.9 | 38.0 |
47
+ | [DeepSeek-R1-Distill-Qwen-7B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B) | ❌ | 57.3 | 33.3 | 92.0 | 89.6 | 47.3 | 48.4 |
48
+ | [OlympicCoder-7B](https://huggingface.co/open-r1/OlympicCoder-7B) | βœ… | 20.7 | 15.3 | 63.0 | 74.8 | 25.3 | 55.4 |
49
+ | [OpenR1-Qwen-7B](https://huggingface.co/open-r1/OpenR1-Qwen-7B) | βœ… | 48.7 | 34.7 | 88.5 | 87.8 | 21.2 | 9.5<br><br> |
50
+
51
+ ## Data
52
+
53
+ This model was trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
54
+
55
+ The [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset was constructed by augmenting [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k) with existing datasets like [OpenR1](https://huggingface.co/open-r1), as well as additional math and code reasoning data.
56
+ We generate the additional math and code data by ablating over 26 different question generation methodologies and sampling from the highest performing ones.
57
+
58
+ See the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset page or our [blog post](https://www.open-thoughts.ai/blog/thinkagain) for additional information.
59
+
60
+
61
+ ## Intended uses & limitations
62
+
63
+ Apache 2.0 License
64
+
65
+
66
+ ## Training procedure
67
+
68
+ We used 32 8xA100 nodes to train the model for 36 hours.
69
+
70
+ ### Training hyperparameters
71
+
72
+ The following hyperparameters were used during training:
73
+ - learning_rate: 8e-05
74
+ - seed: 42
75
+ - distributed_type: multi-GPU
76
+ - num_devices: 256
77
+ - gradient_accumulation_steps: 2
78
+ - total_train_batch_size: 512
79
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
80
+ - lr_scheduler_type: cosine
81
+ - lr_scheduler_warmup_ratio: 0.1
82
+ - num_epochs: 5.0
83
+
84
+ ### Framework versions
85
+
86
+ - Transformers 4.46.1
87
+ - Pytorch 2.3.0
88
+ - Datasets 3.1.0
89
+ - Tokenizers 0.20.3
90
+
91
+ More info can be found in our repository: [https://github.com/open-thoughts/open-thoughts](https://github.com/open-thoughts/open-thoughts).
92
+
93
+ # Citation
94
+ ```
95
+ @misc{openthoughts,
96
+ author = {Team, OpenThoughts},
97
+ month = jan,
98
+ title = {{Open Thoughts}},
99
+ howpublished = {https://open-thoughts.ai},
100
+ year = {2025}
101
+ }
102
+ ```
103
+
104
+ # Links
105
+ - πŸ“Š [OpenThoughts2 and OpenThinker2 Blog Post](https://www.open-thoughts.ai/blog/thinkagain)
106
+ - πŸ’» [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
107
+ - 🧠 [OpenThoughts2-1M dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M)
108
+ - πŸ€– [OpenThinker2-7B model](https://huggingface.co/open-thoughts/OpenThinker2-7B) - this model.
109
+ - πŸ€– [OpenThinker2-32B model](https://huggingface.co/open-thoughts/OpenThinker2-32B)