Add pipeline tag and usage examples
#4
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,7 +1,9 @@
|
|
1 |
---
|
|
|
|
|
|
|
2 |
library_name: transformers
|
3 |
license: apache-2.0
|
4 |
-
base_model: Qwen/Qwen2.5-32B-Instruct
|
5 |
tags:
|
6 |
- llama-factory
|
7 |
- full
|
@@ -9,8 +11,7 @@ tags:
|
|
9 |
model-index:
|
10 |
- name: OpenThinker2-32B
|
11 |
results: []
|
12 |
-
|
13 |
-
- open-thoughts/OpenThoughts2-1M
|
14 |
---
|
15 |
|
16 |
<p align="center">
|
@@ -20,12 +21,11 @@ datasets:
|
|
20 |
> [!NOTE]
|
21 |
> We have released a paper for OpenThoughts! See our paper [here](https://arxiv.org/abs/2506.04178).
|
22 |
|
23 |
-
# OpenThinker2-32B
|
24 |
|
25 |
-
|
26 |
-
[OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
|
27 |
|
28 |
-
The [OpenThinker2-32B](https://huggingface.co/open-thoughts/OpenThinker2-32B) model is the highest performing open-data model.
|
29 |
This model improves upon our previous [OpenThinker-32B](https://huggingface.co/open-thoughts/OpenThinker-32B) model, which was trained on 114k examples from [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
|
30 |
The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
|
31 |
|
@@ -38,6 +38,40 @@ The numbers reported in the table below are evaluated with our open-source tool
|
|
38 |
| [S1.1-32B](https://huggingface.co/simplescaling/s1.1-32B) | β
| 59.3 | 42.7 | 91.5 | 87.4 | 62.0 | 58.7 |
|
39 |
|
40 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
41 |
## Data
|
42 |
|
43 |
This model was trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
|
@@ -85,6 +119,7 @@ More info can be found in our repository: [https://github.com/open-thoughts/open
|
|
85 |
- π [OpenThoughts Paper](https://arxiv.org/abs/2506.04178)
|
86 |
- π [OpenThoughts2 and OpenThinker2 Blog Post](https://www.open-thoughts.ai/blog/thinkagain)
|
87 |
- π» [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
|
|
|
88 |
- π§ [OpenThoughts2-1M dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M)
|
89 |
- π€ [OpenThinker2-7B model](https://huggingface.co/open-thoughts/OpenThinker2-7B)
|
90 |
- π€ [OpenThinker2-32B model](https://huggingface.co/open-thoughts/OpenThinker2-32B) - this model.
|
@@ -100,4 +135,4 @@ More info can be found in our repository: [https://github.com/open-thoughts/open
|
|
100 |
primaryClass={cs.LG},
|
101 |
url={https://arxiv.org/abs/2506.04178},
|
102 |
}
|
103 |
-
```
|
|
|
1 |
---
|
2 |
+
base_model: Qwen/Qwen2.5-32B-Instruct
|
3 |
+
datasets:
|
4 |
+
- open-thoughts/OpenThoughts2-1M
|
5 |
library_name: transformers
|
6 |
license: apache-2.0
|
|
|
7 |
tags:
|
8 |
- llama-factory
|
9 |
- full
|
|
|
11 |
model-index:
|
12 |
- name: OpenThinker2-32B
|
13 |
results: []
|
14 |
+
pipeline_tag: text-generation
|
|
|
15 |
---
|
16 |
|
17 |
<p align="center">
|
|
|
21 |
> [!NOTE]
|
22 |
> We have released a paper for OpenThoughts! See our paper [here](https://arxiv.org/abs/2506.04178).
|
23 |
|
24 |
+
# OpenThinker2-32B: A Powerful Open-Data Reasoning Model
|
25 |
|
26 |
+
OpenThinker2-32B is a fine-tuned version of [Qwen/Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset. This model represents a significant advancement in open-source reasoning capabilities, achieving competitive performance with closed-source models on various benchmarks. OpenThinker2-32B excels at complex reasoning tasks involving math, code, and general knowledge, demonstrating its versatility and robustness.
|
|
|
27 |
|
28 |
+
The [OpenThinker2-32B](https://huggingface.co/open-thoughts/OpenThinker2-32B) model is the highest performing open-data model.
|
29 |
This model improves upon our previous [OpenThinker-32B](https://huggingface.co/open-thoughts/OpenThinker-32B) model, which was trained on 114k examples from [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
|
30 |
The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
|
31 |
|
|
|
38 |
| [S1.1-32B](https://huggingface.co/simplescaling/s1.1-32B) | β
| 59.3 | 42.7 | 91.5 | 87.4 | 62.0 | 58.7 |
|
39 |
|
40 |
|
41 |
+
## Usage Examples
|
42 |
+
|
43 |
+
This model can be easily used with the Hugging Face `pipeline` API for text generation.
|
44 |
+
|
45 |
+
**Example 1: Simple Text Generation**
|
46 |
+
|
47 |
+
```python
|
48 |
+
from transformers import pipeline
|
49 |
+
|
50 |
+
generator = pipeline('text-generation', model='open-thoughts/OpenThinker2-32B')
|
51 |
+
result = generator("Once upon a time,", max_length=50)
|
52 |
+
print(result[0]['generated_text'])
|
53 |
+
```
|
54 |
+
|
55 |
+
**Example 2: Controlling the Length of the Generated Text**
|
56 |
+
|
57 |
+
```python
|
58 |
+
from transformers import pipeline
|
59 |
+
|
60 |
+
generator = pipeline('text-generation', model='open-thoughts/OpenThinker2-32B')
|
61 |
+
result = generator("The quick brown fox jumps over the lazy dog.", max_length=100)
|
62 |
+
print(result[0]['generated_text'])
|
63 |
+
```
|
64 |
+
|
65 |
+
**Example 3: Setting the Temperature**
|
66 |
+
|
67 |
+
```python
|
68 |
+
from transformers import pipeline
|
69 |
+
|
70 |
+
generator = pipeline('text-generation', model='open-thoughts/OpenThinker2-32B')
|
71 |
+
result = generator("Write a short poem about nature:", max_length=50, temperature=0.7)
|
72 |
+
print(result[0]['generated_text'])
|
73 |
+
```
|
74 |
+
|
75 |
## Data
|
76 |
|
77 |
This model was trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
|
|
|
119 |
- π [OpenThoughts Paper](https://arxiv.org/abs/2506.04178)
|
120 |
- π [OpenThoughts2 and OpenThinker2 Blog Post](https://www.open-thoughts.ai/blog/thinkagain)
|
121 |
- π» [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
|
122 |
+
- π [Project Page](https://openthoughts.ai)
|
123 |
- π§ [OpenThoughts2-1M dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M)
|
124 |
- π€ [OpenThinker2-7B model](https://huggingface.co/open-thoughts/OpenThinker2-7B)
|
125 |
- π€ [OpenThinker2-32B model](https://huggingface.co/open-thoughts/OpenThinker2-32B) - this model.
|
|
|
135 |
primaryClass={cs.LG},
|
136 |
url={https://arxiv.org/abs/2506.04178},
|
137 |
}
|
138 |
+
```
|