Add pipeline tag and usage examples

#4
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +43 -8
README.md CHANGED
@@ -1,7 +1,9 @@
1
  ---
 
 
 
2
  library_name: transformers
3
  license: apache-2.0
4
- base_model: Qwen/Qwen2.5-32B-Instruct
5
  tags:
6
  - llama-factory
7
  - full
@@ -9,8 +11,7 @@ tags:
9
  model-index:
10
  - name: OpenThinker2-32B
11
  results: []
12
- datasets:
13
- - open-thoughts/OpenThoughts2-1M
14
  ---
15
 
16
  <p align="center">
@@ -20,12 +21,11 @@ datasets:
20
  > [!NOTE]
21
  > We have released a paper for OpenThoughts! See our paper [here](https://arxiv.org/abs/2506.04178).
22
 
23
- # OpenThinker2-32B
24
 
25
- This model is a fine-tuned version of [Qwen/Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) on the
26
- [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
27
 
28
- The [OpenThinker2-32B](https://huggingface.co/open-thoughts/OpenThinker2-32B) model is the highest performing open-data model.
29
  This model improves upon our previous [OpenThinker-32B](https://huggingface.co/open-thoughts/OpenThinker-32B) model, which was trained on 114k examples from [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
30
  The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
31
 
@@ -38,6 +38,40 @@ The numbers reported in the table below are evaluated with our open-source tool
38
  | [S1.1-32B](https://huggingface.co/simplescaling/s1.1-32B) | βœ… | 59.3 | 42.7 | 91.5 | 87.4 | 62.0 | 58.7 |
39
 
40
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
41
  ## Data
42
 
43
  This model was trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
@@ -85,6 +119,7 @@ More info can be found in our repository: [https://github.com/open-thoughts/open
85
  - πŸ“ [OpenThoughts Paper](https://arxiv.org/abs/2506.04178)
86
  - πŸ“Š [OpenThoughts2 and OpenThinker2 Blog Post](https://www.open-thoughts.ai/blog/thinkagain)
87
  - πŸ’» [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
 
88
  - 🧠 [OpenThoughts2-1M dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M)
89
  - πŸ€– [OpenThinker2-7B model](https://huggingface.co/open-thoughts/OpenThinker2-7B)
90
  - πŸ€– [OpenThinker2-32B model](https://huggingface.co/open-thoughts/OpenThinker2-32B) - this model.
@@ -100,4 +135,4 @@ More info can be found in our repository: [https://github.com/open-thoughts/open
100
  primaryClass={cs.LG},
101
  url={https://arxiv.org/abs/2506.04178},
102
  }
103
- ```
 
1
  ---
2
+ base_model: Qwen/Qwen2.5-32B-Instruct
3
+ datasets:
4
+ - open-thoughts/OpenThoughts2-1M
5
  library_name: transformers
6
  license: apache-2.0
 
7
  tags:
8
  - llama-factory
9
  - full
 
11
  model-index:
12
  - name: OpenThinker2-32B
13
  results: []
14
+ pipeline_tag: text-generation
 
15
  ---
16
 
17
  <p align="center">
 
21
  > [!NOTE]
22
  > We have released a paper for OpenThoughts! See our paper [here](https://arxiv.org/abs/2506.04178).
23
 
24
+ # OpenThinker2-32B: A Powerful Open-Data Reasoning Model
25
 
26
+ OpenThinker2-32B is a fine-tuned version of [Qwen/Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset. This model represents a significant advancement in open-source reasoning capabilities, achieving competitive performance with closed-source models on various benchmarks. OpenThinker2-32B excels at complex reasoning tasks involving math, code, and general knowledge, demonstrating its versatility and robustness.
 
27
 
28
+ The [OpenThinker2-32B](https://huggingface.co/open-thoughts/OpenThinker2-32B) model is the highest performing open-data model.
29
  This model improves upon our previous [OpenThinker-32B](https://huggingface.co/open-thoughts/OpenThinker-32B) model, which was trained on 114k examples from [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
30
  The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
31
 
 
38
  | [S1.1-32B](https://huggingface.co/simplescaling/s1.1-32B) | βœ… | 59.3 | 42.7 | 91.5 | 87.4 | 62.0 | 58.7 |
39
 
40
 
41
+ ## Usage Examples
42
+
43
+ This model can be easily used with the Hugging Face `pipeline` API for text generation.
44
+
45
+ **Example 1: Simple Text Generation**
46
+
47
+ ```python
48
+ from transformers import pipeline
49
+
50
+ generator = pipeline('text-generation', model='open-thoughts/OpenThinker2-32B')
51
+ result = generator("Once upon a time,", max_length=50)
52
+ print(result[0]['generated_text'])
53
+ ```
54
+
55
+ **Example 2: Controlling the Length of the Generated Text**
56
+
57
+ ```python
58
+ from transformers import pipeline
59
+
60
+ generator = pipeline('text-generation', model='open-thoughts/OpenThinker2-32B')
61
+ result = generator("The quick brown fox jumps over the lazy dog.", max_length=100)
62
+ print(result[0]['generated_text'])
63
+ ```
64
+
65
+ **Example 3: Setting the Temperature**
66
+
67
+ ```python
68
+ from transformers import pipeline
69
+
70
+ generator = pipeline('text-generation', model='open-thoughts/OpenThinker2-32B')
71
+ result = generator("Write a short poem about nature:", max_length=50, temperature=0.7)
72
+ print(result[0]['generated_text'])
73
+ ```
74
+
75
  ## Data
76
 
77
  This model was trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
 
119
  - πŸ“ [OpenThoughts Paper](https://arxiv.org/abs/2506.04178)
120
  - πŸ“Š [OpenThoughts2 and OpenThinker2 Blog Post](https://www.open-thoughts.ai/blog/thinkagain)
121
  - πŸ’» [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
122
+ - 🌐 [Project Page](https://openthoughts.ai)
123
  - 🧠 [OpenThoughts2-1M dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M)
124
  - πŸ€– [OpenThinker2-7B model](https://huggingface.co/open-thoughts/OpenThinker2-7B)
125
  - πŸ€– [OpenThinker2-32B model](https://huggingface.co/open-thoughts/OpenThinker2-32B) - this model.
 
135
  primaryClass={cs.LG},
136
  url={https://arxiv.org/abs/2506.04178},
137
  }
138
+ ```