Add paper link to citation
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,18 +1,18 @@
|
|
1 |
---
|
2 |
base_model: LGAI-EXAONE/EXAONE-Deep-32B
|
3 |
-
base_model_relation: quantized
|
4 |
-
license: other
|
5 |
-
license_name: exaone
|
6 |
-
license_link: LICENSE
|
7 |
language:
|
8 |
- en
|
9 |
- ko
|
|
|
|
|
|
|
|
|
|
|
10 |
tags:
|
11 |
- lg-ai
|
12 |
- exaone
|
13 |
- exaone-deep
|
14 |
-
|
15 |
-
library_name: transformers
|
16 |
---
|
17 |
|
18 |
<p align="center">
|
@@ -145,8 +145,11 @@ We provide the pre-quantized EXAONE Deep models with **AWQ** and several quantiz
|
|
145 |
|
146 |
To achieve the expected performance, we recommend using the following configurations:
|
147 |
|
148 |
-
1. Ensure the model starts with `<thought
|
149 |
-
|
|
|
|
|
|
|
150 |
3. Avoid using system prompt, and build the instruction on the user prompt.
|
151 |
4. Additional instructions help the models reason more deeply, so that the models generate better output.
|
152 |
- For math problems, the instructions **"Please reason step by step, and put your final answer within \boxed{}."** are helpful.
|
@@ -173,7 +176,9 @@ outputs violating LG AI’s ethical principles when using EXAONE language models
|
|
173 |
The model is licensed under [EXAONE AI Model License Agreement 1.1 - NC](./LICENSE)
|
174 |
|
175 |
## Citation
|
176 |
-
|
|
|
|
|
177 |
```
|
178 |
@article{exaone-deep,
|
179 |
title={EXAONE Deep: Reasoning Enhanced Language Models},
|
|
|
1 |
---
|
2 |
base_model: LGAI-EXAONE/EXAONE-Deep-32B
|
|
|
|
|
|
|
|
|
3 |
language:
|
4 |
- en
|
5 |
- ko
|
6 |
+
library_name: transformers
|
7 |
+
license: other
|
8 |
+
license_name: exaone
|
9 |
+
license_link: LICENSE
|
10 |
+
pipeline_tag: text-generation
|
11 |
tags:
|
12 |
- lg-ai
|
13 |
- exaone
|
14 |
- exaone-deep
|
15 |
+
base_model_relation: quantized
|
|
|
16 |
---
|
17 |
|
18 |
<p align="center">
|
|
|
145 |
|
146 |
To achieve the expected performance, we recommend using the following configurations:
|
147 |
|
148 |
+
1. Ensure the model starts with `<thought>
|
149 |
+
` for reasoning steps. The model's output quality may be degraded when you omit it. You can easily apply this feature by using `tokenizer.apply_chat_template()` with `add_generation_prompt=True`. Please check the example code on [Quickstart](#quickstart) section.
|
150 |
+
2. The reasoning steps of EXAONE Deep models enclosed by `<thought>
|
151 |
+
...
|
152 |
+
</thought>` usually have lots of tokens, so previous reasoning steps may be necessary to be removed in multi-turn situation. The provided tokenizer handles this automatically.
|
153 |
3. Avoid using system prompt, and build the instruction on the user prompt.
|
154 |
4. Additional instructions help the models reason more deeply, so that the models generate better output.
|
155 |
- For math problems, the instructions **"Please reason step by step, and put your final answer within \boxed{}."** are helpful.
|
|
|
176 |
The model is licensed under [EXAONE AI Model License Agreement 1.1 - NC](./LICENSE)
|
177 |
|
178 |
## Citation
|
179 |
+
|
180 |
+
The model was presented in the paper [EXAONE Deep: Reasoning Enhanced Language Models](https://huggingface.co/papers/2503.12524)
|
181 |
+
|
182 |
```
|
183 |
@article{exaone-deep,
|
184 |
title={EXAONE Deep: Reasoning Enhanced Language Models},
|