Update README.md
Browse files
README.md
CHANGED
@@ -7,7 +7,7 @@ base_model:
|
|
7 |
- allenai/OLMo-2-1124-7B-SFT
|
8 |
library_name: transformers
|
9 |
datasets:
|
10 |
-
- allenai/
|
11 |
---
|
12 |
|
13 |
<img alt="OLMo Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmo2/olmo.png" width="242px">
|
@@ -15,7 +15,7 @@ datasets:
|
|
15 |
# OLMo-2-1124-7B-RM
|
16 |
|
17 |
OLMo 2 7B RM November 2024 is reward model trained on top of the [OLMo 2 7B SFT November 2024](https://huggingface.co/allenai/OLMo2-7B-1124-SFT) model.
|
18 |
-
It has been trained using an OLMo-specific variant of the [Tülu 3 dataset](allenai/tulu-3-sft-olmo-2-mixture) and [this preference dataset](
|
19 |
Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
|
20 |
Check out the OLMo 2 paper (forthcoming) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
|
21 |
|
|
|
7 |
- allenai/OLMo-2-1124-7B-SFT
|
8 |
library_name: transformers
|
9 |
datasets:
|
10 |
+
- allenai/olmo-2-1124-7b-preference-mix-for-rm
|
11 |
---
|
12 |
|
13 |
<img alt="OLMo Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmo2/olmo.png" width="242px">
|
|
|
15 |
# OLMo-2-1124-7B-RM
|
16 |
|
17 |
OLMo 2 7B RM November 2024 is reward model trained on top of the [OLMo 2 7B SFT November 2024](https://huggingface.co/allenai/OLMo2-7B-1124-SFT) model.
|
18 |
+
It has been trained using an OLMo-specific variant of the [Tülu 3 dataset](allenai/tulu-3-sft-olmo-2-mixture) and [this preference dataset](https://huggingface.co/datasets/allenai/olmo-2-1124-7b-preference-mix-for-rm).
|
19 |
Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
|
20 |
Check out the OLMo 2 paper (forthcoming) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
|
21 |
|