hamishivi commited on
Commit
7a709c2
·
verified ·
1 Parent(s): 216f4ad

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -7,7 +7,7 @@ base_model:
7
  - allenai/OLMo-2-1124-7B-SFT
8
  library_name: transformers
9
  datasets:
10
- - allenai/tulu-3-sft-olmo-2-mixture
11
  ---
12
 
13
  <img alt="OLMo Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmo2/olmo.png" width="242px">
@@ -15,7 +15,7 @@ datasets:
15
  # OLMo-2-1124-7B-RM
16
 
17
  OLMo 2 7B RM November 2024 is reward model trained on top of the [OLMo 2 7B SFT November 2024](https://huggingface.co/allenai/OLMo2-7B-1124-SFT) model.
18
- It has been trained using an OLMo-specific variant of the [Tülu 3 dataset](allenai/tulu-3-sft-olmo-2-mixture) and [this preference dataset](todo).
19
  Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
20
  Check out the OLMo 2 paper (forthcoming) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
21
 
 
7
  - allenai/OLMo-2-1124-7B-SFT
8
  library_name: transformers
9
  datasets:
10
+ - allenai/olmo-2-1124-7b-preference-mix-for-rm
11
  ---
12
 
13
  <img alt="OLMo Logo" src="https://huggingface.co/datasets/allenai/blog-images/resolve/main/olmo2/olmo.png" width="242px">
 
15
  # OLMo-2-1124-7B-RM
16
 
17
  OLMo 2 7B RM November 2024 is reward model trained on top of the [OLMo 2 7B SFT November 2024](https://huggingface.co/allenai/OLMo2-7B-1124-SFT) model.
18
+ It has been trained using an OLMo-specific variant of the [Tülu 3 dataset](allenai/tulu-3-sft-olmo-2-mixture) and [this preference dataset](https://huggingface.co/datasets/allenai/olmo-2-1124-7b-preference-mix-for-rm).
19
  Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
20
  Check out the OLMo 2 paper (forthcoming) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
21