Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: other
|
3 |
+
license_name: skywork-community-license
|
4 |
+
license_link: >-
|
5 |
+
https://github.com/SkyworkAI/Skywork-Reward/blob/main/misc/Skywork%20Community%20License.pdf
|
6 |
+
datasets:
|
7 |
+
- nvidia/HelpSteer2
|
8 |
+
language:
|
9 |
+
- en
|
10 |
+
metrics:
|
11 |
+
- accuracy
|
12 |
+
base_model:
|
13 |
+
- Skywork/Skywork-Reward-Gemma-2-27B-v0.2
|
14 |
+
library_name: transformers
|
15 |
+
---
|
16 |
+
# Interpreting Language Model Preferences Through the Lens of Decision Trees
|
17 |
+
|
18 |
+
+ **Author** [Min Li](https://min-li.github.io/)
|
19 |
+
+ **Blog**: https://rlhflow.github.io/posts/2025-01-22-decision-tree-reward-model/
|
20 |
+
+ **Models**:
|
21 |
+
+ [**Decision-Tree-Reward-Gemma-2-27B**](https://huggingface.co/RLHFlow/Decision-Tree-Reward-Gemma-2-27B)
|
22 |
+
+ [**Decision-Tree-Reward-Llama-3.1-8B**](https://huggingface.co/RLHFlow/Decision-Tree-Reward-Llama-3.1-8B)
|
23 |
+
+ **Code Repository:** https://github.com/RLHFlow/RLHF-Reward-Modeling/decision_tree/
|
24 |
+
+ **Tech Report**: To release soon
|
25 |
+
|
26 |
+
## RewardBench Leaderboard (Jan 2025)
|
27 |
+
|
28 |
+
Rank | Model | Base Model | Method | Overall Score | Chat | Chat Hard | Safety | Reasoning |
|
29 |
+
|:------|:------|:-----------|:-------|:------|:-----|:----------|:-------|:----------|
|
30 |
+
1 | [**Decision-Tree-Reward-Gemma-2-27B**](https://huggingface.co/RLHFlow/Decision-Tree-Reward-Gemma-2-27B) | Gemma-2-27B | Decision Tree | **95.3** | 96.9 | **91.4** | 93.7 | **99.1** |
|
31 |
+
2 | INF-QRM-Llama3.1-70B | Llama-3.1-70B | Sequence Classifier | 95.1 | 96.6 | 91.0 | 93.6 | **99.1** |
|
32 |
+
3 | QRM-Gemma-2-27B | Gemma-2-27B | Sequence Classifier | 94.4 | 96.6 | 90.1 | 92.7 | 98.3 |
|
33 |
+
4 | Skywork-Reward-Gemma-2-27B-v0.2 | Gemma-2-27B | Sequence Classifier | 94.3 | 96.1 | 89.9 | 93.0 | 98.1 |
|
34 |
+
5 | [**Decision-Tree-Reward-Llama-3.1-8B**](https://huggingface.co/RLHFlow/Decision-Tree-Reward-Llama-3.1-8B) | Llama-3.1-8B | Decision Tree | 94.3 | 96.9 | 89.3 | 92.9 | 98.5 |
|
35 |
+
6 | Llama-3.1-Nemotron-70B-Reward | Llama-3.1-70B | Custom Classifier | 94.1 | 97.5 | 85.7 | **95.1** | 98.1 |
|
36 |
+
7 | Skywork-Reward-Gemma-2-27B | Gemma-2-27B | Sequence Classifier | 93.8 | 95.8 | **91.4** | 91.9 | 96.1 |
|
37 |
+
8 | TextEval-Llama3.1-70B | Llama-3.1-70B | Generative | 93.5 | 94.1 | 90.1 | 93.2 | 96.4 |
|
38 |
+
9 | MetaMetrics-RM-v1.0 | - | Custom Classifier | 93.4 | **98.3** | 86.4 | 90.8 | 98.2 |
|
39 |
+
10 | Skywork-Critic-Llama-3.1-70B | Llama-3.1-70B | Generative | 93.3 | 96.6 | 87.9 | 93.1 | 95.5 |
|
40 |
+
11 | QRM-Llama3.1-8B-v2 | Llama-3.1-8B | Sequence Classifier | 93.1 | 96.4 | 86.8 | 92.6 | 96.8 |
|
41 |
+
12 | Skywork-Reward-Llama-3.1-8B-v0.2 | Llama-3.1-8B | Sequence Classifier | 93.1 | 94.7 | 88.4 | 92.7 | 96.7 |
|
42 |
+
|
43 |
+
## License
|
44 |
+
**Note**: This model is finetuned from a Skywork model under the following license agreement:
|
45 |
+
|
46 |
+
The community usage of Skywork model requires [Skywork Community License](https://github.com/SkyworkAI/Skywork-Reward/blob/main/misc/Skywork%20Community%20License.pdf). The Skywork model supports commercial use. If you plan to use the Skywork model or its derivatives for commercial purposes, you must abide by terms and conditions within [Skywork Community License](https://github.com/SkyworkAI/Skywork-Reward/blob/main/misc/Skywork%20Community%20License.pdf).
|
47 |
+
|
48 |
+
|
49 |
+
## To-Do
|
50 |
+
+ [ ] Reward Model Usage code
|
51 |
+
+ [ ] Architecture diagram
|