Update README.md
Browse files
README.md
CHANGED
@@ -72,8 +72,8 @@ We show an example leveraging **EurusPRM-Stage1** below:
|
|
72 |
coef=0.001
|
73 |
d = {'query':'111','answer':['111','222']
|
74 |
}
|
75 |
-
model = AutoModelForCausalLM.from_pretrained('PRIME-RL/EurusPRM-
|
76 |
-
tokenizer = AutoTokenizer.from_pretrained('PRIME-RL/EurusPRM-
|
77 |
ref_model = AutoModelForCausalLM.from_pretrained('Qwen/Qwen2.5-Math-7B-Instruct')
|
78 |
input_ids = tokenizer.apply_chat_template([
|
79 |
{"role": "user", "content": d["query"]},
|
|
|
72 |
coef=0.001
|
73 |
d = {'query':'111','answer':['111','222']
|
74 |
}
|
75 |
+
model = AutoModelForCausalLM.from_pretrained('PRIME-RL/EurusPRM-Stage1')
|
76 |
+
tokenizer = AutoTokenizer.from_pretrained('PRIME-RL/EurusPRM-Stage1')
|
77 |
ref_model = AutoModelForCausalLM.from_pretrained('Qwen/Qwen2.5-Math-7B-Instruct')
|
78 |
input_ids = tokenizer.apply_chat_template([
|
79 |
{"role": "user", "content": d["query"]},
|