Update README.md
Browse files
README.md
CHANGED
@@ -59,7 +59,7 @@ The proposition is **agnostic to specific choices of the training objective of O
|
|
59 |
For example, DPO already meets our assumption and serves as a strong variant, while in this work, we instantiate our implicit PRM with cross entropy (CE) loss due to memory efficiency:
|
60 |
|
61 |
$$
|
62 |
-
\small \mathcal{L}_{CE} = l \cdot \log \sigma \left( \beta \log \frac{\
|
63 |
$$
|
64 |
|
65 |
We applied \\(L_{CE}\\) to train implicit PRM. We used a learning rate of 5e-7 and a batch-size of 64 for training.
|
|
|
59 |
For example, DPO already meets our assumption and serves as a strong variant, while in this work, we instantiate our implicit PRM with cross entropy (CE) loss due to memory efficiency:
|
60 |
|
61 |
$$
|
62 |
+
\small \mathcal{L}_{CE} = l \cdot \log \sigma \left( \beta \log \frac{\pi_\phi(\mathbf{y})}{\pi_\text{ref}(\mathbf{y})} \right) + (1 - l) \cdot \log \left[ 1 - \sigma \left( \beta \log \frac{\pi_\phi(\mathbf{y})}{\pi_\text{ref}(\mathbf{y})} \right) \right]
|
63 |
$$
|
64 |
|
65 |
We applied \\(L_{CE}\\) to train implicit PRM. We used a learning rate of 5e-7 and a batch-size of 64 for training.
|