Text Generation
Transformers
PyTorch
English
olmo2
conversational
Inference Endpoints
hamishivi commited on
Commit
965eb40
·
verified ·
1 Parent(s): bb6eb15

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -97,9 +97,9 @@ TODO
97
  Note we use a length-normalized variant of DPO for training.
98
 
99
  DPO:
100
- - **Learning Rate**: 8E-7 (7B), 7.5E-06 (13B)
101
  - **Beta**: 5
102
- - **Effective Batch Size:** 128 (7B), 128 (13B)
103
  - **Max. Sequence Length:** 2048
104
  - **Learning Rate Schedule:** Linear
105
  - **LR Warmup Ratio:** 0.1
 
97
  Note we use a length-normalized variant of DPO for training.
98
 
99
  DPO:
100
+ - **Learning Rate**: 8E-7 (7B, 13B)
101
  - **Beta**: 5
102
+ - **Effective Batch Size:** 128 (7B, 13B)
103
  - **Max. Sequence Length:** 2048
104
  - **Learning Rate Schedule:** Linear
105
  - **LR Warmup Ratio:** 0.1