arynkiewicz commited on
Commit
5d0c265
·
verified ·
1 Parent(s): ed1aefa

Model save

Browse files
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Mistral-7B-v0.1
3
+ tags:
4
+ - trl
5
+ - orpo
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: anydef-orpo
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # anydef-orpo
16
+
17
+ This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on an unknown dataset.
18
+
19
+ ## Model description
20
+
21
+ More information needed
22
+
23
+ ## Intended uses & limitations
24
+
25
+ More information needed
26
+
27
+ ## Training and evaluation data
28
+
29
+ More information needed
30
+
31
+ ## Training procedure
32
+
33
+ ### Training hyperparameters
34
+
35
+ The following hyperparameters were used during training:
36
+ - learning_rate: 5e-06
37
+ - train_batch_size: 8
38
+ - eval_batch_size: 8
39
+ - seed: 42
40
+ - distributed_type: multi-GPU
41
+ - num_devices: 8
42
+ - total_train_batch_size: 64
43
+ - total_eval_batch_size: 64
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: inverse_sqrt
46
+ - lr_scheduler_warmup_steps: 100
47
+ - num_epochs: 3
48
+
49
+ ### Training results
50
+
51
+
52
+
53
+ ### Framework versions
54
+
55
+ - Transformers 4.39.3
56
+ - Pytorch 2.1.2+cu121
57
+ - Datasets 2.18.0
58
+ - Tokenizers 0.15.2
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "train_loss": 0.1731818498170311,
4
+ "train_runtime": 10691.148,
5
+ "train_samples": 31470,
6
+ "train_samples_per_second": 8.831,
7
+ "train_steps_per_second": 0.138
8
+ }
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "mistralai/Mistral-7B-v0.1",
3
+ "architectures": [
4
+ "MistralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 14336,
13
+ "max_position_embeddings": 32768,
14
+ "model_type": "mistral",
15
+ "num_attention_heads": 32,
16
+ "num_hidden_layers": 32,
17
+ "num_key_value_heads": 8,
18
+ "rms_norm_eps": 1e-05,
19
+ "rope_theta": 10000.0,
20
+ "sliding_window": 4096,
21
+ "tie_word_embeddings": false,
22
+ "torch_dtype": "bfloat16",
23
+ "transformers_version": "4.39.3",
24
+ "use_cache": false,
25
+ "vocab_size": 32000
26
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.39.3"
6
+ }
model-00001-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a87011969946f5de6d3a2cec805ccfd353516a64e39b9587dae8a7d0918be8b4
3
+ size 4943162336
model-00002-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f810db6c477e69c7c6f2c7f9251f55f259a1c155b900448fe7b29e09535ede8
3
+ size 4999819336
model-00003-of-00003.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f615b2b7cf7e6e10e97fc78554e9e42f4d2007f4ee3ec22f36e973f9c20ecb63
3
+ size 4540516344
model.safetensors.index.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 14483464192
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00003-of-00003.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00003.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00003.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00003.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00003.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00003.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00003.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00003.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00003.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00002-of-00003.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00002-of-00003.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00002-of-00003.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00002-of-00003.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00002-of-00003.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00003.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
125
+ "model.layers.20.input_layernorm.weight": "model-00002-of-00003.safetensors",
126
+ "model.layers.20.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
127
+ "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
128
+ "model.layers.20.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
129
+ "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
130
+ "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
131
+ "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
132
+ "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
133
+ "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
134
+ "model.layers.21.input_layernorm.weight": "model-00002-of-00003.safetensors",
135
+ "model.layers.21.mlp.down_proj.weight": "model-00002-of-00003.safetensors",
136
+ "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00003.safetensors",
137
+ "model.layers.21.mlp.up_proj.weight": "model-00002-of-00003.safetensors",
138
+ "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00003.safetensors",
139
+ "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
140
+ "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
141
+ "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
142
+ "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
143
+ "model.layers.22.input_layernorm.weight": "model-00003-of-00003.safetensors",
144
+ "model.layers.22.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
145
+ "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
146
+ "model.layers.22.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
147
+ "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
148
+ "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
149
+ "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00003.safetensors",
150
+ "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
151
+ "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00003.safetensors",
152
+ "model.layers.23.input_layernorm.weight": "model-00003-of-00003.safetensors",
153
+ "model.layers.23.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
154
+ "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
155
+ "model.layers.23.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
156
+ "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
157
+ "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
158
+ "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
159
+ "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
160
+ "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
161
+ "model.layers.24.input_layernorm.weight": "model-00003-of-00003.safetensors",
162
+ "model.layers.24.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
163
+ "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
164
+ "model.layers.24.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
165
+ "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
166
+ "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
167
+ "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
168
+ "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
169
+ "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
170
+ "model.layers.25.input_layernorm.weight": "model-00003-of-00003.safetensors",
171
+ "model.layers.25.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
172
+ "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
173
+ "model.layers.25.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
174
+ "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
175
+ "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
176
+ "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
177
+ "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
178
+ "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
179
+ "model.layers.26.input_layernorm.weight": "model-00003-of-00003.safetensors",
180
+ "model.layers.26.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
181
+ "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
182
+ "model.layers.26.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
183
+ "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
184
+ "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
185
+ "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
186
+ "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
187
+ "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
188
+ "model.layers.27.input_layernorm.weight": "model-00003-of-00003.safetensors",
189
+ "model.layers.27.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
190
+ "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
191
+ "model.layers.27.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
192
+ "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
193
+ "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
194
+ "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
195
+ "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
196
+ "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
197
+ "model.layers.28.input_layernorm.weight": "model-00003-of-00003.safetensors",
198
+ "model.layers.28.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
199
+ "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
200
+ "model.layers.28.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
201
+ "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
202
+ "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
203
+ "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
204
+ "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
205
+ "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
206
+ "model.layers.29.input_layernorm.weight": "model-00003-of-00003.safetensors",
207
+ "model.layers.29.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
208
+ "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
209
+ "model.layers.29.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
210
+ "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
211
+ "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
212
+ "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
213
+ "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
214
+ "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
215
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00003.safetensors",
216
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
217
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
218
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
219
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
220
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
221
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
222
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
223
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
224
+ "model.layers.30.input_layernorm.weight": "model-00003-of-00003.safetensors",
225
+ "model.layers.30.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
226
+ "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
227
+ "model.layers.30.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
228
+ "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
229
+ "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
230
+ "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
231
+ "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
232
+ "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
233
+ "model.layers.31.input_layernorm.weight": "model-00003-of-00003.safetensors",
234
+ "model.layers.31.mlp.down_proj.weight": "model-00003-of-00003.safetensors",
235
+ "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00003.safetensors",
236
+ "model.layers.31.mlp.up_proj.weight": "model-00003-of-00003.safetensors",
237
+ "model.layers.31.post_attention_layernorm.weight": "model-00003-of-00003.safetensors",
238
+ "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
239
+ "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00003.safetensors",
240
+ "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00003.safetensors",
241
+ "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00003.safetensors",
242
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00003.safetensors",
243
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
244
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
245
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
246
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
247
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
248
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
249
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
250
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
251
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00003.safetensors",
252
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
253
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
254
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
255
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
256
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
257
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
258
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
259
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
260
+ "model.layers.6.input_layernorm.weight": "model-00001-of-00003.safetensors",
261
+ "model.layers.6.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
262
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
263
+ "model.layers.6.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
264
+ "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
265
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
266
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
267
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
268
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
269
+ "model.layers.7.input_layernorm.weight": "model-00001-of-00003.safetensors",
270
+ "model.layers.7.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
271
+ "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
272
+ "model.layers.7.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
273
+ "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
274
+ "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
275
+ "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
276
+ "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
277
+ "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
278
+ "model.layers.8.input_layernorm.weight": "model-00001-of-00003.safetensors",
279
+ "model.layers.8.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
280
+ "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
281
+ "model.layers.8.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
282
+ "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
283
+ "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
284
+ "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
285
+ "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
286
+ "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
287
+ "model.layers.9.input_layernorm.weight": "model-00001-of-00003.safetensors",
288
+ "model.layers.9.mlp.down_proj.weight": "model-00001-of-00003.safetensors",
289
+ "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00003.safetensors",
290
+ "model.layers.9.mlp.up_proj.weight": "model-00001-of-00003.safetensors",
291
+ "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00003.safetensors",
292
+ "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
293
+ "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00003.safetensors",
294
+ "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00003.safetensors",
295
+ "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00003.safetensors",
296
+ "model.norm.weight": "model-00003-of-00003.safetensors"
297
+ }
298
+ }
runs/Apr19_18-05-38_hgx1/events.out.tfevents.1713551066.hgx1.696066.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09a6539e106dff6c606989e5ab4d877ab373a39ba7db3a348ce3b81e7e6a97f4
3
+ size 131548
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "additional_special_tokens": [],
31
+ "bos_token": "<s>",
32
+ "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}",
33
+ "clean_up_tokenization_spaces": false,
34
+ "eos_token": "</s>",
35
+ "legacy": true,
36
+ "model_max_length": 2048,
37
+ "pad_token": "</s>",
38
+ "sp_model_kwargs": {},
39
+ "spaces_between_special_tokens": false,
40
+ "tokenizer_class": "LlamaTokenizer",
41
+ "unk_token": "<unk>",
42
+ "use_default_system_prompt": false
43
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "train_loss": 0.1731818498170311,
4
+ "train_runtime": 10691.148,
5
+ "train_samples": 31470,
6
+ "train_samples_per_second": 8.831,
7
+ "train_steps_per_second": 0.138
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,2676 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1476,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.02,
13
+ "grad_norm": 420.0,
14
+ "learning_rate": 5.000000000000001e-07,
15
+ "log_odds_chosen": 0.4207151532173157,
16
+ "log_odds_ratio": -0.5201840400695801,
17
+ "logits/chosen": -2.347763776779175,
18
+ "logits/rejected": -2.3541953563690186,
19
+ "logps/chosen": -1.7608455419540405,
20
+ "logps/rejected": -2.1218907833099365,
21
+ "loss": 3.9956,
22
+ "nll_loss": 4.183717250823975,
23
+ "rewards/accuracies": 0.8999999761581421,
24
+ "rewards/chosen": -0.08804227411746979,
25
+ "rewards/margins": 0.018052268773317337,
26
+ "rewards/rejected": -0.10609455406665802,
27
+ "step": 10
28
+ },
29
+ {
30
+ "epoch": 0.04,
31
+ "grad_norm": 44.25,
32
+ "learning_rate": 1.0000000000000002e-06,
33
+ "log_odds_chosen": 0.5392915606498718,
34
+ "log_odds_ratio": -0.4769781231880188,
35
+ "logits/chosen": -2.866349458694458,
36
+ "logits/rejected": -2.866671323776245,
37
+ "logps/chosen": -1.5143439769744873,
38
+ "logps/rejected": -1.9585195779800415,
39
+ "loss": 2.0453,
40
+ "nll_loss": 1.9479516744613647,
41
+ "rewards/accuracies": 0.925000011920929,
42
+ "rewards/chosen": -0.07571719586849213,
43
+ "rewards/margins": 0.02220877632498741,
44
+ "rewards/rejected": -0.09792597591876984,
45
+ "step": 20
46
+ },
47
+ {
48
+ "epoch": 0.06,
49
+ "grad_norm": 7.5,
50
+ "learning_rate": 1.5e-06,
51
+ "log_odds_chosen": 0.9706680178642273,
52
+ "log_odds_ratio": -0.3677331805229187,
53
+ "logits/chosen": -2.6566712856292725,
54
+ "logits/rejected": -2.660308361053467,
55
+ "logps/chosen": -0.7488349676132202,
56
+ "logps/rejected": -1.312819480895996,
57
+ "loss": 1.0008,
58
+ "nll_loss": 0.8921732902526855,
59
+ "rewards/accuracies": 0.9624999761581421,
60
+ "rewards/chosen": -0.03744174912571907,
61
+ "rewards/margins": 0.028199229389429092,
62
+ "rewards/rejected": -0.06564097106456757,
63
+ "step": 30
64
+ },
65
+ {
66
+ "epoch": 0.08,
67
+ "grad_norm": 7.40625,
68
+ "learning_rate": 2.0000000000000003e-06,
69
+ "log_odds_chosen": 2.282021999359131,
70
+ "log_odds_ratio": -0.20370522141456604,
71
+ "logits/chosen": -2.1984050273895264,
72
+ "logits/rejected": -2.202855110168457,
73
+ "logps/chosen": -0.27147307991981506,
74
+ "logps/rejected": -1.2435765266418457,
75
+ "loss": 0.8107,
76
+ "nll_loss": 0.683063268661499,
77
+ "rewards/accuracies": 0.949999988079071,
78
+ "rewards/chosen": -0.013573653995990753,
79
+ "rewards/margins": 0.04860517010092735,
80
+ "rewards/rejected": -0.06217882037162781,
81
+ "step": 40
82
+ },
83
+ {
84
+ "epoch": 0.1,
85
+ "grad_norm": 6.375,
86
+ "learning_rate": 2.5e-06,
87
+ "log_odds_chosen": 2.8677659034729004,
88
+ "log_odds_ratio": -0.14154179394245148,
89
+ "logits/chosen": -2.279298782348633,
90
+ "logits/rejected": -2.2862045764923096,
91
+ "logps/chosen": -0.216520756483078,
92
+ "logps/rejected": -1.3591465950012207,
93
+ "loss": 0.7978,
94
+ "nll_loss": 0.828228771686554,
95
+ "rewards/accuracies": 0.987500011920929,
96
+ "rewards/chosen": -0.01082603819668293,
97
+ "rewards/margins": 0.057131290435791016,
98
+ "rewards/rejected": -0.0679573267698288,
99
+ "step": 50
100
+ },
101
+ {
102
+ "epoch": 0.12,
103
+ "grad_norm": 4.3125,
104
+ "learning_rate": 3e-06,
105
+ "log_odds_chosen": 3.2677154541015625,
106
+ "log_odds_ratio": -0.12433552742004395,
107
+ "logits/chosen": -2.2281601428985596,
108
+ "logits/rejected": -2.2347497940063477,
109
+ "logps/chosen": -0.20724907517433167,
110
+ "logps/rejected": -1.5352025032043457,
111
+ "loss": 0.7919,
112
+ "nll_loss": 0.8021117448806763,
113
+ "rewards/accuracies": 0.9624999761581421,
114
+ "rewards/chosen": -0.010362453758716583,
115
+ "rewards/margins": 0.06639767438173294,
116
+ "rewards/rejected": -0.07676012814044952,
117
+ "step": 60
118
+ },
119
+ {
120
+ "epoch": 0.14,
121
+ "grad_norm": 4.21875,
122
+ "learning_rate": 3.5e-06,
123
+ "log_odds_chosen": 3.9659018516540527,
124
+ "log_odds_ratio": -0.11224009841680527,
125
+ "logits/chosen": -2.0779483318328857,
126
+ "logits/rejected": -2.0807433128356934,
127
+ "logps/chosen": -0.1909479945898056,
128
+ "logps/rejected": -1.9270519018173218,
129
+ "loss": 0.7197,
130
+ "nll_loss": 0.6386193037033081,
131
+ "rewards/accuracies": 0.9750000238418579,
132
+ "rewards/chosen": -0.00954739935696125,
133
+ "rewards/margins": 0.08680519461631775,
134
+ "rewards/rejected": -0.09635259956121445,
135
+ "step": 70
136
+ },
137
+ {
138
+ "epoch": 0.16,
139
+ "grad_norm": 4.15625,
140
+ "learning_rate": 4.000000000000001e-06,
141
+ "log_odds_chosen": 3.9601542949676514,
142
+ "log_odds_ratio": -0.11061631143093109,
143
+ "logits/chosen": -2.1697168350219727,
144
+ "logits/rejected": -2.1794517040252686,
145
+ "logps/chosen": -0.185487762093544,
146
+ "logps/rejected": -1.8811283111572266,
147
+ "loss": 0.7033,
148
+ "nll_loss": 0.7545000314712524,
149
+ "rewards/accuracies": 0.987500011920929,
150
+ "rewards/chosen": -0.009274388663470745,
151
+ "rewards/margins": 0.08478203415870667,
152
+ "rewards/rejected": -0.09405642002820969,
153
+ "step": 80
154
+ },
155
+ {
156
+ "epoch": 0.18,
157
+ "grad_norm": 4.71875,
158
+ "learning_rate": 4.5e-06,
159
+ "log_odds_chosen": 4.599364280700684,
160
+ "log_odds_ratio": -0.07719282805919647,
161
+ "logits/chosen": -2.150642156600952,
162
+ "logits/rejected": -2.15342378616333,
163
+ "logps/chosen": -0.15149493515491486,
164
+ "logps/rejected": -2.1990609169006348,
165
+ "loss": 0.6116,
166
+ "nll_loss": 0.6107597947120667,
167
+ "rewards/accuracies": 0.987500011920929,
168
+ "rewards/chosen": -0.007574746850878,
169
+ "rewards/margins": 0.10237829387187958,
170
+ "rewards/rejected": -0.10995303094387054,
171
+ "step": 90
172
+ },
173
+ {
174
+ "epoch": 0.2,
175
+ "grad_norm": 4.53125,
176
+ "learning_rate": 5e-06,
177
+ "log_odds_chosen": 4.817965507507324,
178
+ "log_odds_ratio": -0.07800901681184769,
179
+ "logits/chosen": -2.1689393520355225,
180
+ "logits/rejected": -2.1756036281585693,
181
+ "logps/chosen": -0.1730736345052719,
182
+ "logps/rejected": -2.246196746826172,
183
+ "loss": 0.6043,
184
+ "nll_loss": 0.6302344799041748,
185
+ "rewards/accuracies": 1.0,
186
+ "rewards/chosen": -0.008653681725263596,
187
+ "rewards/margins": 0.10365615785121918,
188
+ "rewards/rejected": -0.11230983585119247,
189
+ "step": 100
190
+ },
191
+ {
192
+ "epoch": 0.22,
193
+ "grad_norm": 5.625,
194
+ "learning_rate": 4.767312946227961e-06,
195
+ "log_odds_chosen": 4.97894287109375,
196
+ "log_odds_ratio": -0.10877744108438492,
197
+ "logits/chosen": -2.2094526290893555,
198
+ "logits/rejected": -2.216020107269287,
199
+ "logps/chosen": -0.18459849059581757,
200
+ "logps/rejected": -2.573103904724121,
201
+ "loss": 0.5724,
202
+ "nll_loss": 0.6197156310081482,
203
+ "rewards/accuracies": 0.9375,
204
+ "rewards/chosen": -0.009229923598468304,
205
+ "rewards/margins": 0.11942527443170547,
206
+ "rewards/rejected": -0.12865519523620605,
207
+ "step": 110
208
+ },
209
+ {
210
+ "epoch": 0.24,
211
+ "grad_norm": 5.8125,
212
+ "learning_rate": 4.564354645876385e-06,
213
+ "log_odds_chosen": 5.132683277130127,
214
+ "log_odds_ratio": -0.13557349145412445,
215
+ "logits/chosen": -2.178772449493408,
216
+ "logits/rejected": -2.184910774230957,
217
+ "logps/chosen": -0.154138445854187,
218
+ "logps/rejected": -2.2600817680358887,
219
+ "loss": 0.4748,
220
+ "nll_loss": 0.459114134311676,
221
+ "rewards/accuracies": 0.949999988079071,
222
+ "rewards/chosen": -0.0077069224789738655,
223
+ "rewards/margins": 0.10529716312885284,
224
+ "rewards/rejected": -0.11300408840179443,
225
+ "step": 120
226
+ },
227
+ {
228
+ "epoch": 0.26,
229
+ "grad_norm": 5.0,
230
+ "learning_rate": 4.385290096535147e-06,
231
+ "log_odds_chosen": 5.768979072570801,
232
+ "log_odds_ratio": -0.08052898198366165,
233
+ "logits/chosen": -2.101267099380493,
234
+ "logits/rejected": -2.104369878768921,
235
+ "logps/chosen": -0.1255403310060501,
236
+ "logps/rejected": -2.850447416305542,
237
+ "loss": 0.4379,
238
+ "nll_loss": 0.40499457716941833,
239
+ "rewards/accuracies": 0.9750000238418579,
240
+ "rewards/chosen": -0.006277016364037991,
241
+ "rewards/margins": 0.13624534010887146,
242
+ "rewards/rejected": -0.14252236485481262,
243
+ "step": 130
244
+ },
245
+ {
246
+ "epoch": 0.28,
247
+ "grad_norm": 6.0,
248
+ "learning_rate": 4.2257712736425835e-06,
249
+ "log_odds_chosen": 5.547473907470703,
250
+ "log_odds_ratio": -0.056151650846004486,
251
+ "logits/chosen": -2.0569746494293213,
252
+ "logits/rejected": -2.059227466583252,
253
+ "logps/chosen": -0.12779685854911804,
254
+ "logps/rejected": -2.394996166229248,
255
+ "loss": 0.4122,
256
+ "nll_loss": 0.33014988899230957,
257
+ "rewards/accuracies": 1.0,
258
+ "rewards/chosen": -0.006389842834323645,
259
+ "rewards/margins": 0.1133599653840065,
260
+ "rewards/rejected": -0.11974982172250748,
261
+ "step": 140
262
+ },
263
+ {
264
+ "epoch": 0.3,
265
+ "grad_norm": 5.21875,
266
+ "learning_rate": 4.082482904638631e-06,
267
+ "log_odds_chosen": 5.67286491394043,
268
+ "log_odds_ratio": -0.06435154378414154,
269
+ "logits/chosen": -2.166313886642456,
270
+ "logits/rejected": -2.1680521965026855,
271
+ "logps/chosen": -0.13122043013572693,
272
+ "logps/rejected": -2.701577663421631,
273
+ "loss": 0.4055,
274
+ "nll_loss": 0.4321494996547699,
275
+ "rewards/accuracies": 0.9750000238418579,
276
+ "rewards/chosen": -0.0065610213205218315,
277
+ "rewards/margins": 0.12851786613464355,
278
+ "rewards/rejected": -0.13507887721061707,
279
+ "step": 150
280
+ },
281
+ {
282
+ "epoch": 0.33,
283
+ "grad_norm": 6.5625,
284
+ "learning_rate": 3.952847075210474e-06,
285
+ "log_odds_chosen": 5.581357002258301,
286
+ "log_odds_ratio": -0.094394750893116,
287
+ "logits/chosen": -2.121497392654419,
288
+ "logits/rejected": -2.1272549629211426,
289
+ "logps/chosen": -0.15410242974758148,
290
+ "logps/rejected": -2.376707077026367,
291
+ "loss": 0.3695,
292
+ "nll_loss": 0.43908724188804626,
293
+ "rewards/accuracies": 0.987500011920929,
294
+ "rewards/chosen": -0.007705122232437134,
295
+ "rewards/margins": 0.1111302375793457,
296
+ "rewards/rejected": -0.11883536726236343,
297
+ "step": 160
298
+ },
299
+ {
300
+ "epoch": 0.35,
301
+ "grad_norm": 6.0,
302
+ "learning_rate": 3.834824944236852e-06,
303
+ "log_odds_chosen": 6.281979560852051,
304
+ "log_odds_ratio": -0.059505023062229156,
305
+ "logits/chosen": -2.1019725799560547,
306
+ "logits/rejected": -2.1044390201568604,
307
+ "logps/chosen": -0.15183117985725403,
308
+ "logps/rejected": -3.044116497039795,
309
+ "loss": 0.3264,
310
+ "nll_loss": 0.3386436998844147,
311
+ "rewards/accuracies": 0.987500011920929,
312
+ "rewards/chosen": -0.007591559551656246,
313
+ "rewards/margins": 0.14461426436901093,
314
+ "rewards/rejected": -0.15220583975315094,
315
+ "step": 170
316
+ },
317
+ {
318
+ "epoch": 0.37,
319
+ "grad_norm": 6.3125,
320
+ "learning_rate": 3.72677996249965e-06,
321
+ "log_odds_chosen": 6.111880302429199,
322
+ "log_odds_ratio": -0.07030317932367325,
323
+ "logits/chosen": -2.0706937313079834,
324
+ "logits/rejected": -2.067783832550049,
325
+ "logps/chosen": -0.14989063143730164,
326
+ "logps/rejected": -3.172281265258789,
327
+ "loss": 0.3324,
328
+ "nll_loss": 0.38789159059524536,
329
+ "rewards/accuracies": 0.987500011920929,
330
+ "rewards/chosen": -0.007494532503187656,
331
+ "rewards/margins": 0.15111951529979706,
332
+ "rewards/rejected": -0.15861406922340393,
333
+ "step": 180
334
+ },
335
+ {
336
+ "epoch": 0.39,
337
+ "grad_norm": 6.0625,
338
+ "learning_rate": 3.6273812505500587e-06,
339
+ "log_odds_chosen": 6.445823669433594,
340
+ "log_odds_ratio": -0.05266156792640686,
341
+ "logits/chosen": -2.0448713302612305,
342
+ "logits/rejected": -2.0481276512145996,
343
+ "logps/chosen": -0.13432614505290985,
344
+ "logps/rejected": -2.9074063301086426,
345
+ "loss": 0.3063,
346
+ "nll_loss": 0.28705888986587524,
347
+ "rewards/accuracies": 1.0,
348
+ "rewards/chosen": -0.006716308183968067,
349
+ "rewards/margins": 0.13865402340888977,
350
+ "rewards/rejected": -0.14537033438682556,
351
+ "step": 190
352
+ },
353
+ {
354
+ "epoch": 0.41,
355
+ "grad_norm": 6.0,
356
+ "learning_rate": 3.5355339059327378e-06,
357
+ "log_odds_chosen": 5.775099754333496,
358
+ "log_odds_ratio": -0.08722616732120514,
359
+ "logits/chosen": -2.087759256362915,
360
+ "logits/rejected": -2.0876097679138184,
361
+ "logps/chosen": -0.11597369611263275,
362
+ "logps/rejected": -2.6853129863739014,
363
+ "loss": 0.2962,
364
+ "nll_loss": 0.31049779057502747,
365
+ "rewards/accuracies": 0.987500011920929,
366
+ "rewards/chosen": -0.005798685364425182,
367
+ "rewards/margins": 0.12846694886684418,
368
+ "rewards/rejected": -0.13426566123962402,
369
+ "step": 200
370
+ },
371
+ {
372
+ "epoch": 0.43,
373
+ "grad_norm": 5.90625,
374
+ "learning_rate": 3.450327796711771e-06,
375
+ "log_odds_chosen": 7.5723066329956055,
376
+ "log_odds_ratio": -0.040629759430885315,
377
+ "logits/chosen": -1.973818063735962,
378
+ "logits/rejected": -1.976941704750061,
379
+ "logps/chosen": -0.09340523928403854,
380
+ "logps/rejected": -3.5397675037384033,
381
+ "loss": 0.2637,
382
+ "nll_loss": 0.20889191329479218,
383
+ "rewards/accuracies": 0.987500011920929,
384
+ "rewards/chosen": -0.00467026187106967,
385
+ "rewards/margins": 0.17231810092926025,
386
+ "rewards/rejected": -0.1769883632659912,
387
+ "step": 210
388
+ },
389
+ {
390
+ "epoch": 0.45,
391
+ "grad_norm": 6.25,
392
+ "learning_rate": 3.3709993123162106e-06,
393
+ "log_odds_chosen": 7.618361473083496,
394
+ "log_odds_ratio": -0.074308380484581,
395
+ "logits/chosen": -2.0433554649353027,
396
+ "logits/rejected": -2.0466837882995605,
397
+ "logps/chosen": -0.09071488678455353,
398
+ "logps/rejected": -3.5269036293029785,
399
+ "loss": 0.2727,
400
+ "nll_loss": 0.24802179634571075,
401
+ "rewards/accuracies": 0.9624999761581421,
402
+ "rewards/chosen": -0.004535744432359934,
403
+ "rewards/margins": 0.17180944979190826,
404
+ "rewards/rejected": -0.17634519934654236,
405
+ "step": 220
406
+ },
407
+ {
408
+ "epoch": 0.47,
409
+ "grad_norm": 5.875,
410
+ "learning_rate": 3.296902366978936e-06,
411
+ "log_odds_chosen": 6.698166847229004,
412
+ "log_odds_ratio": -0.04654809832572937,
413
+ "logits/chosen": -1.968505620956421,
414
+ "logits/rejected": -1.9712941646575928,
415
+ "logps/chosen": -0.1543513834476471,
416
+ "logps/rejected": -3.186583995819092,
417
+ "loss": 0.2015,
418
+ "nll_loss": 0.1845930814743042,
419
+ "rewards/accuracies": 1.0,
420
+ "rewards/chosen": -0.00771756935864687,
421
+ "rewards/margins": 0.15161164104938507,
422
+ "rewards/rejected": -0.15932922065258026,
423
+ "step": 230
424
+ },
425
+ {
426
+ "epoch": 0.49,
427
+ "grad_norm": 5.28125,
428
+ "learning_rate": 3.2274861218395142e-06,
429
+ "log_odds_chosen": 7.480630397796631,
430
+ "log_odds_ratio": -0.06602514535188675,
431
+ "logits/chosen": -1.995495080947876,
432
+ "logits/rejected": -1.9951270818710327,
433
+ "logps/chosen": -0.12399391829967499,
434
+ "logps/rejected": -3.7262351512908936,
435
+ "loss": 0.2181,
436
+ "nll_loss": 0.1835182011127472,
437
+ "rewards/accuracies": 0.987500011920929,
438
+ "rewards/chosen": -0.006199696101248264,
439
+ "rewards/margins": 0.1801120638847351,
440
+ "rewards/rejected": -0.1863117516040802,
441
+ "step": 240
442
+ },
443
+ {
444
+ "epoch": 0.51,
445
+ "grad_norm": 6.65625,
446
+ "learning_rate": 3.1622776601683796e-06,
447
+ "log_odds_chosen": 7.167886257171631,
448
+ "log_odds_ratio": -0.020891310647130013,
449
+ "logits/chosen": -1.9453086853027344,
450
+ "logits/rejected": -1.9443309307098389,
451
+ "logps/chosen": -0.07993815094232559,
452
+ "logps/rejected": -3.2065398693084717,
453
+ "loss": 0.1966,
454
+ "nll_loss": 0.21496787667274475,
455
+ "rewards/accuracies": 1.0,
456
+ "rewards/chosen": -0.003996907267719507,
457
+ "rewards/margins": 0.15633007884025574,
458
+ "rewards/rejected": -0.1603269875049591,
459
+ "step": 250
460
+ },
461
+ {
462
+ "epoch": 0.53,
463
+ "grad_norm": 7.03125,
464
+ "learning_rate": 3.1008683647302113e-06,
465
+ "log_odds_chosen": 6.189971446990967,
466
+ "log_odds_ratio": -0.06337615847587585,
467
+ "logits/chosen": -1.983068823814392,
468
+ "logits/rejected": -1.9870706796646118,
469
+ "logps/chosen": -0.13599030673503876,
470
+ "logps/rejected": -2.809053897857666,
471
+ "loss": 0.1981,
472
+ "nll_loss": 0.23529119789600372,
473
+ "rewards/accuracies": 0.987500011920929,
474
+ "rewards/chosen": -0.0067995162680745125,
475
+ "rewards/margins": 0.1336531937122345,
476
+ "rewards/rejected": -0.14045271277427673,
477
+ "step": 260
478
+ },
479
+ {
480
+ "epoch": 0.55,
481
+ "grad_norm": 5.125,
482
+ "learning_rate": 3.0429030972509227e-06,
483
+ "log_odds_chosen": 6.417479038238525,
484
+ "log_odds_ratio": -0.05179642513394356,
485
+ "logits/chosen": -1.9953663349151611,
486
+ "logits/rejected": -1.9997583627700806,
487
+ "logps/chosen": -0.12870252132415771,
488
+ "logps/rejected": -2.9075326919555664,
489
+ "loss": 0.1753,
490
+ "nll_loss": 0.22951626777648926,
491
+ "rewards/accuracies": 0.987500011920929,
492
+ "rewards/chosen": -0.006435126066207886,
493
+ "rewards/margins": 0.13894154131412506,
494
+ "rewards/rejected": -0.14537665247917175,
495
+ "step": 270
496
+ },
497
+ {
498
+ "epoch": 0.57,
499
+ "grad_norm": 5.90625,
500
+ "learning_rate": 2.988071523335984e-06,
501
+ "log_odds_chosen": 6.808651924133301,
502
+ "log_odds_ratio": -0.06405957788228989,
503
+ "logits/chosen": -2.0137691497802734,
504
+ "logits/rejected": -2.015270709991455,
505
+ "logps/chosen": -0.11106196790933609,
506
+ "logps/rejected": -3.0401816368103027,
507
+ "loss": 0.1885,
508
+ "nll_loss": 0.1923011988401413,
509
+ "rewards/accuracies": 0.987500011920929,
510
+ "rewards/chosen": -0.005553097929805517,
511
+ "rewards/margins": 0.14645597338676453,
512
+ "rewards/rejected": -0.15200909972190857,
513
+ "step": 280
514
+ },
515
+ {
516
+ "epoch": 0.59,
517
+ "grad_norm": 5.375,
518
+ "learning_rate": 2.9361010975735177e-06,
519
+ "log_odds_chosen": 8.074934959411621,
520
+ "log_odds_ratio": -0.022188549861311913,
521
+ "logits/chosen": -1.94329833984375,
522
+ "logits/rejected": -1.943985939025879,
523
+ "logps/chosen": -0.06887535750865936,
524
+ "logps/rejected": -3.7694950103759766,
525
+ "loss": 0.1729,
526
+ "nll_loss": 0.15832534432411194,
527
+ "rewards/accuracies": 0.987500011920929,
528
+ "rewards/chosen": -0.0034437670838087797,
529
+ "rewards/margins": 0.185030996799469,
530
+ "rewards/rejected": -0.18847474455833435,
531
+ "step": 290
532
+ },
533
+ {
534
+ "epoch": 0.61,
535
+ "grad_norm": 4.875,
536
+ "learning_rate": 2.8867513459481293e-06,
537
+ "log_odds_chosen": 6.823982238769531,
538
+ "log_odds_ratio": -0.04056983441114426,
539
+ "logits/chosen": -1.9236243963241577,
540
+ "logits/rejected": -1.9263416528701782,
541
+ "logps/chosen": -0.15296845138072968,
542
+ "logps/rejected": -3.20212984085083,
543
+ "loss": 0.1766,
544
+ "nll_loss": 0.1584191620349884,
545
+ "rewards/accuracies": 1.0,
546
+ "rewards/chosen": -0.0076484219171106815,
547
+ "rewards/margins": 0.152458056807518,
548
+ "rewards/rejected": -0.16010649502277374,
549
+ "step": 300
550
+ },
551
+ {
552
+ "epoch": 0.63,
553
+ "grad_norm": 5.0,
554
+ "learning_rate": 2.839809171235324e-06,
555
+ "log_odds_chosen": 7.289086818695068,
556
+ "log_odds_ratio": -0.07533205300569534,
557
+ "logits/chosen": -2.0198066234588623,
558
+ "logits/rejected": -2.022125720977783,
559
+ "logps/chosen": -0.12243582308292389,
560
+ "logps/rejected": -3.1500906944274902,
561
+ "loss": 0.1636,
562
+ "nll_loss": 0.1975417137145996,
563
+ "rewards/accuracies": 0.9624999761581421,
564
+ "rewards/chosen": -0.0061217909678816795,
565
+ "rewards/margins": 0.15138274431228638,
566
+ "rewards/rejected": -0.15750452876091003,
567
+ "step": 310
568
+ },
569
+ {
570
+ "epoch": 0.65,
571
+ "grad_norm": 9.0,
572
+ "learning_rate": 2.7950849718747376e-06,
573
+ "log_odds_chosen": 6.9169487953186035,
574
+ "log_odds_ratio": -0.03769087791442871,
575
+ "logits/chosen": -1.9343087673187256,
576
+ "logits/rejected": -1.9343109130859375,
577
+ "logps/chosen": -0.10367435216903687,
578
+ "logps/rejected": -3.0478665828704834,
579
+ "loss": 0.1696,
580
+ "nll_loss": 0.15094847977161407,
581
+ "rewards/accuracies": 0.987500011920929,
582
+ "rewards/chosen": -0.005183717701584101,
583
+ "rewards/margins": 0.14720964431762695,
584
+ "rewards/rejected": -0.15239334106445312,
585
+ "step": 320
586
+ },
587
+ {
588
+ "epoch": 0.67,
589
+ "grad_norm": 5.90625,
590
+ "learning_rate": 2.752409412815902e-06,
591
+ "log_odds_chosen": 6.86618185043335,
592
+ "log_odds_ratio": -0.038672395050525665,
593
+ "logits/chosen": -2.047116279602051,
594
+ "logits/rejected": -2.044132947921753,
595
+ "logps/chosen": -0.13211274147033691,
596
+ "logps/rejected": -3.158360004425049,
597
+ "loss": 0.1625,
598
+ "nll_loss": 0.17262741923332214,
599
+ "rewards/accuracies": 0.987500011920929,
600
+ "rewards/chosen": -0.006605637725442648,
601
+ "rewards/margins": 0.15131238102912903,
602
+ "rewards/rejected": -0.15791800618171692,
603
+ "step": 330
604
+ },
605
+ {
606
+ "epoch": 0.69,
607
+ "grad_norm": 5.40625,
608
+ "learning_rate": 2.711630722733202e-06,
609
+ "log_odds_chosen": 7.859139442443848,
610
+ "log_odds_ratio": -0.05241536349058151,
611
+ "logits/chosen": -1.951807975769043,
612
+ "logits/rejected": -1.9466063976287842,
613
+ "logps/chosen": -0.08200397342443466,
614
+ "logps/rejected": -3.5420470237731934,
615
+ "loss": 0.1465,
616
+ "nll_loss": 0.1444871723651886,
617
+ "rewards/accuracies": 0.9750000238418579,
618
+ "rewards/chosen": -0.004100198391824961,
619
+ "rewards/margins": 0.17300216853618622,
620
+ "rewards/rejected": -0.17710235714912415,
621
+ "step": 340
622
+ },
623
+ {
624
+ "epoch": 0.71,
625
+ "grad_norm": 5.71875,
626
+ "learning_rate": 2.6726124191242444e-06,
627
+ "log_odds_chosen": 7.041175842285156,
628
+ "log_odds_ratio": -0.03843230381608009,
629
+ "logits/chosen": -1.9341367483139038,
630
+ "logits/rejected": -1.9377237558364868,
631
+ "logps/chosen": -0.105894073843956,
632
+ "logps/rejected": -2.9890971183776855,
633
+ "loss": 0.145,
634
+ "nll_loss": 0.15412884950637817,
635
+ "rewards/accuracies": 0.987500011920929,
636
+ "rewards/chosen": -0.005294703878462315,
637
+ "rewards/margins": 0.14416013658046722,
638
+ "rewards/rejected": -0.14945486187934875,
639
+ "step": 350
640
+ },
641
+ {
642
+ "epoch": 0.73,
643
+ "grad_norm": 6.90625,
644
+ "learning_rate": 2.6352313834736496e-06,
645
+ "log_odds_chosen": 6.8190507888793945,
646
+ "log_odds_ratio": -0.04403644800186157,
647
+ "logits/chosen": -1.984197974205017,
648
+ "logits/rejected": -1.987479567527771,
649
+ "logps/chosen": -0.11883167922496796,
650
+ "logps/rejected": -3.1214818954467773,
651
+ "loss": 0.1577,
652
+ "nll_loss": 0.12197007238864899,
653
+ "rewards/accuracies": 0.987500011920929,
654
+ "rewards/chosen": -0.0059415847063064575,
655
+ "rewards/margins": 0.15013250708580017,
656
+ "rewards/rejected": -0.15607409179210663,
657
+ "step": 360
658
+ },
659
+ {
660
+ "epoch": 0.75,
661
+ "grad_norm": 5.28125,
662
+ "learning_rate": 2.599376224550182e-06,
663
+ "log_odds_chosen": 6.533886909484863,
664
+ "log_odds_ratio": -0.03479010984301567,
665
+ "logits/chosen": -1.898244857788086,
666
+ "logits/rejected": -1.8991117477416992,
667
+ "logps/chosen": -0.10942695289850235,
668
+ "logps/rejected": -2.946617841720581,
669
+ "loss": 0.1199,
670
+ "nll_loss": 0.11430631577968597,
671
+ "rewards/accuracies": 1.0,
672
+ "rewards/chosen": -0.005471347831189632,
673
+ "rewards/margins": 0.14185954630374908,
674
+ "rewards/rejected": -0.14733090996742249,
675
+ "step": 370
676
+ },
677
+ {
678
+ "epoch": 0.77,
679
+ "grad_norm": 5.6875,
680
+ "learning_rate": 2.564945880212886e-06,
681
+ "log_odds_chosen": 6.550747871398926,
682
+ "log_odds_ratio": -0.0334211066365242,
683
+ "logits/chosen": -1.9049135446548462,
684
+ "logits/rejected": -1.9040100574493408,
685
+ "logps/chosen": -0.12769648432731628,
686
+ "logps/rejected": -3.1679463386535645,
687
+ "loss": 0.1321,
688
+ "nll_loss": 0.12015843391418457,
689
+ "rewards/accuracies": 0.987500011920929,
690
+ "rewards/chosen": -0.006384824402630329,
691
+ "rewards/margins": 0.15201251208782196,
692
+ "rewards/rejected": -0.15839733183383942,
693
+ "step": 380
694
+ },
695
+ {
696
+ "epoch": 0.79,
697
+ "grad_norm": 4.46875,
698
+ "learning_rate": 2.5318484177091667e-06,
699
+ "log_odds_chosen": 6.963167667388916,
700
+ "log_odds_ratio": -0.07767531275749207,
701
+ "logits/chosen": -1.9564456939697266,
702
+ "logits/rejected": -1.957360029220581,
703
+ "logps/chosen": -0.09807981550693512,
704
+ "logps/rejected": -2.8958752155303955,
705
+ "loss": 0.1334,
706
+ "nll_loss": 0.13979199528694153,
707
+ "rewards/accuracies": 0.9750000238418579,
708
+ "rewards/chosen": -0.004903990775346756,
709
+ "rewards/margins": 0.13988977670669556,
710
+ "rewards/rejected": -0.1447937786579132,
711
+ "step": 390
712
+ },
713
+ {
714
+ "epoch": 0.81,
715
+ "grad_norm": 6.46875,
716
+ "learning_rate": 2.5e-06,
717
+ "log_odds_chosen": 7.737925052642822,
718
+ "log_odds_ratio": -0.011308287270367146,
719
+ "logits/chosen": -1.8833544254302979,
720
+ "logits/rejected": -1.8816306591033936,
721
+ "logps/chosen": -0.07808590680360794,
722
+ "logps/rejected": -3.517468214035034,
723
+ "loss": 0.116,
724
+ "nll_loss": 0.09308944642543793,
725
+ "rewards/accuracies": 1.0,
726
+ "rewards/chosen": -0.0039042953867465258,
727
+ "rewards/margins": 0.17196913063526154,
728
+ "rewards/rejected": -0.17587339878082275,
729
+ "step": 400
730
+ },
731
+ {
732
+ "epoch": 0.83,
733
+ "grad_norm": 5.09375,
734
+ "learning_rate": 2.4693239916239746e-06,
735
+ "log_odds_chosen": 8.026070594787598,
736
+ "log_odds_ratio": -0.009079935029149055,
737
+ "logits/chosen": -1.8916471004486084,
738
+ "logits/rejected": -1.8905131816864014,
739
+ "logps/chosen": -0.08094353973865509,
740
+ "logps/rejected": -3.705895185470581,
741
+ "loss": 0.1292,
742
+ "nll_loss": 0.10612641274929047,
743
+ "rewards/accuracies": 1.0,
744
+ "rewards/chosen": -0.00404717680066824,
745
+ "rewards/margins": 0.18124757707118988,
746
+ "rewards/rejected": -0.1852947473526001,
747
+ "step": 410
748
+ },
749
+ {
750
+ "epoch": 0.85,
751
+ "grad_norm": 4.15625,
752
+ "learning_rate": 2.4397501823713327e-06,
753
+ "log_odds_chosen": 6.706048488616943,
754
+ "log_odds_ratio": -0.04705270379781723,
755
+ "logits/chosen": -1.9574733972549438,
756
+ "logits/rejected": -1.9553906917572021,
757
+ "logps/chosen": -0.13810355961322784,
758
+ "logps/rejected": -3.244007110595703,
759
+ "loss": 0.1116,
760
+ "nll_loss": 0.13714413344860077,
761
+ "rewards/accuracies": 0.9750000238418579,
762
+ "rewards/chosen": -0.00690517807379365,
763
+ "rewards/margins": 0.15529517829418182,
764
+ "rewards/rejected": -0.16220036149024963,
765
+ "step": 420
766
+ },
767
+ {
768
+ "epoch": 0.87,
769
+ "grad_norm": 4.9375,
770
+ "learning_rate": 2.411214110852061e-06,
771
+ "log_odds_chosen": 7.185335636138916,
772
+ "log_odds_ratio": -0.049024853855371475,
773
+ "logits/chosen": -1.9474496841430664,
774
+ "logits/rejected": -1.94386887550354,
775
+ "logps/chosen": -0.12190796434879303,
776
+ "logps/rejected": -3.319507122039795,
777
+ "loss": 0.1158,
778
+ "nll_loss": 0.13795073330402374,
779
+ "rewards/accuracies": 0.987500011920929,
780
+ "rewards/chosen": -0.0060953982174396515,
781
+ "rewards/margins": 0.15987998247146606,
782
+ "rewards/rejected": -0.16597537696361542,
783
+ "step": 430
784
+ },
785
+ {
786
+ "epoch": 0.89,
787
+ "grad_norm": 4.8125,
788
+ "learning_rate": 2.3836564731139807e-06,
789
+ "log_odds_chosen": 7.590296268463135,
790
+ "log_odds_ratio": -0.021134180948138237,
791
+ "logits/chosen": -1.8886020183563232,
792
+ "logits/rejected": -1.886425256729126,
793
+ "logps/chosen": -0.09185922890901566,
794
+ "logps/rejected": -3.507038116455078,
795
+ "loss": 0.1033,
796
+ "nll_loss": 0.06822694092988968,
797
+ "rewards/accuracies": 0.987500011920929,
798
+ "rewards/chosen": -0.00459296116605401,
799
+ "rewards/margins": 0.1707589477300644,
800
+ "rewards/rejected": -0.17535191774368286,
801
+ "step": 440
802
+ },
803
+ {
804
+ "epoch": 0.91,
805
+ "grad_norm": 6.28125,
806
+ "learning_rate": 2.357022603955159e-06,
807
+ "log_odds_chosen": 6.999331474304199,
808
+ "log_odds_ratio": -0.03660160303115845,
809
+ "logits/chosen": -1.861541748046875,
810
+ "logits/rejected": -1.861534833908081,
811
+ "logps/chosen": -0.12671752274036407,
812
+ "logps/rejected": -3.262852430343628,
813
+ "loss": 0.1048,
814
+ "nll_loss": 0.09448351711034775,
815
+ "rewards/accuracies": 0.9750000238418579,
816
+ "rewards/chosen": -0.0063358754850924015,
817
+ "rewards/margins": 0.15680675208568573,
818
+ "rewards/rejected": -0.1631426364183426,
819
+ "step": 450
820
+ },
821
+ {
822
+ "epoch": 0.93,
823
+ "grad_norm": 4.78125,
824
+ "learning_rate": 2.3312620206007847e-06,
825
+ "log_odds_chosen": 6.333981037139893,
826
+ "log_odds_ratio": -0.049268852919340134,
827
+ "logits/chosen": -1.8751580715179443,
828
+ "logits/rejected": -1.8740007877349854,
829
+ "logps/chosen": -0.13210389018058777,
830
+ "logps/rejected": -2.7039647102355957,
831
+ "loss": 0.1094,
832
+ "nll_loss": 0.09672755748033524,
833
+ "rewards/accuracies": 1.0,
834
+ "rewards/chosen": -0.006605193950235844,
835
+ "rewards/margins": 0.12859304249286652,
836
+ "rewards/rejected": -0.13519823551177979,
837
+ "step": 460
838
+ },
839
+ {
840
+ "epoch": 0.96,
841
+ "grad_norm": 6.65625,
842
+ "learning_rate": 2.3063280200722128e-06,
843
+ "log_odds_chosen": 7.438237190246582,
844
+ "log_odds_ratio": -0.02831345424056053,
845
+ "logits/chosen": -1.9090907573699951,
846
+ "logits/rejected": -1.9076954126358032,
847
+ "logps/chosen": -0.07157810777425766,
848
+ "logps/rejected": -3.022064447402954,
849
+ "loss": 0.0996,
850
+ "nll_loss": 0.09075023233890533,
851
+ "rewards/accuracies": 1.0,
852
+ "rewards/chosen": -0.003578905714675784,
853
+ "rewards/margins": 0.14752431213855743,
854
+ "rewards/rejected": -0.151103213429451,
855
+ "step": 470
856
+ },
857
+ {
858
+ "epoch": 0.98,
859
+ "grad_norm": 5.46875,
860
+ "learning_rate": 2.2821773229381924e-06,
861
+ "log_odds_chosen": 8.533849716186523,
862
+ "log_odds_ratio": -0.010285028256475925,
863
+ "logits/chosen": -1.958752989768982,
864
+ "logits/rejected": -1.961133599281311,
865
+ "logps/chosen": -0.07796324789524078,
866
+ "logps/rejected": -3.159803628921509,
867
+ "loss": 0.1074,
868
+ "nll_loss": 0.10170219838619232,
869
+ "rewards/accuracies": 1.0,
870
+ "rewards/chosen": -0.0038981616962701082,
871
+ "rewards/margins": 0.15409204363822937,
872
+ "rewards/rejected": -0.15799018740653992,
873
+ "step": 480
874
+ },
875
+ {
876
+ "epoch": 1.0,
877
+ "grad_norm": 4.40625,
878
+ "learning_rate": 2.2587697572631284e-06,
879
+ "log_odds_chosen": 7.246984958648682,
880
+ "log_odds_ratio": -0.03945430740714073,
881
+ "logits/chosen": -1.9564424753189087,
882
+ "logits/rejected": -1.9600261449813843,
883
+ "logps/chosen": -0.1045081838965416,
884
+ "logps/rejected": -3.074450731277466,
885
+ "loss": 0.1118,
886
+ "nll_loss": 0.11936734616756439,
887
+ "rewards/accuracies": 1.0,
888
+ "rewards/chosen": -0.00522540882229805,
889
+ "rewards/margins": 0.14849711954593658,
890
+ "rewards/rejected": -0.15372253954410553,
891
+ "step": 490
892
+ },
893
+ {
894
+ "epoch": 1.02,
895
+ "grad_norm": 4.78125,
896
+ "learning_rate": 2.23606797749979e-06,
897
+ "log_odds_chosen": 7.313413143157959,
898
+ "log_odds_ratio": -0.06336325407028198,
899
+ "logits/chosen": -1.925005316734314,
900
+ "logits/rejected": -1.9246110916137695,
901
+ "logps/chosen": -0.06910906732082367,
902
+ "logps/rejected": -3.088639974594116,
903
+ "loss": 0.0748,
904
+ "nll_loss": 0.06068875640630722,
905
+ "rewards/accuracies": 0.987500011920929,
906
+ "rewards/chosen": -0.0034554533194750547,
907
+ "rewards/margins": 0.15097656846046448,
908
+ "rewards/rejected": -0.1544320285320282,
909
+ "step": 500
910
+ },
911
+ {
912
+ "epoch": 1.04,
913
+ "grad_norm": 3.984375,
914
+ "learning_rate": 2.2140372138502386e-06,
915
+ "log_odds_chosen": 8.353876113891602,
916
+ "log_odds_ratio": -0.011013227514922619,
917
+ "logits/chosen": -1.9190175533294678,
918
+ "logits/rejected": -1.9186252355575562,
919
+ "logps/chosen": -0.060610998421907425,
920
+ "logps/rejected": -3.5158791542053223,
921
+ "loss": 0.073,
922
+ "nll_loss": 0.0933486819267273,
923
+ "rewards/accuracies": 1.0,
924
+ "rewards/chosen": -0.0030305504333227873,
925
+ "rewards/margins": 0.1727634072303772,
926
+ "rewards/rejected": -0.17579396069049835,
927
+ "step": 510
928
+ },
929
+ {
930
+ "epoch": 1.06,
931
+ "grad_norm": 4.6875,
932
+ "learning_rate": 2.1926450482675734e-06,
933
+ "log_odds_chosen": 8.398624420166016,
934
+ "log_odds_ratio": -0.01034296303987503,
935
+ "logits/chosen": -1.9005162715911865,
936
+ "logits/rejected": -1.8965606689453125,
937
+ "logps/chosen": -0.08098746836185455,
938
+ "logps/rejected": -3.6455941200256348,
939
+ "loss": 0.0753,
940
+ "nll_loss": 0.0635688304901123,
941
+ "rewards/accuracies": 1.0,
942
+ "rewards/chosen": -0.00404937332496047,
943
+ "rewards/margins": 0.17823031544685364,
944
+ "rewards/rejected": -0.18227970600128174,
945
+ "step": 520
946
+ },
947
+ {
948
+ "epoch": 1.08,
949
+ "grad_norm": 5.4375,
950
+ "learning_rate": 2.1718612138153473e-06,
951
+ "log_odds_chosen": 8.60987663269043,
952
+ "log_odds_ratio": -0.00940676499158144,
953
+ "logits/chosen": -1.8967382907867432,
954
+ "logits/rejected": -1.8937251567840576,
955
+ "logps/chosen": -0.07309839874505997,
956
+ "logps/rejected": -3.786154270172119,
957
+ "loss": 0.0827,
958
+ "nll_loss": 0.07813505828380585,
959
+ "rewards/accuracies": 1.0,
960
+ "rewards/chosen": -0.003654920030385256,
961
+ "rewards/margins": 0.18565279245376587,
962
+ "rewards/rejected": -0.18930771946907043,
963
+ "step": 530
964
+ },
965
+ {
966
+ "epoch": 1.1,
967
+ "grad_norm": 4.34375,
968
+ "learning_rate": 2.151657414559676e-06,
969
+ "log_odds_chosen": 9.486946105957031,
970
+ "log_odds_ratio": -0.006830220576375723,
971
+ "logits/chosen": -1.8446986675262451,
972
+ "logits/rejected": -1.839769721031189,
973
+ "logps/chosen": -0.045620597898960114,
974
+ "logps/rejected": -4.278280258178711,
975
+ "loss": 0.0759,
976
+ "nll_loss": 0.061400435864925385,
977
+ "rewards/accuracies": 1.0,
978
+ "rewards/chosen": -0.002281030174344778,
979
+ "rewards/margins": 0.21163301169872284,
980
+ "rewards/rejected": -0.21391403675079346,
981
+ "step": 540
982
+ },
983
+ {
984
+ "epoch": 1.12,
985
+ "grad_norm": 3.328125,
986
+ "learning_rate": 2.132007163556104e-06,
987
+ "log_odds_chosen": 8.17650032043457,
988
+ "log_odds_ratio": -0.015669699758291245,
989
+ "logits/chosen": -1.83901047706604,
990
+ "logits/rejected": -1.835837960243225,
991
+ "logps/chosen": -0.08997470885515213,
992
+ "logps/rejected": -3.697235584259033,
993
+ "loss": 0.0766,
994
+ "nll_loss": 0.05432655289769173,
995
+ "rewards/accuracies": 1.0,
996
+ "rewards/chosen": -0.004498735070228577,
997
+ "rewards/margins": 0.1803630292415619,
998
+ "rewards/rejected": -0.18486176431179047,
999
+ "step": 550
1000
+ },
1001
+ {
1002
+ "epoch": 1.14,
1003
+ "grad_norm": 4.75,
1004
+ "learning_rate": 2.1128856368212917e-06,
1005
+ "log_odds_chosen": 7.758215427398682,
1006
+ "log_odds_ratio": -0.027782147750258446,
1007
+ "logits/chosen": -1.887261152267456,
1008
+ "logits/rejected": -1.8827438354492188,
1009
+ "logps/chosen": -0.14876721799373627,
1010
+ "logps/rejected": -3.6743240356445312,
1011
+ "loss": 0.0828,
1012
+ "nll_loss": 0.11095689237117767,
1013
+ "rewards/accuracies": 1.0,
1014
+ "rewards/chosen": -0.007438362576067448,
1015
+ "rewards/margins": 0.17627784609794617,
1016
+ "rewards/rejected": -0.18371620774269104,
1017
+ "step": 560
1018
+ },
1019
+ {
1020
+ "epoch": 1.16,
1021
+ "grad_norm": 4.5625,
1022
+ "learning_rate": 2.0942695414584777e-06,
1023
+ "log_odds_chosen": 8.34414291381836,
1024
+ "log_odds_ratio": -0.021350722759962082,
1025
+ "logits/chosen": -1.8923943042755127,
1026
+ "logits/rejected": -1.8904056549072266,
1027
+ "logps/chosen": -0.09869461506605148,
1028
+ "logps/rejected": -3.8127264976501465,
1029
+ "loss": 0.0633,
1030
+ "nll_loss": 0.06047019362449646,
1031
+ "rewards/accuracies": 0.987500011920929,
1032
+ "rewards/chosen": -0.004934730939567089,
1033
+ "rewards/margins": 0.18570157885551453,
1034
+ "rewards/rejected": -0.19063633680343628,
1035
+ "step": 570
1036
+ },
1037
+ {
1038
+ "epoch": 1.18,
1039
+ "grad_norm": 4.09375,
1040
+ "learning_rate": 2.0761369963434992e-06,
1041
+ "log_odds_chosen": 8.360025405883789,
1042
+ "log_odds_ratio": -0.022461308166384697,
1043
+ "logits/chosen": -1.9567257165908813,
1044
+ "logits/rejected": -1.9553306102752686,
1045
+ "logps/chosen": -0.057451434433460236,
1046
+ "logps/rejected": -3.338317394256592,
1047
+ "loss": 0.0647,
1048
+ "nll_loss": 0.09119249880313873,
1049
+ "rewards/accuracies": 0.987500011920929,
1050
+ "rewards/chosen": -0.0028725718148052692,
1051
+ "rewards/margins": 0.1640433371067047,
1052
+ "rewards/rejected": -0.1669158935546875,
1053
+ "step": 580
1054
+ },
1055
+ {
1056
+ "epoch": 1.2,
1057
+ "grad_norm": 4.53125,
1058
+ "learning_rate": 2.058467423981546e-06,
1059
+ "log_odds_chosen": 8.788118362426758,
1060
+ "log_odds_ratio": -0.014825952239334583,
1061
+ "logits/chosen": -2.0043745040893555,
1062
+ "logits/rejected": -2.00369930267334,
1063
+ "logps/chosen": -0.06359820067882538,
1064
+ "logps/rejected": -3.9866626262664795,
1065
+ "loss": 0.0737,
1066
+ "nll_loss": 0.10766579210758209,
1067
+ "rewards/accuracies": 1.0,
1068
+ "rewards/chosen": -0.00317991035990417,
1069
+ "rewards/margins": 0.19615323841571808,
1070
+ "rewards/rejected": -0.19933313131332397,
1071
+ "step": 590
1072
+ },
1073
+ {
1074
+ "epoch": 1.22,
1075
+ "grad_norm": 4.53125,
1076
+ "learning_rate": 2.0412414523193154e-06,
1077
+ "log_odds_chosen": 8.787096977233887,
1078
+ "log_odds_ratio": -0.02094932086765766,
1079
+ "logits/chosen": -2.0005838871002197,
1080
+ "logits/rejected": -1.9980299472808838,
1081
+ "logps/chosen": -0.06464333832263947,
1082
+ "logps/rejected": -3.8292534351348877,
1083
+ "loss": 0.0665,
1084
+ "nll_loss": 0.07510735094547272,
1085
+ "rewards/accuracies": 1.0,
1086
+ "rewards/chosen": -0.0032321668695658445,
1087
+ "rewards/margins": 0.1882304847240448,
1088
+ "rewards/rejected": -0.1914626806974411,
1089
+ "step": 600
1090
+ },
1091
+ {
1092
+ "epoch": 1.24,
1093
+ "grad_norm": 5.0,
1094
+ "learning_rate": 2.0244408254472904e-06,
1095
+ "log_odds_chosen": 8.794069290161133,
1096
+ "log_odds_ratio": -0.0068406336940824986,
1097
+ "logits/chosen": -1.9200255870819092,
1098
+ "logits/rejected": -1.9168468713760376,
1099
+ "logps/chosen": -0.07637156546115875,
1100
+ "logps/rejected": -4.043463230133057,
1101
+ "loss": 0.0774,
1102
+ "nll_loss": 0.06419046223163605,
1103
+ "rewards/accuracies": 1.0,
1104
+ "rewards/chosen": -0.00381857855245471,
1105
+ "rewards/margins": 0.1983545869588852,
1106
+ "rewards/rejected": -0.2021731585264206,
1107
+ "step": 610
1108
+ },
1109
+ {
1110
+ "epoch": 1.26,
1111
+ "grad_norm": 3.359375,
1112
+ "learning_rate": 2.0080483222562476e-06,
1113
+ "log_odds_chosen": 9.725520133972168,
1114
+ "log_odds_ratio": -0.010235416702926159,
1115
+ "logits/chosen": -1.8615529537200928,
1116
+ "logits/rejected": -1.8584880828857422,
1117
+ "logps/chosen": -0.07194074988365173,
1118
+ "logps/rejected": -4.24954891204834,
1119
+ "loss": 0.0592,
1120
+ "nll_loss": 0.045302603393793106,
1121
+ "rewards/accuracies": 1.0,
1122
+ "rewards/chosen": -0.0035970378667116165,
1123
+ "rewards/margins": 0.20888039469718933,
1124
+ "rewards/rejected": -0.212477445602417,
1125
+ "step": 620
1126
+ },
1127
+ {
1128
+ "epoch": 1.28,
1129
+ "grad_norm": 3.703125,
1130
+ "learning_rate": 1.9920476822239895e-06,
1131
+ "log_odds_chosen": 8.134096145629883,
1132
+ "log_odds_ratio": -0.02648916281759739,
1133
+ "logits/chosen": -1.987565279006958,
1134
+ "logits/rejected": -1.985645055770874,
1135
+ "logps/chosen": -0.07914531975984573,
1136
+ "logps/rejected": -3.5829226970672607,
1137
+ "loss": 0.076,
1138
+ "nll_loss": 0.06963127106428146,
1139
+ "rewards/accuracies": 0.987500011920929,
1140
+ "rewards/chosen": -0.003957265987992287,
1141
+ "rewards/margins": 0.17518886923789978,
1142
+ "rewards/rejected": -0.17914614081382751,
1143
+ "step": 630
1144
+ },
1145
+ {
1146
+ "epoch": 1.3,
1147
+ "grad_norm": 4.65625,
1148
+ "learning_rate": 1.976423537605237e-06,
1149
+ "log_odds_chosen": 8.905769348144531,
1150
+ "log_odds_ratio": -0.0063996403478085995,
1151
+ "logits/chosen": -1.8973668813705444,
1152
+ "logits/rejected": -1.895581841468811,
1153
+ "logps/chosen": -0.04883747547864914,
1154
+ "logps/rejected": -3.7183055877685547,
1155
+ "loss": 0.0691,
1156
+ "nll_loss": 0.060373760759830475,
1157
+ "rewards/accuracies": 1.0,
1158
+ "rewards/chosen": -0.0024418740067631006,
1159
+ "rewards/margins": 0.1834733784198761,
1160
+ "rewards/rejected": -0.1859152615070343,
1161
+ "step": 640
1162
+ },
1163
+ {
1164
+ "epoch": 1.32,
1165
+ "grad_norm": 3.671875,
1166
+ "learning_rate": 1.961161351381841e-06,
1167
+ "log_odds_chosen": 8.213096618652344,
1168
+ "log_odds_ratio": -0.012629789300262928,
1169
+ "logits/chosen": -1.8932087421417236,
1170
+ "logits/rejected": -1.8902513980865479,
1171
+ "logps/chosen": -0.07331077754497528,
1172
+ "logps/rejected": -3.4578137397766113,
1173
+ "loss": 0.0703,
1174
+ "nll_loss": 0.060921620577573776,
1175
+ "rewards/accuracies": 1.0,
1176
+ "rewards/chosen": -0.0036655396688729525,
1177
+ "rewards/margins": 0.16922515630722046,
1178
+ "rewards/rejected": -0.17289067804813385,
1179
+ "step": 650
1180
+ },
1181
+ {
1182
+ "epoch": 1.34,
1183
+ "grad_norm": 4.1875,
1184
+ "learning_rate": 1.9462473604038077e-06,
1185
+ "log_odds_chosen": 8.2732572555542,
1186
+ "log_odds_ratio": -0.015530670061707497,
1187
+ "logits/chosen": -1.9228805303573608,
1188
+ "logits/rejected": -1.9188134670257568,
1189
+ "logps/chosen": -0.08276908099651337,
1190
+ "logps/rejected": -3.572401523590088,
1191
+ "loss": 0.0672,
1192
+ "nll_loss": 0.03623514249920845,
1193
+ "rewards/accuracies": 1.0,
1194
+ "rewards/chosen": -0.004138454794883728,
1195
+ "rewards/margins": 0.1744816154241562,
1196
+ "rewards/rejected": -0.17862007021903992,
1197
+ "step": 660
1198
+ },
1199
+ {
1200
+ "epoch": 1.36,
1201
+ "grad_norm": 2.953125,
1202
+ "learning_rate": 1.9316685232156397e-06,
1203
+ "log_odds_chosen": 8.805624961853027,
1204
+ "log_odds_ratio": -0.0074573298916220665,
1205
+ "logits/chosen": -1.8942184448242188,
1206
+ "logits/rejected": -1.8891878128051758,
1207
+ "logps/chosen": -0.07351969927549362,
1208
+ "logps/rejected": -3.884014844894409,
1209
+ "loss": 0.0656,
1210
+ "nll_loss": 0.05570251867175102,
1211
+ "rewards/accuracies": 1.0,
1212
+ "rewards/chosen": -0.0036759853828698397,
1213
+ "rewards/margins": 0.19052477180957794,
1214
+ "rewards/rejected": -0.1942007690668106,
1215
+ "step": 670
1216
+ },
1217
+ {
1218
+ "epoch": 1.38,
1219
+ "grad_norm": 3.328125,
1220
+ "learning_rate": 1.917412472118426e-06,
1221
+ "log_odds_chosen": 8.748679161071777,
1222
+ "log_odds_ratio": -0.024869108572602272,
1223
+ "logits/chosen": -1.9717075824737549,
1224
+ "logits/rejected": -1.9675525426864624,
1225
+ "logps/chosen": -0.06451079994440079,
1226
+ "logps/rejected": -3.9762165546417236,
1227
+ "loss": 0.073,
1228
+ "nll_loss": 0.11092261970043182,
1229
+ "rewards/accuracies": 0.987500011920929,
1230
+ "rewards/chosen": -0.003225540043786168,
1231
+ "rewards/margins": 0.19558526575565338,
1232
+ "rewards/rejected": -0.19881080090999603,
1233
+ "step": 680
1234
+ },
1235
+ {
1236
+ "epoch": 1.4,
1237
+ "grad_norm": 4.90625,
1238
+ "learning_rate": 1.9034674690672024e-06,
1239
+ "log_odds_chosen": 8.910208702087402,
1240
+ "log_odds_ratio": -0.005920908413827419,
1241
+ "logits/chosen": -1.8858320713043213,
1242
+ "logits/rejected": -1.8857864141464233,
1243
+ "logps/chosen": -0.06332524120807648,
1244
+ "logps/rejected": -3.758251667022705,
1245
+ "loss": 0.0613,
1246
+ "nll_loss": 0.07212791591882706,
1247
+ "rewards/accuracies": 1.0,
1248
+ "rewards/chosen": -0.003166262526065111,
1249
+ "rewards/margins": 0.18474632501602173,
1250
+ "rewards/rejected": -0.18791258335113525,
1251
+ "step": 690
1252
+ },
1253
+ {
1254
+ "epoch": 1.42,
1255
+ "grad_norm": 3.25,
1256
+ "learning_rate": 1.8898223650461362e-06,
1257
+ "log_odds_chosen": 8.65866756439209,
1258
+ "log_odds_ratio": -0.012265409342944622,
1259
+ "logits/chosen": -1.9048351049423218,
1260
+ "logits/rejected": -1.9031221866607666,
1261
+ "logps/chosen": -0.0720919519662857,
1262
+ "logps/rejected": -3.809000015258789,
1263
+ "loss": 0.0596,
1264
+ "nll_loss": 0.054995644837617874,
1265
+ "rewards/accuracies": 1.0,
1266
+ "rewards/chosen": -0.003604597644880414,
1267
+ "rewards/margins": 0.18684543669223785,
1268
+ "rewards/rejected": -0.1904500126838684,
1269
+ "step": 700
1270
+ },
1271
+ {
1272
+ "epoch": 1.44,
1273
+ "grad_norm": 3.34375,
1274
+ "learning_rate": 1.876466562602004e-06,
1275
+ "log_odds_chosen": 8.907628059387207,
1276
+ "log_odds_ratio": -0.01111848745495081,
1277
+ "logits/chosen": -1.945853590965271,
1278
+ "logits/rejected": -1.942375898361206,
1279
+ "logps/chosen": -0.07798963785171509,
1280
+ "logps/rejected": -4.1663408279418945,
1281
+ "loss": 0.0662,
1282
+ "nll_loss": 0.08023327589035034,
1283
+ "rewards/accuracies": 1.0,
1284
+ "rewards/chosen": -0.003899481613188982,
1285
+ "rewards/margins": 0.20441755652427673,
1286
+ "rewards/rejected": -0.20831704139709473,
1287
+ "step": 710
1288
+ },
1289
+ {
1290
+ "epoch": 1.46,
1291
+ "grad_norm": 5.28125,
1292
+ "learning_rate": 1.863389981249825e-06,
1293
+ "log_odds_chosen": 9.416102409362793,
1294
+ "log_odds_ratio": -0.007732940372079611,
1295
+ "logits/chosen": -1.8631290197372437,
1296
+ "logits/rejected": -1.8593025207519531,
1297
+ "logps/chosen": -0.06107889488339424,
1298
+ "logps/rejected": -4.503323554992676,
1299
+ "loss": 0.0585,
1300
+ "nll_loss": 0.05343944951891899,
1301
+ "rewards/accuracies": 1.0,
1302
+ "rewards/chosen": -0.0030539450235664845,
1303
+ "rewards/margins": 0.22211222350597382,
1304
+ "rewards/rejected": -0.2251661717891693,
1305
+ "step": 720
1306
+ },
1307
+ {
1308
+ "epoch": 1.48,
1309
+ "grad_norm": 2.78125,
1310
+ "learning_rate": 1.8505830254940132e-06,
1311
+ "log_odds_chosen": 9.134099960327148,
1312
+ "log_odds_ratio": -0.011219611391425133,
1313
+ "logits/chosen": -1.9615509510040283,
1314
+ "logits/rejected": -1.9607875347137451,
1315
+ "logps/chosen": -0.05891376733779907,
1316
+ "logps/rejected": -4.027497291564941,
1317
+ "loss": 0.0582,
1318
+ "nll_loss": 0.050385020673274994,
1319
+ "rewards/accuracies": 1.0,
1320
+ "rewards/chosen": -0.0029456880874931812,
1321
+ "rewards/margins": 0.1984291821718216,
1322
+ "rewards/rejected": -0.2013748586177826,
1323
+ "step": 730
1324
+ },
1325
+ {
1326
+ "epoch": 1.5,
1327
+ "grad_norm": 4.0625,
1328
+ "learning_rate": 1.8380365552345197e-06,
1329
+ "log_odds_chosen": 8.314208984375,
1330
+ "log_odds_ratio": -0.017240628600120544,
1331
+ "logits/chosen": -1.9306405782699585,
1332
+ "logits/rejected": -1.9276716709136963,
1333
+ "logps/chosen": -0.12811905145645142,
1334
+ "logps/rejected": -3.857971668243408,
1335
+ "loss": 0.0666,
1336
+ "nll_loss": 0.07230392098426819,
1337
+ "rewards/accuracies": 1.0,
1338
+ "rewards/chosen": -0.006405952386558056,
1339
+ "rewards/margins": 0.18649262189865112,
1340
+ "rewards/rejected": -0.19289858639240265,
1341
+ "step": 740
1342
+ },
1343
+ {
1344
+ "epoch": 1.52,
1345
+ "grad_norm": 6.8125,
1346
+ "learning_rate": 1.8257418583505536e-06,
1347
+ "log_odds_chosen": 8.118074417114258,
1348
+ "log_odds_ratio": -0.04242696613073349,
1349
+ "logits/chosen": -1.951964020729065,
1350
+ "logits/rejected": -1.9483181238174438,
1351
+ "logps/chosen": -0.07924682646989822,
1352
+ "logps/rejected": -3.6926894187927246,
1353
+ "loss": 0.0608,
1354
+ "nll_loss": 0.05164622515439987,
1355
+ "rewards/accuracies": 0.987500011920929,
1356
+ "rewards/chosen": -0.003962341230362654,
1357
+ "rewards/margins": 0.18067213892936707,
1358
+ "rewards/rejected": -0.1846344769001007,
1359
+ "step": 750
1360
+ },
1361
+ {
1362
+ "epoch": 1.54,
1363
+ "grad_norm": 2.890625,
1364
+ "learning_rate": 1.8136906252750293e-06,
1365
+ "log_odds_chosen": 8.884002685546875,
1366
+ "log_odds_ratio": -0.007163527421653271,
1367
+ "logits/chosen": -1.8522392511367798,
1368
+ "logits/rejected": -1.8525257110595703,
1369
+ "logps/chosen": -0.06406822055578232,
1370
+ "logps/rejected": -3.8873114585876465,
1371
+ "loss": 0.0508,
1372
+ "nll_loss": 0.052127063274383545,
1373
+ "rewards/accuracies": 1.0,
1374
+ "rewards/chosen": -0.003203411353752017,
1375
+ "rewards/margins": 0.19116216897964478,
1376
+ "rewards/rejected": -0.19436559081077576,
1377
+ "step": 760
1378
+ },
1379
+ {
1380
+ "epoch": 1.57,
1381
+ "grad_norm": 4.8125,
1382
+ "learning_rate": 1.801874925391118e-06,
1383
+ "log_odds_chosen": 8.48697566986084,
1384
+ "log_odds_ratio": -0.026209736242890358,
1385
+ "logits/chosen": -1.9563274383544922,
1386
+ "logits/rejected": -1.9566619396209717,
1387
+ "logps/chosen": -0.075563944876194,
1388
+ "logps/rejected": -3.6374893188476562,
1389
+ "loss": 0.0665,
1390
+ "nll_loss": 0.05961776524782181,
1391
+ "rewards/accuracies": 0.987500011920929,
1392
+ "rewards/chosen": -0.0037781973369419575,
1393
+ "rewards/margins": 0.17809626460075378,
1394
+ "rewards/rejected": -0.18187445402145386,
1395
+ "step": 770
1396
+ },
1397
+ {
1398
+ "epoch": 1.59,
1399
+ "grad_norm": 3.703125,
1400
+ "learning_rate": 1.7902871850985824e-06,
1401
+ "log_odds_chosen": 9.119174003601074,
1402
+ "log_odds_ratio": -0.024956868961453438,
1403
+ "logits/chosen": -1.918013572692871,
1404
+ "logits/rejected": -1.9134963750839233,
1405
+ "logps/chosen": -0.05534970015287399,
1406
+ "logps/rejected": -4.010322570800781,
1407
+ "loss": 0.0595,
1408
+ "nll_loss": 0.04103861376643181,
1409
+ "rewards/accuracies": 0.987500011920929,
1410
+ "rewards/chosen": -0.0027674853336066008,
1411
+ "rewards/margins": 0.19774861633777618,
1412
+ "rewards/rejected": -0.20051613450050354,
1413
+ "step": 780
1414
+ },
1415
+ {
1416
+ "epoch": 1.61,
1417
+ "grad_norm": 3.6875,
1418
+ "learning_rate": 1.7789201674120502e-06,
1419
+ "log_odds_chosen": 8.733691215515137,
1420
+ "log_odds_ratio": -0.008985884487628937,
1421
+ "logits/chosen": -1.8730590343475342,
1422
+ "logits/rejected": -1.8695628643035889,
1423
+ "logps/chosen": -0.05646519735455513,
1424
+ "logps/rejected": -3.503816604614258,
1425
+ "loss": 0.0506,
1426
+ "nll_loss": 0.047934651374816895,
1427
+ "rewards/accuracies": 1.0,
1428
+ "rewards/chosen": -0.002823260147124529,
1429
+ "rewards/margins": 0.17236760258674622,
1430
+ "rewards/rejected": -0.17519085109233856,
1431
+ "step": 790
1432
+ },
1433
+ {
1434
+ "epoch": 1.63,
1435
+ "grad_norm": 2.140625,
1436
+ "learning_rate": 1.7677669529663689e-06,
1437
+ "log_odds_chosen": 8.943314552307129,
1438
+ "log_odds_ratio": -0.00942598469555378,
1439
+ "logits/chosen": -1.8948551416397095,
1440
+ "logits/rejected": -1.892686128616333,
1441
+ "logps/chosen": -0.05206329748034477,
1442
+ "logps/rejected": -3.80363130569458,
1443
+ "loss": 0.0526,
1444
+ "nll_loss": 0.06757352501153946,
1445
+ "rewards/accuracies": 1.0,
1446
+ "rewards/chosen": -0.0026031648740172386,
1447
+ "rewards/margins": 0.18757839500904083,
1448
+ "rewards/rejected": -0.19018153846263885,
1449
+ "step": 800
1450
+ },
1451
+ {
1452
+ "epoch": 1.65,
1453
+ "grad_norm": 3.515625,
1454
+ "learning_rate": 1.7568209223157664e-06,
1455
+ "log_odds_chosen": 9.602978706359863,
1456
+ "log_odds_ratio": -0.0054091801866889,
1457
+ "logits/chosen": -1.8749698400497437,
1458
+ "logits/rejected": -1.8726646900177002,
1459
+ "logps/chosen": -0.049051497131586075,
1460
+ "logps/rejected": -4.167695045471191,
1461
+ "loss": 0.0527,
1462
+ "nll_loss": 0.08227180689573288,
1463
+ "rewards/accuracies": 1.0,
1464
+ "rewards/chosen": -0.0024525749031454325,
1465
+ "rewards/margins": 0.20593218505382538,
1466
+ "rewards/rejected": -0.20838478207588196,
1467
+ "step": 810
1468
+ },
1469
+ {
1470
+ "epoch": 1.67,
1471
+ "grad_norm": 4.1875,
1472
+ "learning_rate": 1.7460757394239458e-06,
1473
+ "log_odds_chosen": 8.731072425842285,
1474
+ "log_odds_ratio": -0.006553919520229101,
1475
+ "logits/chosen": -1.9533252716064453,
1476
+ "logits/rejected": -1.9529660940170288,
1477
+ "logps/chosen": -0.0727374479174614,
1478
+ "logps/rejected": -3.5265769958496094,
1479
+ "loss": 0.0659,
1480
+ "nll_loss": 0.0733790472149849,
1481
+ "rewards/accuracies": 1.0,
1482
+ "rewards/chosen": -0.003636872861534357,
1483
+ "rewards/margins": 0.17269198596477509,
1484
+ "rewards/rejected": -0.1763288527727127,
1485
+ "step": 820
1486
+ },
1487
+ {
1488
+ "epoch": 1.69,
1489
+ "grad_norm": 2.984375,
1490
+ "learning_rate": 1.7355253362515584e-06,
1491
+ "log_odds_chosen": 9.135687828063965,
1492
+ "log_odds_ratio": -0.03666002303361893,
1493
+ "logits/chosen": -1.9465566873550415,
1494
+ "logits/rejected": -1.9425933361053467,
1495
+ "logps/chosen": -0.07385378330945969,
1496
+ "logps/rejected": -3.989952564239502,
1497
+ "loss": 0.0478,
1498
+ "nll_loss": 0.06535151600837708,
1499
+ "rewards/accuracies": 0.987500011920929,
1500
+ "rewards/chosen": -0.003692689584568143,
1501
+ "rewards/margins": 0.1958049088716507,
1502
+ "rewards/rejected": -0.19949761033058167,
1503
+ "step": 830
1504
+ },
1505
+ {
1506
+ "epoch": 1.71,
1507
+ "grad_norm": 2.9375,
1508
+ "learning_rate": 1.7251638983558855e-06,
1509
+ "log_odds_chosen": 8.61412239074707,
1510
+ "log_odds_ratio": -0.006854639388620853,
1511
+ "logits/chosen": -1.8970463275909424,
1512
+ "logits/rejected": -1.8957879543304443,
1513
+ "logps/chosen": -0.09370087087154388,
1514
+ "logps/rejected": -3.841550827026367,
1515
+ "loss": 0.0587,
1516
+ "nll_loss": 0.046143922954797745,
1517
+ "rewards/accuracies": 1.0,
1518
+ "rewards/chosen": -0.004685043357312679,
1519
+ "rewards/margins": 0.18739250302314758,
1520
+ "rewards/rejected": -0.19207754731178284,
1521
+ "step": 840
1522
+ },
1523
+ {
1524
+ "epoch": 1.73,
1525
+ "grad_norm": 3.40625,
1526
+ "learning_rate": 1.7149858514250883e-06,
1527
+ "log_odds_chosen": 9.044164657592773,
1528
+ "log_odds_ratio": -0.023376554250717163,
1529
+ "logits/chosen": -1.9251768589019775,
1530
+ "logits/rejected": -1.9241018295288086,
1531
+ "logps/chosen": -0.05019516497850418,
1532
+ "logps/rejected": -3.740339994430542,
1533
+ "loss": 0.0537,
1534
+ "nll_loss": 0.04762560874223709,
1535
+ "rewards/accuracies": 0.987500011920929,
1536
+ "rewards/chosen": -0.0025097583420574665,
1537
+ "rewards/margins": 0.18450722098350525,
1538
+ "rewards/rejected": -0.18701700866222382,
1539
+ "step": 850
1540
+ },
1541
+ {
1542
+ "epoch": 1.75,
1543
+ "grad_norm": 2.703125,
1544
+ "learning_rate": 1.704985848676184e-06,
1545
+ "log_odds_chosen": 8.560470581054688,
1546
+ "log_odds_ratio": -0.015030615031719208,
1547
+ "logits/chosen": -1.893930435180664,
1548
+ "logits/rejected": -1.8902561664581299,
1549
+ "logps/chosen": -0.07954917848110199,
1550
+ "logps/rejected": -3.786511182785034,
1551
+ "loss": 0.0544,
1552
+ "nll_loss": 0.054652292281389236,
1553
+ "rewards/accuracies": 1.0,
1554
+ "rewards/chosen": -0.0039774589240550995,
1555
+ "rewards/margins": 0.18534812331199646,
1556
+ "rewards/rejected": -0.18932557106018066,
1557
+ "step": 860
1558
+ },
1559
+ {
1560
+ "epoch": 1.77,
1561
+ "grad_norm": 2.9375,
1562
+ "learning_rate": 1.6951587590520263e-06,
1563
+ "log_odds_chosen": 9.6688232421875,
1564
+ "log_odds_ratio": -0.010795338079333305,
1565
+ "logits/chosen": -1.9225839376449585,
1566
+ "logits/rejected": -1.9220905303955078,
1567
+ "logps/chosen": -0.048293761909008026,
1568
+ "logps/rejected": -4.323627948760986,
1569
+ "loss": 0.0488,
1570
+ "nll_loss": 0.043466441333293915,
1571
+ "rewards/accuracies": 0.987500011920929,
1572
+ "rewards/chosen": -0.002414688002318144,
1573
+ "rewards/margins": 0.21376672387123108,
1574
+ "rewards/rejected": -0.2161814272403717,
1575
+ "step": 870
1576
+ },
1577
+ {
1578
+ "epoch": 1.79,
1579
+ "grad_norm": 3.234375,
1580
+ "learning_rate": 1.6854996561581053e-06,
1581
+ "log_odds_chosen": 9.363157272338867,
1582
+ "log_odds_ratio": -0.026759719476103783,
1583
+ "logits/chosen": -1.9485028982162476,
1584
+ "logits/rejected": -1.9452836513519287,
1585
+ "logps/chosen": -0.042739786207675934,
1586
+ "logps/rejected": -3.5607593059539795,
1587
+ "loss": 0.0555,
1588
+ "nll_loss": 0.04253282770514488,
1589
+ "rewards/accuracies": 0.9750000238418579,
1590
+ "rewards/chosen": -0.002136989263817668,
1591
+ "rewards/margins": 0.17590096592903137,
1592
+ "rewards/rejected": -0.17803797125816345,
1593
+ "step": 880
1594
+ },
1595
+ {
1596
+ "epoch": 1.81,
1597
+ "grad_norm": 4.1875,
1598
+ "learning_rate": 1.6760038078849776e-06,
1599
+ "log_odds_chosen": 8.319189071655273,
1600
+ "log_odds_ratio": -0.023240404203534126,
1601
+ "logits/chosen": -1.851123571395874,
1602
+ "logits/rejected": -1.849542260169983,
1603
+ "logps/chosen": -0.06788856536149979,
1604
+ "logps/rejected": -3.406628131866455,
1605
+ "loss": 0.0603,
1606
+ "nll_loss": 0.04571394622325897,
1607
+ "rewards/accuracies": 0.987500011920929,
1608
+ "rewards/chosen": -0.003394428174942732,
1609
+ "rewards/margins": 0.1669369637966156,
1610
+ "rewards/rejected": -0.17033140361309052,
1611
+ "step": 890
1612
+ },
1613
+ {
1614
+ "epoch": 1.83,
1615
+ "grad_norm": 4.21875,
1616
+ "learning_rate": 1.6666666666666667e-06,
1617
+ "log_odds_chosen": 8.61945915222168,
1618
+ "log_odds_ratio": -0.017158176749944687,
1619
+ "logits/chosen": -1.9368407726287842,
1620
+ "logits/rejected": -1.9338048696517944,
1621
+ "logps/chosen": -0.05057498812675476,
1622
+ "logps/rejected": -3.4434471130371094,
1623
+ "loss": 0.0499,
1624
+ "nll_loss": 0.05269382521510124,
1625
+ "rewards/accuracies": 1.0,
1626
+ "rewards/chosen": -0.0025287498719990253,
1627
+ "rewards/margins": 0.1696436107158661,
1628
+ "rewards/rejected": -0.17217235267162323,
1629
+ "step": 900
1630
+ },
1631
+ {
1632
+ "epoch": 1.85,
1633
+ "grad_norm": 3.265625,
1634
+ "learning_rate": 1.6574838603294898e-06,
1635
+ "log_odds_chosen": 9.762075424194336,
1636
+ "log_odds_ratio": -0.0037377229891717434,
1637
+ "logits/chosen": -2.0028135776519775,
1638
+ "logits/rejected": -2.0014703273773193,
1639
+ "logps/chosen": -0.034680210053920746,
1640
+ "logps/rejected": -4.053231239318848,
1641
+ "loss": 0.0576,
1642
+ "nll_loss": 0.057746343314647675,
1643
+ "rewards/accuracies": 1.0,
1644
+ "rewards/chosen": -0.0017340105259791017,
1645
+ "rewards/margins": 0.20092757046222687,
1646
+ "rewards/rejected": -0.20266155898571014,
1647
+ "step": 910
1648
+ },
1649
+ {
1650
+ "epoch": 1.87,
1651
+ "grad_norm": 3.0,
1652
+ "learning_rate": 1.648451183489468e-06,
1653
+ "log_odds_chosen": 9.181684494018555,
1654
+ "log_odds_ratio": -0.007746308110654354,
1655
+ "logits/chosen": -1.935484528541565,
1656
+ "logits/rejected": -1.932363748550415,
1657
+ "logps/chosen": -0.08534862101078033,
1658
+ "logps/rejected": -4.044074535369873,
1659
+ "loss": 0.0526,
1660
+ "nll_loss": 0.06775641441345215,
1661
+ "rewards/accuracies": 1.0,
1662
+ "rewards/chosen": -0.004267430864274502,
1663
+ "rewards/margins": 0.1979363113641739,
1664
+ "rewards/rejected": -0.20220375061035156,
1665
+ "step": 920
1666
+ },
1667
+ {
1668
+ "epoch": 1.89,
1669
+ "grad_norm": 1.609375,
1670
+ "learning_rate": 1.6395645894598825e-06,
1671
+ "log_odds_chosen": 9.165306091308594,
1672
+ "log_odds_ratio": -0.01449726801365614,
1673
+ "logits/chosen": -1.9488122463226318,
1674
+ "logits/rejected": -1.9472631216049194,
1675
+ "logps/chosen": -0.07069545984268188,
1676
+ "logps/rejected": -3.8849570751190186,
1677
+ "loss": 0.057,
1678
+ "nll_loss": 0.052025146782398224,
1679
+ "rewards/accuracies": 1.0,
1680
+ "rewards/chosen": -0.0035347728990018368,
1681
+ "rewards/margins": 0.1907130777835846,
1682
+ "rewards/rejected": -0.19424787163734436,
1683
+ "step": 930
1684
+ },
1685
+ {
1686
+ "epoch": 1.91,
1687
+ "grad_norm": 3.90625,
1688
+ "learning_rate": 1.6308201826336057e-06,
1689
+ "log_odds_chosen": 9.020658493041992,
1690
+ "log_odds_ratio": -0.00849425233900547,
1691
+ "logits/chosen": -1.916486144065857,
1692
+ "logits/rejected": -1.9109046459197998,
1693
+ "logps/chosen": -0.06480909883975983,
1694
+ "logps/rejected": -4.059305667877197,
1695
+ "loss": 0.0504,
1696
+ "nll_loss": 0.058377087116241455,
1697
+ "rewards/accuracies": 1.0,
1698
+ "rewards/chosen": -0.0032404554076492786,
1699
+ "rewards/margins": 0.19972485303878784,
1700
+ "rewards/rejected": -0.20296530425548553,
1701
+ "step": 940
1702
+ },
1703
+ {
1704
+ "epoch": 1.93,
1705
+ "grad_norm": 3.75,
1706
+ "learning_rate": 1.6222142113076255e-06,
1707
+ "log_odds_chosen": 9.191986083984375,
1708
+ "log_odds_ratio": -0.007987757213413715,
1709
+ "logits/chosen": -1.8939701318740845,
1710
+ "logits/rejected": -1.892139196395874,
1711
+ "logps/chosen": -0.042362868785858154,
1712
+ "logps/rejected": -3.922192335128784,
1713
+ "loss": 0.0543,
1714
+ "nll_loss": 0.0392763689160347,
1715
+ "rewards/accuracies": 1.0,
1716
+ "rewards/chosen": -0.0021181434858590364,
1717
+ "rewards/margins": 0.19399148225784302,
1718
+ "rewards/rejected": -0.1961096078157425,
1719
+ "step": 950
1720
+ },
1721
+ {
1722
+ "epoch": 1.95,
1723
+ "grad_norm": 3.65625,
1724
+ "learning_rate": 1.6137430609197571e-06,
1725
+ "log_odds_chosen": 8.950517654418945,
1726
+ "log_odds_ratio": -0.01621365174651146,
1727
+ "logits/chosen": -1.9614036083221436,
1728
+ "logits/rejected": -1.9606664180755615,
1729
+ "logps/chosen": -0.07848796993494034,
1730
+ "logps/rejected": -4.038422584533691,
1731
+ "loss": 0.0554,
1732
+ "nll_loss": 0.0618586540222168,
1733
+ "rewards/accuracies": 1.0,
1734
+ "rewards/chosen": -0.003924398683011532,
1735
+ "rewards/margins": 0.19799675047397614,
1736
+ "rewards/rejected": -0.20192115008831024,
1737
+ "step": 960
1738
+ },
1739
+ {
1740
+ "epoch": 1.97,
1741
+ "grad_norm": 2.875,
1742
+ "learning_rate": 1.605403247669839e-06,
1743
+ "log_odds_chosen": 8.41962718963623,
1744
+ "log_odds_ratio": -0.020061850547790527,
1745
+ "logits/chosen": -1.908539056777954,
1746
+ "logits/rejected": -1.9044193029403687,
1747
+ "logps/chosen": -0.07218164205551147,
1748
+ "logps/rejected": -3.6821017265319824,
1749
+ "loss": 0.0528,
1750
+ "nll_loss": 0.03915850818157196,
1751
+ "rewards/accuracies": 0.987500011920929,
1752
+ "rewards/chosen": -0.003609082894399762,
1753
+ "rewards/margins": 0.1804960072040558,
1754
+ "rewards/rejected": -0.18410511314868927,
1755
+ "step": 970
1756
+ },
1757
+ {
1758
+ "epoch": 1.99,
1759
+ "grad_norm": 4.4375,
1760
+ "learning_rate": 1.59719141249985e-06,
1761
+ "log_odds_chosen": 8.960528373718262,
1762
+ "log_odds_ratio": -0.007182210683822632,
1763
+ "logits/chosen": -2.059950351715088,
1764
+ "logits/rejected": -2.058908224105835,
1765
+ "logps/chosen": -0.07827770709991455,
1766
+ "logps/rejected": -3.9805305004119873,
1767
+ "loss": 0.0589,
1768
+ "nll_loss": 0.08598147332668304,
1769
+ "rewards/accuracies": 1.0,
1770
+ "rewards/chosen": -0.003913884982466698,
1771
+ "rewards/margins": 0.19511264562606812,
1772
+ "rewards/rejected": -0.19902652502059937,
1773
+ "step": 980
1774
+ },
1775
+ {
1776
+ "epoch": 2.01,
1777
+ "grad_norm": 2.421875,
1778
+ "learning_rate": 1.5891043154093205e-06,
1779
+ "log_odds_chosen": 8.75261402130127,
1780
+ "log_odds_ratio": -0.008246051147580147,
1781
+ "logits/chosen": -1.9300010204315186,
1782
+ "logits/rejected": -1.9284627437591553,
1783
+ "logps/chosen": -0.059024132788181305,
1784
+ "logps/rejected": -3.7112643718719482,
1785
+ "loss": 0.0501,
1786
+ "nll_loss": 0.0379130020737648,
1787
+ "rewards/accuracies": 1.0,
1788
+ "rewards/chosen": -0.0029512066394090652,
1789
+ "rewards/margins": 0.18261203169822693,
1790
+ "rewards/rejected": -0.18556323647499084,
1791
+ "step": 990
1792
+ },
1793
+ {
1794
+ "epoch": 2.03,
1795
+ "grad_norm": 2.4375,
1796
+ "learning_rate": 1.5811388300841898e-06,
1797
+ "log_odds_chosen": 9.132776260375977,
1798
+ "log_odds_ratio": -0.010855279862880707,
1799
+ "logits/chosen": -1.9048831462860107,
1800
+ "logits/rejected": -1.9021356105804443,
1801
+ "logps/chosen": -0.05851290374994278,
1802
+ "logps/rejected": -3.9342761039733887,
1803
+ "loss": 0.0399,
1804
+ "nll_loss": 0.045368582010269165,
1805
+ "rewards/accuracies": 1.0,
1806
+ "rewards/chosen": -0.002925645327195525,
1807
+ "rewards/margins": 0.19378817081451416,
1808
+ "rewards/rejected": -0.19671380519866943,
1809
+ "step": 1000
1810
+ },
1811
+ {
1812
+ "epoch": 2.05,
1813
+ "grad_norm": 3.34375,
1814
+ "learning_rate": 1.5732919388188816e-06,
1815
+ "log_odds_chosen": 9.254007339477539,
1816
+ "log_odds_ratio": -0.004844398703426123,
1817
+ "logits/chosen": -1.906938910484314,
1818
+ "logits/rejected": -1.9027900695800781,
1819
+ "logps/chosen": -0.0639704093337059,
1820
+ "logps/rejected": -4.133778095245361,
1821
+ "loss": 0.0377,
1822
+ "nll_loss": 0.03271421417593956,
1823
+ "rewards/accuracies": 1.0,
1824
+ "rewards/chosen": -0.003198520513251424,
1825
+ "rewards/margins": 0.20349040627479553,
1826
+ "rewards/rejected": -0.20668891072273254,
1827
+ "step": 1010
1828
+ },
1829
+ {
1830
+ "epoch": 2.07,
1831
+ "grad_norm": 2.34375,
1832
+ "learning_rate": 1.565560727712874e-06,
1833
+ "log_odds_chosen": 9.79780387878418,
1834
+ "log_odds_ratio": -0.0035797320306301117,
1835
+ "logits/chosen": -1.934853196144104,
1836
+ "logits/rejected": -1.9338003396987915,
1837
+ "logps/chosen": -0.05114441365003586,
1838
+ "logps/rejected": -4.095462799072266,
1839
+ "loss": 0.0456,
1840
+ "nll_loss": 0.055598996579647064,
1841
+ "rewards/accuracies": 1.0,
1842
+ "rewards/chosen": -0.002557220868766308,
1843
+ "rewards/margins": 0.20221590995788574,
1844
+ "rewards/rejected": -0.20477311313152313,
1845
+ "step": 1020
1846
+ },
1847
+ {
1848
+ "epoch": 2.09,
1849
+ "grad_norm": 2.140625,
1850
+ "learning_rate": 1.5579423821243897e-06,
1851
+ "log_odds_chosen": 10.34232234954834,
1852
+ "log_odds_ratio": -0.002075557829812169,
1853
+ "logits/chosen": -1.908393144607544,
1854
+ "logits/rejected": -1.9075828790664673,
1855
+ "logps/chosen": -0.03475125879049301,
1856
+ "logps/rejected": -4.130476474761963,
1857
+ "loss": 0.0436,
1858
+ "nll_loss": 0.04652316868305206,
1859
+ "rewards/accuracies": 1.0,
1860
+ "rewards/chosen": -0.0017375629395246506,
1861
+ "rewards/margins": 0.2047862708568573,
1862
+ "rewards/rejected": -0.2065238207578659,
1863
+ "step": 1030
1864
+ },
1865
+ {
1866
+ "epoch": 2.11,
1867
+ "grad_norm": 3.796875,
1868
+ "learning_rate": 1.5504341823651056e-06,
1869
+ "log_odds_chosen": 9.534278869628906,
1870
+ "log_odds_ratio": -0.014330191537737846,
1871
+ "logits/chosen": -1.8937572240829468,
1872
+ "logits/rejected": -1.8925514221191406,
1873
+ "logps/chosen": -0.05281580239534378,
1874
+ "logps/rejected": -3.8376965522766113,
1875
+ "loss": 0.0506,
1876
+ "nll_loss": 0.04847393184900284,
1877
+ "rewards/accuracies": 1.0,
1878
+ "rewards/chosen": -0.0026407900732010603,
1879
+ "rewards/margins": 0.18924403190612793,
1880
+ "rewards/rejected": -0.1918848156929016,
1881
+ "step": 1040
1882
+ },
1883
+ {
1884
+ "epoch": 2.13,
1885
+ "grad_norm": 2.765625,
1886
+ "learning_rate": 1.5430334996209192e-06,
1887
+ "log_odds_chosen": 9.982309341430664,
1888
+ "log_odds_ratio": -0.0034835089463740587,
1889
+ "logits/chosen": -1.9073753356933594,
1890
+ "logits/rejected": -1.9033172130584717,
1891
+ "logps/chosen": -0.03734086453914642,
1892
+ "logps/rejected": -4.198154449462891,
1893
+ "loss": 0.0399,
1894
+ "nll_loss": 0.03343256562948227,
1895
+ "rewards/accuracies": 1.0,
1896
+ "rewards/chosen": -0.0018670432036742568,
1897
+ "rewards/margins": 0.20804063975811005,
1898
+ "rewards/rejected": -0.20990769565105438,
1899
+ "step": 1050
1900
+ },
1901
+ {
1902
+ "epoch": 2.15,
1903
+ "grad_norm": 2.015625,
1904
+ "learning_rate": 1.5357377920848783e-06,
1905
+ "log_odds_chosen": 9.932341575622559,
1906
+ "log_odds_ratio": -0.00664928276091814,
1907
+ "logits/chosen": -1.968000054359436,
1908
+ "logits/rejected": -1.9650490283966064,
1909
+ "logps/chosen": -0.053252361714839935,
1910
+ "logps/rejected": -4.3154191970825195,
1911
+ "loss": 0.0398,
1912
+ "nll_loss": 0.03708261623978615,
1913
+ "rewards/accuracies": 1.0,
1914
+ "rewards/chosen": -0.0026626181788742542,
1915
+ "rewards/margins": 0.2131083756685257,
1916
+ "rewards/rejected": -0.21577095985412598,
1917
+ "step": 1060
1918
+ },
1919
+ {
1920
+ "epoch": 2.17,
1921
+ "grad_norm": 3.203125,
1922
+ "learning_rate": 1.5285446012893579e-06,
1923
+ "log_odds_chosen": 10.412023544311523,
1924
+ "log_odds_ratio": -0.004382356535643339,
1925
+ "logits/chosen": -1.8630377054214478,
1926
+ "logits/rejected": -1.858724594116211,
1927
+ "logps/chosen": -0.03156786039471626,
1928
+ "logps/rejected": -4.467475414276123,
1929
+ "loss": 0.0435,
1930
+ "nll_loss": 0.043990008533000946,
1931
+ "rewards/accuracies": 1.0,
1932
+ "rewards/chosen": -0.0015783928101882339,
1933
+ "rewards/margins": 0.22179536521434784,
1934
+ "rewards/rejected": -0.22337380051612854,
1935
+ "step": 1070
1936
+ },
1937
+ {
1938
+ "epoch": 2.2,
1939
+ "grad_norm": 2.859375,
1940
+ "learning_rate": 1.5214515486254614e-06,
1941
+ "log_odds_chosen": 9.860450744628906,
1942
+ "log_odds_ratio": -0.00421870406717062,
1943
+ "logits/chosen": -1.9334264993667603,
1944
+ "logits/rejected": -1.9309759140014648,
1945
+ "logps/chosen": -0.04163174703717232,
1946
+ "logps/rejected": -4.278031349182129,
1947
+ "loss": 0.0399,
1948
+ "nll_loss": 0.03083736263215542,
1949
+ "rewards/accuracies": 1.0,
1950
+ "rewards/chosen": -0.0020815874449908733,
1951
+ "rewards/margins": 0.21182000637054443,
1952
+ "rewards/rejected": -0.2139015644788742,
1953
+ "step": 1080
1954
+ },
1955
+ {
1956
+ "epoch": 2.22,
1957
+ "grad_norm": 3.375,
1958
+ "learning_rate": 1.5144563320384566e-06,
1959
+ "log_odds_chosen": 9.61739444732666,
1960
+ "log_odds_ratio": -0.007051125168800354,
1961
+ "logits/chosen": -1.9654079675674438,
1962
+ "logits/rejected": -1.9627532958984375,
1963
+ "logps/chosen": -0.04155623912811279,
1964
+ "logps/rejected": -3.9517130851745605,
1965
+ "loss": 0.0419,
1966
+ "nll_loss": 0.0399235263466835,
1967
+ "rewards/accuracies": 1.0,
1968
+ "rewards/chosen": -0.002077811863273382,
1969
+ "rewards/margins": 0.19550786912441254,
1970
+ "rewards/rejected": -0.19758570194244385,
1971
+ "step": 1090
1972
+ },
1973
+ {
1974
+ "epoch": 2.24,
1975
+ "grad_norm": 2.5625,
1976
+ "learning_rate": 1.5075567228888182e-06,
1977
+ "log_odds_chosen": 9.472233772277832,
1978
+ "log_odds_ratio": -0.004241319373250008,
1979
+ "logits/chosen": -1.8757314682006836,
1980
+ "logits/rejected": -1.875422477722168,
1981
+ "logps/chosen": -0.04146653413772583,
1982
+ "logps/rejected": -3.9626433849334717,
1983
+ "loss": 0.0388,
1984
+ "nll_loss": 0.044396962970495224,
1985
+ "rewards/accuracies": 1.0,
1986
+ "rewards/chosen": -0.002073326613754034,
1987
+ "rewards/margins": 0.19605886936187744,
1988
+ "rewards/rejected": -0.19813218712806702,
1989
+ "step": 1100
1990
+ },
1991
+ {
1992
+ "epoch": 2.26,
1993
+ "grad_norm": 2.84375,
1994
+ "learning_rate": 1.5007505629691608e-06,
1995
+ "log_odds_chosen": 10.060752868652344,
1996
+ "log_odds_ratio": -0.005245304200798273,
1997
+ "logits/chosen": -1.884495735168457,
1998
+ "logits/rejected": -1.8812463283538818,
1999
+ "logps/chosen": -0.042206455022096634,
2000
+ "logps/rejected": -4.164919853210449,
2001
+ "loss": 0.0437,
2002
+ "nll_loss": 0.03563828393816948,
2003
+ "rewards/accuracies": 1.0,
2004
+ "rewards/chosen": -0.002110322704538703,
2005
+ "rewards/margins": 0.20613567531108856,
2006
+ "rewards/rejected": -0.20824599266052246,
2007
+ "step": 1110
2008
+ },
2009
+ {
2010
+ "epoch": 2.28,
2011
+ "grad_norm": 1.875,
2012
+ "learning_rate": 1.494035761667992e-06,
2013
+ "log_odds_chosen": 10.209965705871582,
2014
+ "log_odds_ratio": -0.00312315602786839,
2015
+ "logits/chosen": -1.921362280845642,
2016
+ "logits/rejected": -1.9180587530136108,
2017
+ "logps/chosen": -0.04360593855381012,
2018
+ "logps/rejected": -4.336488723754883,
2019
+ "loss": 0.0425,
2020
+ "nll_loss": 0.050070591270923615,
2021
+ "rewards/accuracies": 1.0,
2022
+ "rewards/chosen": -0.002180296927690506,
2023
+ "rewards/margins": 0.2146441489458084,
2024
+ "rewards/rejected": -0.21682441234588623,
2025
+ "step": 1120
2026
+ },
2027
+ {
2028
+ "epoch": 2.3,
2029
+ "grad_norm": 3.34375,
2030
+ "learning_rate": 1.487410293271824e-06,
2031
+ "log_odds_chosen": 9.571352005004883,
2032
+ "log_odds_ratio": -0.0124302227050066,
2033
+ "logits/chosen": -1.9245920181274414,
2034
+ "logits/rejected": -1.9231822490692139,
2035
+ "logps/chosen": -0.051782481372356415,
2036
+ "logps/rejected": -4.130411624908447,
2037
+ "loss": 0.042,
2038
+ "nll_loss": 0.04519005864858627,
2039
+ "rewards/accuracies": 0.987500011920929,
2040
+ "rewards/chosen": -0.002589124022051692,
2041
+ "rewards/margins": 0.20393142104148865,
2042
+ "rewards/rejected": -0.20652055740356445,
2043
+ "step": 1130
2044
+ },
2045
+ {
2046
+ "epoch": 2.32,
2047
+ "grad_norm": 2.0625,
2048
+ "learning_rate": 1.480872194397731e-06,
2049
+ "log_odds_chosen": 9.836039543151855,
2050
+ "log_odds_ratio": -0.004647588822990656,
2051
+ "logits/chosen": -1.8972570896148682,
2052
+ "logits/rejected": -1.8939294815063477,
2053
+ "logps/chosen": -0.0661562830209732,
2054
+ "logps/rejected": -4.237457752227783,
2055
+ "loss": 0.0435,
2056
+ "nll_loss": 0.04812124744057655,
2057
+ "rewards/accuracies": 1.0,
2058
+ "rewards/chosen": -0.0033078144770115614,
2059
+ "rewards/margins": 0.20856507122516632,
2060
+ "rewards/rejected": -0.211872860789299,
2061
+ "step": 1140
2062
+ },
2063
+ {
2064
+ "epoch": 2.34,
2065
+ "grad_norm": 3.546875,
2066
+ "learning_rate": 1.4744195615489715e-06,
2067
+ "log_odds_chosen": 10.264336585998535,
2068
+ "log_odds_ratio": -0.0017655363772064447,
2069
+ "logits/chosen": -1.9511454105377197,
2070
+ "logits/rejected": -1.9473825693130493,
2071
+ "logps/chosen": -0.04254579171538353,
2072
+ "logps/rejected": -4.465078353881836,
2073
+ "loss": 0.0408,
2074
+ "nll_loss": 0.03954671695828438,
2075
+ "rewards/accuracies": 1.0,
2076
+ "rewards/chosen": -0.0021272895392030478,
2077
+ "rewards/margins": 0.22112664580345154,
2078
+ "rewards/rejected": -0.22325392067432404,
2079
+ "step": 1150
2080
+ },
2081
+ {
2082
+ "epoch": 2.36,
2083
+ "grad_norm": 2.203125,
2084
+ "learning_rate": 1.4680505487867589e-06,
2085
+ "log_odds_chosen": 9.718263626098633,
2086
+ "log_odds_ratio": -0.002675818046554923,
2087
+ "logits/chosen": -1.9049984216690063,
2088
+ "logits/rejected": -1.9013592004776,
2089
+ "logps/chosen": -0.04851164296269417,
2090
+ "logps/rejected": -4.104137420654297,
2091
+ "loss": 0.0396,
2092
+ "nll_loss": 0.03321639448404312,
2093
+ "rewards/accuracies": 1.0,
2094
+ "rewards/chosen": -0.0024255819153040648,
2095
+ "rewards/margins": 0.2027812898159027,
2096
+ "rewards/rejected": -0.20520684123039246,
2097
+ "step": 1160
2098
+ },
2099
+ {
2100
+ "epoch": 2.38,
2101
+ "grad_norm": 1.8828125,
2102
+ "learning_rate": 1.4617633655117156e-06,
2103
+ "log_odds_chosen": 9.280858993530273,
2104
+ "log_odds_ratio": -0.008265355601906776,
2105
+ "logits/chosen": -1.964536428451538,
2106
+ "logits/rejected": -1.9617042541503906,
2107
+ "logps/chosen": -0.06000961735844612,
2108
+ "logps/rejected": -4.104333400726318,
2109
+ "loss": 0.0435,
2110
+ "nll_loss": 0.04580981656908989,
2111
+ "rewards/accuracies": 1.0,
2112
+ "rewards/chosen": -0.003000481054186821,
2113
+ "rewards/margins": 0.20221619307994843,
2114
+ "rewards/rejected": -0.2052166759967804,
2115
+ "step": 1170
2116
+ },
2117
+ {
2118
+ "epoch": 2.4,
2119
+ "grad_norm": 2.578125,
2120
+ "learning_rate": 1.4555562743489552e-06,
2121
+ "log_odds_chosen": 10.110807418823242,
2122
+ "log_odds_ratio": -0.003054619999602437,
2123
+ "logits/chosen": -1.925559401512146,
2124
+ "logits/rejected": -1.924526572227478,
2125
+ "logps/chosen": -0.03492622449994087,
2126
+ "logps/rejected": -4.048843860626221,
2127
+ "loss": 0.0415,
2128
+ "nll_loss": 0.040362752974033356,
2129
+ "rewards/accuracies": 1.0,
2130
+ "rewards/chosen": -0.001746311434544623,
2131
+ "rewards/margins": 0.20069590210914612,
2132
+ "rewards/rejected": -0.2024422138929367,
2133
+ "step": 1180
2134
+ },
2135
+ {
2136
+ "epoch": 2.42,
2137
+ "grad_norm": 2.53125,
2138
+ "learning_rate": 1.4494275891311214e-06,
2139
+ "log_odds_chosen": 9.7469482421875,
2140
+ "log_odds_ratio": -0.012390781193971634,
2141
+ "logits/chosen": -1.8723634481430054,
2142
+ "logits/rejected": -1.8694031238555908,
2143
+ "logps/chosen": -0.045554567128419876,
2144
+ "logps/rejected": -4.15027379989624,
2145
+ "loss": 0.0354,
2146
+ "nll_loss": 0.03426050767302513,
2147
+ "rewards/accuracies": 1.0,
2148
+ "rewards/chosen": -0.002277728170156479,
2149
+ "rewards/margins": 0.20523595809936523,
2150
+ "rewards/rejected": -0.2075137197971344,
2151
+ "step": 1190
2152
+ },
2153
+ {
2154
+ "epoch": 2.44,
2155
+ "grad_norm": 2.140625,
2156
+ "learning_rate": 1.4433756729740647e-06,
2157
+ "log_odds_chosen": 10.098226547241211,
2158
+ "log_odds_ratio": -0.0019627574365586042,
2159
+ "logits/chosen": -1.9441982507705688,
2160
+ "logits/rejected": -1.9426124095916748,
2161
+ "logps/chosen": -0.02355217933654785,
2162
+ "logps/rejected": -4.209158897399902,
2163
+ "loss": 0.0457,
2164
+ "nll_loss": 0.053135812282562256,
2165
+ "rewards/accuracies": 1.0,
2166
+ "rewards/chosen": -0.0011776090832427144,
2167
+ "rewards/margins": 0.20928029716014862,
2168
+ "rewards/rejected": -0.2104579210281372,
2169
+ "step": 1200
2170
+ },
2171
+ {
2172
+ "epoch": 2.46,
2173
+ "grad_norm": 3.609375,
2174
+ "learning_rate": 1.4373989364401727e-06,
2175
+ "log_odds_chosen": 9.691657066345215,
2176
+ "log_odds_ratio": -0.004629252012819052,
2177
+ "logits/chosen": -1.9337596893310547,
2178
+ "logits/rejected": -1.931654930114746,
2179
+ "logps/chosen": -0.06260120123624802,
2180
+ "logps/rejected": -4.166058540344238,
2181
+ "loss": 0.0411,
2182
+ "nll_loss": 0.03751087933778763,
2183
+ "rewards/accuracies": 1.0,
2184
+ "rewards/chosen": -0.003130060387775302,
2185
+ "rewards/margins": 0.20517286658287048,
2186
+ "rewards/rejected": -0.20830294489860535,
2187
+ "step": 1210
2188
+ },
2189
+ {
2190
+ "epoch": 2.48,
2191
+ "grad_norm": 1.3984375,
2192
+ "learning_rate": 1.4314958357846706e-06,
2193
+ "log_odds_chosen": 9.472856521606445,
2194
+ "log_odds_ratio": -0.00650912756100297,
2195
+ "logits/chosen": -1.95502507686615,
2196
+ "logits/rejected": -1.951743721961975,
2197
+ "logps/chosen": -0.07376778870820999,
2198
+ "logps/rejected": -4.022158145904541,
2199
+ "loss": 0.0357,
2200
+ "nll_loss": 0.03547949343919754,
2201
+ "rewards/accuracies": 1.0,
2202
+ "rewards/chosen": -0.003688389901071787,
2203
+ "rewards/margins": 0.19741952419281006,
2204
+ "rewards/rejected": -0.2011079341173172,
2205
+ "step": 1220
2206
+ },
2207
+ {
2208
+ "epoch": 2.5,
2209
+ "grad_norm": 2.59375,
2210
+ "learning_rate": 1.4256648712805027e-06,
2211
+ "log_odds_chosen": 10.319469451904297,
2212
+ "log_odds_ratio": -0.001875861780717969,
2213
+ "logits/chosen": -1.9282852411270142,
2214
+ "logits/rejected": -1.9243885278701782,
2215
+ "logps/chosen": -0.0374315045773983,
2216
+ "logps/rejected": -4.336482524871826,
2217
+ "loss": 0.0412,
2218
+ "nll_loss": 0.0424610860645771,
2219
+ "rewards/accuracies": 1.0,
2220
+ "rewards/chosen": -0.0018715756013989449,
2221
+ "rewards/margins": 0.21495255827903748,
2222
+ "rewards/rejected": -0.21682414412498474,
2223
+ "step": 1230
2224
+ },
2225
+ {
2226
+ "epoch": 2.52,
2227
+ "grad_norm": 3.78125,
2228
+ "learning_rate": 1.419904585617662e-06,
2229
+ "log_odds_chosen": 10.347482681274414,
2230
+ "log_odds_ratio": -0.0033975064288824797,
2231
+ "logits/chosen": -1.9401060342788696,
2232
+ "logits/rejected": -1.9385757446289062,
2233
+ "logps/chosen": -0.04415961354970932,
2234
+ "logps/rejected": -4.426904201507568,
2235
+ "loss": 0.0403,
2236
+ "nll_loss": 0.04613417014479637,
2237
+ "rewards/accuracies": 1.0,
2238
+ "rewards/chosen": -0.0022079809568822384,
2239
+ "rewards/margins": 0.21913722157478333,
2240
+ "rewards/rejected": -0.2213451862335205,
2241
+ "step": 1240
2242
+ },
2243
+ {
2244
+ "epoch": 2.54,
2245
+ "grad_norm": 2.53125,
2246
+ "learning_rate": 1.4142135623730952e-06,
2247
+ "log_odds_chosen": 10.331408500671387,
2248
+ "log_odds_ratio": -0.0018885558238252997,
2249
+ "logits/chosen": -1.940860390663147,
2250
+ "logits/rejected": -1.9384444952011108,
2251
+ "logps/chosen": -0.03379340097308159,
2252
+ "logps/rejected": -4.056138038635254,
2253
+ "loss": 0.0393,
2254
+ "nll_loss": 0.030756259337067604,
2255
+ "rewards/accuracies": 1.0,
2256
+ "rewards/chosen": -0.0016896702582016587,
2257
+ "rewards/margins": 0.20111723244190216,
2258
+ "rewards/rejected": -0.20280690491199493,
2259
+ "step": 1250
2260
+ },
2261
+ {
2262
+ "epoch": 2.56,
2263
+ "grad_norm": 2.109375,
2264
+ "learning_rate": 1.4085904245475275e-06,
2265
+ "log_odds_chosen": 10.41731071472168,
2266
+ "log_odds_ratio": -0.012277336791157722,
2267
+ "logits/chosen": -1.9542973041534424,
2268
+ "logits/rejected": -1.953238844871521,
2269
+ "logps/chosen": -0.030469322577118874,
2270
+ "logps/rejected": -4.482726097106934,
2271
+ "loss": 0.0407,
2272
+ "nll_loss": 0.04968537390232086,
2273
+ "rewards/accuracies": 1.0,
2274
+ "rewards/chosen": -0.001523466082289815,
2275
+ "rewards/margins": 0.22261285781860352,
2276
+ "rewards/rejected": -0.2241363227367401,
2277
+ "step": 1260
2278
+ },
2279
+ {
2280
+ "epoch": 2.58,
2281
+ "grad_norm": 2.75,
2282
+ "learning_rate": 1.4030338331657844e-06,
2283
+ "log_odds_chosen": 10.663671493530273,
2284
+ "log_odds_ratio": -0.005900475196540356,
2285
+ "logits/chosen": -1.8991870880126953,
2286
+ "logits/rejected": -1.8935034275054932,
2287
+ "logps/chosen": -0.036128245294094086,
2288
+ "logps/rejected": -4.496649742126465,
2289
+ "loss": 0.036,
2290
+ "nll_loss": 0.032302360981702805,
2291
+ "rewards/accuracies": 1.0,
2292
+ "rewards/chosen": -0.00180641224142164,
2293
+ "rewards/margins": 0.2230260670185089,
2294
+ "rewards/rejected": -0.22483249008655548,
2295
+ "step": 1270
2296
+ },
2297
+ {
2298
+ "epoch": 2.6,
2299
+ "grad_norm": 1.984375,
2300
+ "learning_rate": 1.3975424859373688e-06,
2301
+ "log_odds_chosen": 9.953948020935059,
2302
+ "log_odds_ratio": -0.002804636722430587,
2303
+ "logits/chosen": -1.9552701711654663,
2304
+ "logits/rejected": -1.9517933130264282,
2305
+ "logps/chosen": -0.04314829781651497,
2306
+ "logps/rejected": -4.2899580001831055,
2307
+ "loss": 0.0389,
2308
+ "nll_loss": 0.03743810951709747,
2309
+ "rewards/accuracies": 1.0,
2310
+ "rewards/chosen": -0.002157414797693491,
2311
+ "rewards/margins": 0.21234047412872314,
2312
+ "rewards/rejected": -0.21449792385101318,
2313
+ "step": 1280
2314
+ },
2315
+ {
2316
+ "epoch": 2.62,
2317
+ "grad_norm": 2.046875,
2318
+ "learning_rate": 1.3921151159742616e-06,
2319
+ "log_odds_chosen": 9.60466194152832,
2320
+ "log_odds_ratio": -0.021028626710176468,
2321
+ "logits/chosen": -1.8756396770477295,
2322
+ "logits/rejected": -1.8728761672973633,
2323
+ "logps/chosen": -0.043241944164037704,
2324
+ "logps/rejected": -4.110201835632324,
2325
+ "loss": 0.0374,
2326
+ "nll_loss": 0.03517068549990654,
2327
+ "rewards/accuracies": 0.987500011920929,
2328
+ "rewards/chosen": -0.0021620974875986576,
2329
+ "rewards/margins": 0.20334799587726593,
2330
+ "rewards/rejected": -0.20551009476184845,
2331
+ "step": 1290
2332
+ },
2333
+ {
2334
+ "epoch": 2.64,
2335
+ "grad_norm": 2.171875,
2336
+ "learning_rate": 1.386750490563073e-06,
2337
+ "log_odds_chosen": 10.577958106994629,
2338
+ "log_odds_ratio": -0.0030032056383788586,
2339
+ "logits/chosen": -1.9832937717437744,
2340
+ "logits/rejected": -1.9783260822296143,
2341
+ "logps/chosen": -0.06359104067087173,
2342
+ "logps/rejected": -4.623613357543945,
2343
+ "loss": 0.0396,
2344
+ "nll_loss": 0.04861373454332352,
2345
+ "rewards/accuracies": 1.0,
2346
+ "rewards/chosen": -0.003179552499204874,
2347
+ "rewards/margins": 0.22800114750862122,
2348
+ "rewards/rejected": -0.23118066787719727,
2349
+ "step": 1300
2350
+ },
2351
+ {
2352
+ "epoch": 2.66,
2353
+ "grad_norm": 4.1875,
2354
+ "learning_rate": 1.3814474099888442e-06,
2355
+ "log_odds_chosen": 10.487229347229004,
2356
+ "log_odds_ratio": -0.0037563578225672245,
2357
+ "logits/chosen": -1.9412161111831665,
2358
+ "logits/rejected": -1.9390010833740234,
2359
+ "logps/chosen": -0.04359542205929756,
2360
+ "logps/rejected": -4.6636810302734375,
2361
+ "loss": 0.0433,
2362
+ "nll_loss": 0.03972161188721657,
2363
+ "rewards/accuracies": 1.0,
2364
+ "rewards/chosen": -0.0021797711960971355,
2365
+ "rewards/margins": 0.2310042679309845,
2366
+ "rewards/rejected": -0.2331840544939041,
2367
+ "step": 1310
2368
+ },
2369
+ {
2370
+ "epoch": 2.68,
2371
+ "grad_norm": 1.796875,
2372
+ "learning_rate": 1.376204706407951e-06,
2373
+ "log_odds_chosen": 10.476730346679688,
2374
+ "log_odds_ratio": -0.002192417625337839,
2375
+ "logits/chosen": -1.9325330257415771,
2376
+ "logits/rejected": -1.930166244506836,
2377
+ "logps/chosen": -0.02953045442700386,
2378
+ "logps/rejected": -4.279176712036133,
2379
+ "loss": 0.0353,
2380
+ "nll_loss": 0.03177348151803017,
2381
+ "rewards/accuracies": 1.0,
2382
+ "rewards/chosen": -0.001476522651501,
2383
+ "rewards/margins": 0.2124823033809662,
2384
+ "rewards/rejected": -0.21395882964134216,
2385
+ "step": 1320
2386
+ },
2387
+ {
2388
+ "epoch": 2.7,
2389
+ "grad_norm": 2.484375,
2390
+ "learning_rate": 1.3710212427677044e-06,
2391
+ "log_odds_chosen": 10.246335983276367,
2392
+ "log_odds_ratio": -0.0470537394285202,
2393
+ "logits/chosen": -1.9275156259536743,
2394
+ "logits/rejected": -1.9267303943634033,
2395
+ "logps/chosen": -0.07270321995019913,
2396
+ "logps/rejected": -4.131697177886963,
2397
+ "loss": 0.036,
2398
+ "nll_loss": 0.031709469854831696,
2399
+ "rewards/accuracies": 0.987500011920929,
2400
+ "rewards/chosen": -0.0036351613234728575,
2401
+ "rewards/margins": 0.20294968783855438,
2402
+ "rewards/rejected": -0.2065848559141159,
2403
+ "step": 1330
2404
+ },
2405
+ {
2406
+ "epoch": 2.72,
2407
+ "grad_norm": 2.578125,
2408
+ "learning_rate": 1.3658959117703826e-06,
2409
+ "log_odds_chosen": 10.568267822265625,
2410
+ "log_odds_ratio": -0.003112158738076687,
2411
+ "logits/chosen": -1.948743462562561,
2412
+ "logits/rejected": -1.9471248388290405,
2413
+ "logps/chosen": -0.028313973918557167,
2414
+ "logps/rejected": -4.5648627281188965,
2415
+ "loss": 0.0378,
2416
+ "nll_loss": 0.041449058800935745,
2417
+ "rewards/accuracies": 1.0,
2418
+ "rewards/chosen": -0.001415698672644794,
2419
+ "rewards/margins": 0.22682742774486542,
2420
+ "rewards/rejected": -0.2282431572675705,
2421
+ "step": 1340
2422
+ },
2423
+ {
2424
+ "epoch": 2.74,
2425
+ "grad_norm": 3.546875,
2426
+ "learning_rate": 1.3608276348795436e-06,
2427
+ "log_odds_chosen": 10.491666793823242,
2428
+ "log_odds_ratio": -0.0048934281803667545,
2429
+ "logits/chosen": -1.9433883428573608,
2430
+ "logits/rejected": -1.9420783519744873,
2431
+ "logps/chosen": -0.03949974104762077,
2432
+ "logps/rejected": -4.376409530639648,
2433
+ "loss": 0.0414,
2434
+ "nll_loss": 0.03402886912226677,
2435
+ "rewards/accuracies": 1.0,
2436
+ "rewards/chosen": -0.001974987331777811,
2437
+ "rewards/margins": 0.21684548258781433,
2438
+ "rewards/rejected": -0.2188204824924469,
2439
+ "step": 1350
2440
+ },
2441
+ {
2442
+ "epoch": 2.76,
2443
+ "grad_norm": 2.265625,
2444
+ "learning_rate": 1.355815361366601e-06,
2445
+ "log_odds_chosen": 10.092164993286133,
2446
+ "log_odds_ratio": -0.008559630252420902,
2447
+ "logits/chosen": -1.945858359336853,
2448
+ "logits/rejected": -1.9418405294418335,
2449
+ "logps/chosen": -0.06214701011776924,
2450
+ "logps/rejected": -4.4873552322387695,
2451
+ "loss": 0.0359,
2452
+ "nll_loss": 0.03512818366289139,
2453
+ "rewards/accuracies": 1.0,
2454
+ "rewards/chosen": -0.0031073507852852345,
2455
+ "rewards/margins": 0.2212604284286499,
2456
+ "rewards/rejected": -0.22436778247356415,
2457
+ "step": 1360
2458
+ },
2459
+ {
2460
+ "epoch": 2.78,
2461
+ "grad_norm": 1.9296875,
2462
+ "learning_rate": 1.350858067395748e-06,
2463
+ "log_odds_chosen": 10.725711822509766,
2464
+ "log_odds_ratio": -0.020754020661115646,
2465
+ "logits/chosen": -1.9590380191802979,
2466
+ "logits/rejected": -1.953980803489685,
2467
+ "logps/chosen": -0.08093880116939545,
2468
+ "logps/rejected": -4.96276330947876,
2469
+ "loss": 0.0419,
2470
+ "nll_loss": 0.0406344048678875,
2471
+ "rewards/accuracies": 0.987500011920929,
2472
+ "rewards/chosen": -0.004046940244734287,
2473
+ "rewards/margins": 0.2440912276506424,
2474
+ "rewards/rejected": -0.2481381893157959,
2475
+ "step": 1370
2476
+ },
2477
+ {
2478
+ "epoch": 2.8,
2479
+ "grad_norm": 3.359375,
2480
+ "learning_rate": 1.345954755145414e-06,
2481
+ "log_odds_chosen": 10.852327346801758,
2482
+ "log_odds_ratio": -0.0012642950750887394,
2483
+ "logits/chosen": -1.8603143692016602,
2484
+ "logits/rejected": -1.85476815700531,
2485
+ "logps/chosen": -0.03734128549695015,
2486
+ "logps/rejected": -4.723127841949463,
2487
+ "loss": 0.04,
2488
+ "nll_loss": 0.025891494005918503,
2489
+ "rewards/accuracies": 1.0,
2490
+ "rewards/chosen": -0.0018670640420168638,
2491
+ "rewards/margins": 0.23428933322429657,
2492
+ "rewards/rejected": -0.2361563891172409,
2493
+ "step": 1380
2494
+ },
2495
+ {
2496
+ "epoch": 2.83,
2497
+ "grad_norm": 2.390625,
2498
+ "learning_rate": 1.3411044519645502e-06,
2499
+ "log_odds_chosen": 10.230730056762695,
2500
+ "log_odds_ratio": -0.002351417439058423,
2501
+ "logits/chosen": -1.8683515787124634,
2502
+ "logits/rejected": -1.86406672000885,
2503
+ "logps/chosen": -0.044098712503910065,
2504
+ "logps/rejected": -4.59433126449585,
2505
+ "loss": 0.0331,
2506
+ "nll_loss": 0.03638423979282379,
2507
+ "rewards/accuracies": 1.0,
2508
+ "rewards/chosen": -0.0022049355320632458,
2509
+ "rewards/margins": 0.22751164436340332,
2510
+ "rewards/rejected": -0.22971658408641815,
2511
+ "step": 1390
2512
+ },
2513
+ {
2514
+ "epoch": 2.85,
2515
+ "grad_norm": 2.296875,
2516
+ "learning_rate": 1.3363062095621222e-06,
2517
+ "log_odds_chosen": 9.968645095825195,
2518
+ "log_odds_ratio": -0.004948216024786234,
2519
+ "logits/chosen": -1.9218571186065674,
2520
+ "logits/rejected": -1.9199250936508179,
2521
+ "logps/chosen": -0.06343747675418854,
2522
+ "logps/rejected": -4.219951629638672,
2523
+ "loss": 0.0353,
2524
+ "nll_loss": 0.03614867106080055,
2525
+ "rewards/accuracies": 1.0,
2526
+ "rewards/chosen": -0.0031718737445771694,
2527
+ "rewards/margins": 0.20782573521137238,
2528
+ "rewards/rejected": -0.21099761128425598,
2529
+ "step": 1400
2530
+ },
2531
+ {
2532
+ "epoch": 2.87,
2533
+ "grad_norm": 2.609375,
2534
+ "learning_rate": 1.3315591032282687e-06,
2535
+ "log_odds_chosen": 10.9727783203125,
2536
+ "log_odds_ratio": -0.004531107842922211,
2537
+ "logits/chosen": -1.9281492233276367,
2538
+ "logits/rejected": -1.9244182109832764,
2539
+ "logps/chosen": -0.03406985104084015,
2540
+ "logps/rejected": -4.806418418884277,
2541
+ "loss": 0.0409,
2542
+ "nll_loss": 0.04478253796696663,
2543
+ "rewards/accuracies": 1.0,
2544
+ "rewards/chosen": -0.0017034925986081362,
2545
+ "rewards/margins": 0.2386174201965332,
2546
+ "rewards/rejected": -0.24032095074653625,
2547
+ "step": 1410
2548
+ },
2549
+ {
2550
+ "epoch": 2.89,
2551
+ "grad_norm": 2.21875,
2552
+ "learning_rate": 1.3268622310856882e-06,
2553
+ "log_odds_chosen": 11.176974296569824,
2554
+ "log_odds_ratio": -0.003099204506725073,
2555
+ "logits/chosen": -1.9768095016479492,
2556
+ "logits/rejected": -1.9735969305038452,
2557
+ "logps/chosen": -0.03214428573846817,
2558
+ "logps/rejected": -4.844766616821289,
2559
+ "loss": 0.0368,
2560
+ "nll_loss": 0.03034246899187565,
2561
+ "rewards/accuracies": 1.0,
2562
+ "rewards/chosen": -0.0016072141006588936,
2563
+ "rewards/margins": 0.240631103515625,
2564
+ "rewards/rejected": -0.24223831295967102,
2565
+ "step": 1420
2566
+ },
2567
+ {
2568
+ "epoch": 2.91,
2569
+ "grad_norm": 3.953125,
2570
+ "learning_rate": 1.3222147133698626e-06,
2571
+ "log_odds_chosen": 9.971961975097656,
2572
+ "log_odds_ratio": -0.009039332158863544,
2573
+ "logits/chosen": -1.8924024105072021,
2574
+ "logits/rejected": -1.8878147602081299,
2575
+ "logps/chosen": -0.05458871275186539,
2576
+ "logps/rejected": -4.532398700714111,
2577
+ "loss": 0.041,
2578
+ "nll_loss": 0.033775750547647476,
2579
+ "rewards/accuracies": 1.0,
2580
+ "rewards/chosen": -0.002729435684159398,
2581
+ "rewards/margins": 0.2238904982805252,
2582
+ "rewards/rejected": -0.2266199290752411,
2583
+ "step": 1430
2584
+ },
2585
+ {
2586
+ "epoch": 2.93,
2587
+ "grad_norm": 2.09375,
2588
+ "learning_rate": 1.3176156917368248e-06,
2589
+ "log_odds_chosen": 10.354188919067383,
2590
+ "log_odds_ratio": -0.0086409542709589,
2591
+ "logits/chosen": -1.871347427368164,
2592
+ "logits/rejected": -1.8664233684539795,
2593
+ "logps/chosen": -0.0295790396630764,
2594
+ "logps/rejected": -4.4826579093933105,
2595
+ "loss": 0.0313,
2596
+ "nll_loss": 0.030693087726831436,
2597
+ "rewards/accuracies": 1.0,
2598
+ "rewards/chosen": -0.0014789520064368844,
2599
+ "rewards/margins": 0.22265395522117615,
2600
+ "rewards/rejected": -0.22413286566734314,
2601
+ "step": 1440
2602
+ },
2603
+ {
2604
+ "epoch": 2.95,
2605
+ "grad_norm": 2.578125,
2606
+ "learning_rate": 1.3130643285972255e-06,
2607
+ "log_odds_chosen": 10.23839282989502,
2608
+ "log_odds_ratio": -0.008670750074088573,
2609
+ "logits/chosen": -1.9214890003204346,
2610
+ "logits/rejected": -1.9176349639892578,
2611
+ "logps/chosen": -0.04931662604212761,
2612
+ "logps/rejected": -4.222093105316162,
2613
+ "loss": 0.0354,
2614
+ "nll_loss": 0.03125245124101639,
2615
+ "rewards/accuracies": 1.0,
2616
+ "rewards/chosen": -0.002465831348672509,
2617
+ "rewards/margins": 0.20863883197307587,
2618
+ "rewards/rejected": -0.21110467612743378,
2619
+ "step": 1450
2620
+ },
2621
+ {
2622
+ "epoch": 2.97,
2623
+ "grad_norm": 1.890625,
2624
+ "learning_rate": 1.3085598064755342e-06,
2625
+ "log_odds_chosen": 10.399072647094727,
2626
+ "log_odds_ratio": -0.003747476963326335,
2627
+ "logits/chosen": -1.8761295080184937,
2628
+ "logits/rejected": -1.8740062713623047,
2629
+ "logps/chosen": -0.048522982746362686,
2630
+ "logps/rejected": -4.560256004333496,
2631
+ "loss": 0.0359,
2632
+ "nll_loss": 0.042360566556453705,
2633
+ "rewards/accuracies": 1.0,
2634
+ "rewards/chosen": -0.0024261490907520056,
2635
+ "rewards/margins": 0.2255866527557373,
2636
+ "rewards/rejected": -0.2280128002166748,
2637
+ "step": 1460
2638
+ },
2639
+ {
2640
+ "epoch": 2.99,
2641
+ "grad_norm": 2.140625,
2642
+ "learning_rate": 1.3041013273932528e-06,
2643
+ "log_odds_chosen": 11.28456974029541,
2644
+ "log_odds_ratio": -0.0021586515940725803,
2645
+ "logits/chosen": -1.9435927867889404,
2646
+ "logits/rejected": -1.9409191608428955,
2647
+ "logps/chosen": -0.03155381605029106,
2648
+ "logps/rejected": -5.088595390319824,
2649
+ "loss": 0.0357,
2650
+ "nll_loss": 0.028978243470191956,
2651
+ "rewards/accuracies": 1.0,
2652
+ "rewards/chosen": -0.0015776908257976174,
2653
+ "rewards/margins": 0.25285208225250244,
2654
+ "rewards/rejected": -0.25442975759506226,
2655
+ "step": 1470
2656
+ },
2657
+ {
2658
+ "epoch": 3.0,
2659
+ "step": 1476,
2660
+ "total_flos": 0.0,
2661
+ "train_loss": 0.1731818498170311,
2662
+ "train_runtime": 10691.148,
2663
+ "train_samples_per_second": 8.831,
2664
+ "train_steps_per_second": 0.138
2665
+ }
2666
+ ],
2667
+ "logging_steps": 10,
2668
+ "max_steps": 1476,
2669
+ "num_input_tokens_seen": 0,
2670
+ "num_train_epochs": 3,
2671
+ "save_steps": 500,
2672
+ "total_flos": 0.0,
2673
+ "train_batch_size": 8,
2674
+ "trial_name": null,
2675
+ "trial_params": null
2676
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e60dea34ea29f59d1e18b2a0572c65d481d8e1906b0ca685e3f89aa8e1c7e981
3
+ size 5240