Dragon116rus commited on
Commit
20934e4
·
verified ·
1 Parent(s): ff27f3c

Saving train state of step 1000

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "openai/whisper-base",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
@@ -13,17 +13,17 @@
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
16
- "d_model": 512,
17
- "decoder_attention_heads": 8,
18
- "decoder_ffn_dim": 2048,
19
  "decoder_layerdrop": 0.0,
20
- "decoder_layers": 6,
21
  "decoder_start_token_id": 50258,
22
  "dropout": 0.0,
23
- "encoder_attention_heads": 8,
24
- "encoder_ffn_dim": 2048,
25
  "encoder_layerdrop": 0.0,
26
- "encoder_layers": 6,
27
  "eos_token_id": 50257,
28
  "forced_decoder_ids": [
29
  [
@@ -52,7 +52,7 @@
52
  "max_target_positions": 448,
53
  "median_filter_width": 7,
54
  "model_type": "whisper",
55
- "num_hidden_layers": 6,
56
  "num_mel_bins": 80,
57
  "pad_token_id": 50257,
58
  "scale_embedding": false,
 
1
  {
2
+ "_name_or_path": "openai/whisper-tiny",
3
  "activation_dropout": 0.0,
4
  "activation_function": "gelu",
5
  "apply_spec_augment": false,
 
13
  ],
14
  "bos_token_id": 50257,
15
  "classifier_proj_size": 256,
16
+ "d_model": 384,
17
+ "decoder_attention_heads": 6,
18
+ "decoder_ffn_dim": 1536,
19
  "decoder_layerdrop": 0.0,
20
+ "decoder_layers": 4,
21
  "decoder_start_token_id": 50258,
22
  "dropout": 0.0,
23
+ "encoder_attention_heads": 6,
24
+ "encoder_ffn_dim": 1536,
25
  "encoder_layerdrop": 0.0,
26
+ "encoder_layers": 4,
27
  "eos_token_id": 50257,
28
  "forced_decoder_ids": [
29
  [
 
52
  "max_target_positions": 448,
53
  "median_filter_width": 7,
54
  "model_type": "whisper",
55
+ "num_hidden_layers": 4,
56
  "num_mel_bins": 80,
57
  "pad_token_id": 50257,
58
  "scale_embedding": false,
distil-whisper/events.out.tfevents.1712414653.train02.3753928.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e83f1cd21a4e6043c2334dce2bfa96ecf8cb0d84dcb2d86fcbd7d1775027605f
3
- size 63446
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b51a92acd49bca43622c5913fb11ace5cd11f669217a6686bea3d17315306783
3
+ size 63793
distil-whisper/events.out.tfevents.1712421884.train02.3768589.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc0c06b7543dfcbb3a6135880755720796f84a317e6e45d303bc1d8515193bba
3
+ size 12458
dvclive/metrics.json CHANGED
@@ -1,14 +1,8 @@
1
  {
2
  "train": {
3
- "time": 2253.3530826568604,
4
- "epoch": 5,
5
  "learning_rate": 0.0001
6
  },
7
- "step": 5000,
8
- "eval": {
9
- "wer": 22.24298186860319,
10
- "wer_ortho": 28.8339752069018,
11
- "time": 245.54304361343384,
12
- "epoch": 4
13
- }
14
  }
 
1
  {
2
  "train": {
3
+ "time": 356.29300141334534,
4
+ "epoch": 1,
5
  "learning_rate": 0.0001
6
  },
7
+ "step": 1000
 
 
 
 
 
 
8
  }
dvclive/plots/metrics/train/epoch.tsv CHANGED
@@ -39,163 +39,3 @@ step epoch
39
  950 1
40
  975 1
41
  1000 1
42
- 1025 1
43
- 1050 1
44
- 1075 1
45
- 1100 1
46
- 1125 1
47
- 1150 1
48
- 1175 1
49
- 1200 1
50
- 1225 1
51
- 1250 1
52
- 1275 1
53
- 1300 1
54
- 1325 1
55
- 1350 1
56
- 1375 1
57
- 1400 1
58
- 1425 1
59
- 1450 1
60
- 1475 1
61
- 1500 1
62
- 1525 1
63
- 1550 1
64
- 1575 1
65
- 1600 1
66
- 1625 1
67
- 1650 1
68
- 1675 1
69
- 1700 1
70
- 1725 2
71
- 1750 2
72
- 1775 2
73
- 1800 2
74
- 1825 2
75
- 1850 2
76
- 1875 2
77
- 1900 2
78
- 1925 2
79
- 1950 2
80
- 1975 2
81
- 2000 2
82
- 2025 2
83
- 2050 2
84
- 2075 2
85
- 2100 2
86
- 2125 2
87
- 2150 2
88
- 2175 2
89
- 2200 2
90
- 2225 2
91
- 2250 2
92
- 2275 2
93
- 2300 2
94
- 2325 2
95
- 2350 2
96
- 2375 2
97
- 2400 2
98
- 2425 2
99
- 2450 2
100
- 2475 2
101
- 2500 2
102
- 2525 2
103
- 2550 2
104
- 2575 2
105
- 2600 3
106
- 2625 3
107
- 2650 3
108
- 2675 3
109
- 2700 3
110
- 2725 3
111
- 2750 3
112
- 2775 3
113
- 2800 3
114
- 2825 3
115
- 2850 3
116
- 2875 3
117
- 2900 3
118
- 2925 3
119
- 2950 3
120
- 2975 3
121
- 3000 3
122
- 3025 3
123
- 3050 3
124
- 3075 3
125
- 3100 3
126
- 3125 3
127
- 3150 3
128
- 3175 3
129
- 3200 3
130
- 3225 3
131
- 3250 3
132
- 3275 3
133
- 3300 3
134
- 3325 3
135
- 3350 3
136
- 3375 3
137
- 3400 3
138
- 3425 3
139
- 3450 4
140
- 3475 4
141
- 3500 4
142
- 3525 4
143
- 3550 4
144
- 3575 4
145
- 3600 4
146
- 3625 4
147
- 3650 4
148
- 3675 4
149
- 3700 4
150
- 3725 4
151
- 3750 4
152
- 3775 4
153
- 3800 4
154
- 3825 4
155
- 3850 4
156
- 3875 4
157
- 3900 4
158
- 3925 4
159
- 3950 4
160
- 3975 4
161
- 4000 4
162
- 4025 4
163
- 4050 4
164
- 4075 4
165
- 4100 4
166
- 4125 4
167
- 4150 4
168
- 4175 4
169
- 4200 4
170
- 4225 4
171
- 4250 4
172
- 4275 4
173
- 4300 4
174
- 4325 5
175
- 4350 5
176
- 4375 5
177
- 4400 5
178
- 4425 5
179
- 4450 5
180
- 4475 5
181
- 4500 5
182
- 4525 5
183
- 4550 5
184
- 4575 5
185
- 4600 5
186
- 4625 5
187
- 4650 5
188
- 4675 5
189
- 4700 5
190
- 4725 5
191
- 4750 5
192
- 4775 5
193
- 4800 5
194
- 4825 5
195
- 4850 5
196
- 4875 5
197
- 4900 5
198
- 4925 5
199
- 4950 5
200
- 4975 5
201
- 5000 5
 
39
  950 1
40
  975 1
41
  1000 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dvclive/plots/metrics/train/learning_rate.tsv CHANGED
@@ -1,6 +1,6 @@
1
  step learning_rate
2
- 25 4.4000000000000006e-05
3
- 50 9.4e-05
4
  75 0.0001
5
  100 0.0001
6
  125 0.0001
@@ -39,163 +39,3 @@ step learning_rate
39
  950 0.0001
40
  975 0.0001
41
  1000 0.0001
42
- 1025 0.0001
43
- 1050 0.0001
44
- 1075 0.0001
45
- 1100 0.0001
46
- 1125 0.0001
47
- 1150 0.0001
48
- 1175 0.0001
49
- 1200 0.0001
50
- 1225 0.0001
51
- 1250 0.0001
52
- 1275 0.0001
53
- 1300 0.0001
54
- 1325 0.0001
55
- 1350 0.0001
56
- 1375 0.0001
57
- 1400 0.0001
58
- 1425 0.0001
59
- 1450 0.0001
60
- 1475 0.0001
61
- 1500 0.0001
62
- 1525 0.0001
63
- 1550 0.0001
64
- 1575 0.0001
65
- 1600 0.0001
66
- 1625 0.0001
67
- 1650 0.0001
68
- 1675 0.0001
69
- 1700 0.0001
70
- 1725 0.0001
71
- 1750 0.0001
72
- 1775 0.0001
73
- 1800 0.0001
74
- 1825 0.0001
75
- 1850 0.0001
76
- 1875 0.0001
77
- 1900 0.0001
78
- 1925 0.0001
79
- 1950 0.0001
80
- 1975 0.0001
81
- 2000 0.0001
82
- 2025 0.0001
83
- 2050 0.0001
84
- 2075 0.0001
85
- 2100 0.0001
86
- 2125 0.0001
87
- 2150 0.0001
88
- 2175 0.0001
89
- 2200 0.0001
90
- 2225 0.0001
91
- 2250 0.0001
92
- 2275 0.0001
93
- 2300 0.0001
94
- 2325 0.0001
95
- 2350 0.0001
96
- 2375 0.0001
97
- 2400 0.0001
98
- 2425 0.0001
99
- 2450 0.0001
100
- 2475 0.0001
101
- 2500 0.0001
102
- 2525 0.0001
103
- 2550 0.0001
104
- 2575 0.0001
105
- 2600 0.0001
106
- 2625 0.0001
107
- 2650 0.0001
108
- 2675 0.0001
109
- 2700 0.0001
110
- 2725 0.0001
111
- 2750 0.0001
112
- 2775 0.0001
113
- 2800 0.0001
114
- 2825 0.0001
115
- 2850 0.0001
116
- 2875 0.0001
117
- 2900 0.0001
118
- 2925 0.0001
119
- 2950 0.0001
120
- 2975 0.0001
121
- 3000 0.0001
122
- 3025 0.0001
123
- 3050 0.0001
124
- 3075 0.0001
125
- 3100 0.0001
126
- 3125 0.0001
127
- 3150 0.0001
128
- 3175 0.0001
129
- 3200 0.0001
130
- 3225 0.0001
131
- 3250 0.0001
132
- 3275 0.0001
133
- 3300 0.0001
134
- 3325 0.0001
135
- 3350 0.0001
136
- 3375 0.0001
137
- 3400 0.0001
138
- 3425 0.0001
139
- 3450 0.0001
140
- 3475 0.0001
141
- 3500 0.0001
142
- 3525 0.0001
143
- 3550 0.0001
144
- 3575 0.0001
145
- 3600 0.0001
146
- 3625 0.0001
147
- 3650 0.0001
148
- 3675 0.0001
149
- 3700 0.0001
150
- 3725 0.0001
151
- 3750 0.0001
152
- 3775 0.0001
153
- 3800 0.0001
154
- 3825 0.0001
155
- 3850 0.0001
156
- 3875 0.0001
157
- 3900 0.0001
158
- 3925 0.0001
159
- 3950 0.0001
160
- 3975 0.0001
161
- 4000 0.0001
162
- 4025 0.0001
163
- 4050 0.0001
164
- 4075 0.0001
165
- 4100 0.0001
166
- 4125 0.0001
167
- 4150 0.0001
168
- 4175 0.0001
169
- 4200 0.0001
170
- 4225 0.0001
171
- 4250 0.0001
172
- 4275 0.0001
173
- 4300 0.0001
174
- 4325 0.0001
175
- 4350 0.0001
176
- 4375 0.0001
177
- 4400 0.0001
178
- 4425 0.0001
179
- 4450 0.0001
180
- 4475 0.0001
181
- 4500 0.0001
182
- 4525 0.0001
183
- 4550 0.0001
184
- 4575 0.0001
185
- 4600 0.0001
186
- 4625 0.0001
187
- 4650 0.0001
188
- 4675 0.0001
189
- 4700 0.0001
190
- 4725 0.0001
191
- 4750 0.0001
192
- 4775 0.0001
193
- 4800 0.0001
194
- 4825 0.0001
195
- 4850 0.0001
196
- 4875 0.0001
197
- 4900 0.0001
198
- 4925 0.0001
199
- 4950 0.0001
200
- 4975 0.0001
201
- 5000 0.0001
 
1
  step learning_rate
2
+ 25 4e-05
3
+ 50 9e-05
4
  75 0.0001
5
  100 0.0001
6
  125 0.0001
 
39
  950 0.0001
40
  975 0.0001
41
  1000 0.0001
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dvclive/plots/metrics/train/time.tsv CHANGED
@@ -1,201 +1,41 @@
1
  step time
2
- 25 11.684435844421387
3
- 50 21.286893844604492
4
- 75 30.893161058425903
5
- 100 40.50417184829712
6
- 125 50.126877784729004
7
- 150 59.74510478973389
8
- 175 69.3599362373352
9
- 200 78.9830493927002
10
- 225 88.60592222213745
11
- 250 98.23294377326965
12
- 275 107.85366106033325
13
- 300 117.4715666770935
14
- 325 127.09286832809448
15
- 350 136.72641706466675
16
- 375 146.37023186683655
17
- 400 156.02068185806274
18
- 425 165.67155933380127
19
- 450 175.32003140449524
20
- 475 184.960928440094
21
- 500 194.60969734191895
22
- 525 204.2479236125946
23
- 550 213.8882131576538
24
- 575 223.52437162399292
25
- 600 233.17730355262756
26
- 625 242.82540273666382
27
- 650 252.46957540512085
28
- 675 262.1174519062042
29
- 700 271.77419233322144
30
- 725 281.43049025535583
31
- 750 291.0802092552185
32
- 775 300.7343370914459
33
- 800 310.38205337524414
34
- 825 320.03323912620544
35
- 850 329.6830699443817
36
- 875 341.0324788093567
37
- 900 350.6875424385071
38
- 925 360.3380663394928
39
- 950 369.98790550231934
40
- 975 379.62979459762573
41
- 1000 389.28764247894287
42
- 1025 477.43934750556946
43
- 1050 487.06905603408813
44
- 1075 496.70161390304565
45
- 1100 506.34014201164246
46
- 1125 515.968909740448
47
- 1150 525.6247124671936
48
- 1175 535.2812929153442
49
- 1200 544.94753074646
50
- 1225 554.6170375347137
51
- 1250 564.2695646286011
52
- 1275 573.9344782829285
53
- 1300 583.5934343338013
54
- 1325 593.2619092464447
55
- 1350 602.9200406074524
56
- 1375 612.5834064483643
57
- 1400 622.2427930831909
58
- 1425 631.9039030075073
59
- 1450 641.5672912597656
60
- 1475 651.2363836765289
61
- 1500 660.9025223255157
62
- 1525 670.5803966522217
63
- 1550 680.2368001937866
64
- 1575 689.8935434818268
65
- 1600 699.561598777771
66
- 1625 709.2326860427856
67
- 1650 718.9015653133392
68
- 1675 728.5651636123657
69
- 1700 738.2388761043549
70
- 1725 749.058812379837
71
- 1750 758.715169429779
72
- 1775 768.3791604042053
73
- 1800 778.0330898761749
74
- 1825 787.7043738365173
75
- 1850 797.3711838722229
76
- 1875 807.033620595932
77
- 1900 816.7156624794006
78
- 1925 826.369535446167
79
- 1950 836.0318987369537
80
- 1975 845.6994693279266
81
- 2000 855.3777701854706
82
- 2025 945.9205918312073
83
- 2050 955.5570874214172
84
- 2075 965.183685541153
85
- 2100 974.8230764865875
86
- 2125 984.4629333019257
87
- 2150 994.1015644073486
88
- 2175 1003.7421622276306
89
- 2200 1013.3943798542023
90
- 2225 1023.0340690612793
91
- 2250 1032.6920757293701
92
- 2275 1042.355932712555
93
- 2300 1052.0133595466614
94
- 2325 1061.6764154434204
95
- 2350 1071.3519291877747
96
- 2375 1081.009160041809
97
- 2400 1090.6628205776215
98
- 2425 1100.3322820663452
99
- 2450 1109.9939272403717
100
- 2475 1119.6580357551575
101
- 2500 1129.3195736408234
102
- 2525 1138.9884753227234
103
- 2550 1148.6439034938812
104
- 2575 1158.2940447330475
105
- 2600 1169.1666514873505
106
- 2625 1178.8291101455688
107
- 2650 1188.4843485355377
108
- 2675 1198.1412012577057
109
- 2700 1207.7901220321655
110
- 2725 1217.4453997612
111
- 2750 1227.106692790985
112
- 2775 1236.758858203888
113
- 2800 1246.4241180419922
114
- 2825 1256.0781526565552
115
- 2850 1265.7430965900421
116
- 2875 1275.3995866775513
117
- 2900 1285.067929983139
118
- 2925 1294.7319355010986
119
- 2950 1304.398491859436
120
- 2975 1314.061797618866
121
- 3000 1323.7229840755463
122
- 3025 1409.9547500610352
123
- 3050 1419.5985000133514
124
- 3075 1429.2499001026154
125
- 3100 1438.9136247634888
126
- 3125 1448.574645280838
127
- 3150 1458.2276797294617
128
- 3175 1467.8867342472076
129
- 3200 1477.555083990097
130
- 3225 1487.21031498909
131
- 3250 1496.8673069477081
132
- 3275 1506.5304148197174
133
- 3300 1516.191492319107
134
- 3325 1525.8438050746918
135
- 3350 1535.5006895065308
136
- 3375 1545.1577517986298
137
- 3400 1554.8170964717865
138
- 3425 1564.486836194992
139
- 3450 1575.3070306777954
140
- 3475 1584.952998638153
141
- 3500 1594.5967934131622
142
- 3525 1604.240556716919
143
- 3550 1613.887773990631
144
- 3575 1623.5305771827698
145
- 3600 1633.1819751262665
146
- 3625 1642.8412942886353
147
- 3650 1652.4885885715485
148
- 3675 1662.1482954025269
149
- 3700 1671.7933938503265
150
- 3725 1681.4414370059967
151
- 3750 1691.099225282669
152
- 3775 1700.7563977241516
153
- 3800 1710.4122366905212
154
- 3825 1720.0654191970825
155
- 3850 1729.7193558216095
156
- 3875 1739.3637006282806
157
- 3900 1749.0059382915497
158
- 3925 1758.6463613510132
159
- 3950 1768.295547246933
160
- 3975 1777.9552805423737
161
- 4000 1787.6112430095673
162
- 4025 1875.8111493587494
163
- 4050 1885.4643814563751
164
- 4075 1895.1054847240448
165
- 4100 1904.7525160312653
166
- 4125 1914.390187740326
167
- 4150 1924.0471398830414
168
- 4175 1933.6979429721832
169
- 4200 1943.349021911621
170
- 4225 1952.9981617927551
171
- 4250 1962.63041639328
172
- 4275 1972.275469303131
173
- 4300 1981.9206290245056
174
- 4325 1992.7996625900269
175
- 4350 2002.4559121131897
176
- 4375 2012.1016597747803
177
- 4400 2021.749123096466
178
- 4425 2031.40646982193
179
- 4450 2041.0575802326202
180
- 4475 2050.7022540569305
181
- 4500 2060.351977586746
182
- 4525 2069.997896671295
183
- 4550 2079.6516761779785
184
- 4575 2089.2996559143066
185
- 4600 2098.954682826996
186
- 4625 2108.6132814884186
187
- 4650 2118.261932849884
188
- 4675 2127.917363882065
189
- 4700 2137.573076248169
190
- 4725 2147.217692375183
191
- 4750 2156.8727049827576
192
- 4775 2166.5152747631073
193
- 4800 2176.1649520397186
194
- 4825 2185.8083612918854
195
- 4850 2195.4502625465393
196
- 4875 2205.109563589096
197
- 4900 2214.761839389801
198
- 4925 2224.407284975052
199
- 4950 2234.056025505066
200
- 4975 2243.701906681061
201
- 5000 2253.3530826568604
 
1
  step time
2
+ 25 11.024884700775146
3
+ 50 19.85481309890747
4
+ 75 28.683847665786743
5
+ 100 37.506720781326294
6
+ 125 46.30107140541077
7
+ 150 55.08619236946106
8
+ 175 63.87367296218872
9
+ 200 72.66220188140869
10
+ 225 81.45210266113281
11
+ 250 90.23731517791748
12
+ 275 99.02079439163208
13
+ 300 107.81453824043274
14
+ 325 116.60759544372559
15
+ 350 125.42598938941956
16
+ 375 134.24483466148376
17
+ 400 143.07194328308105
18
+ 425 151.8877546787262
19
+ 450 160.70777201652527
20
+ 475 169.5274999141693
21
+ 500 178.3484025001526
22
+ 525 187.16752195358276
23
+ 550 196.0018846988678
24
+ 575 204.82303142547607
25
+ 600 213.64335584640503
26
+ 625 222.46671533584595
27
+ 650 231.29865288734436
28
+ 675 240.13090705871582
29
+ 700 248.95567631721497
30
+ 725 257.780916929245
31
+ 750 266.6054456233978
32
+ 775 275.4310476779938
33
+ 800 284.2579245567322
34
+ 825 293.0841419696808
35
+ 850 301.9183979034424
36
+ 875 312.1040503978729
37
+ 900 320.93589448928833
38
+ 925 329.77263498306274
39
+ 950 338.6202914714813
40
+ 975 347.453581571579
41
+ 1000 356.29300141334534
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
generation_config.json CHANGED
@@ -1,36 +1,28 @@
1
  {
2
  "alignment_heads": [
3
  [
4
- 3,
5
- 1
6
- ],
7
- [
8
- 4,
9
  2
10
  ],
11
  [
12
- 4,
13
- 3
14
- ],
15
- [
16
- 4,
17
- 7
18
  ],
19
  [
20
- 5,
21
- 1
22
  ],
23
  [
24
- 5,
25
- 2
26
  ],
27
  [
28
- 5,
29
  4
30
  ],
31
  [
32
- 5,
33
- 6
34
  ]
35
  ],
36
  "begin_suppress_tokens": [
 
1
  {
2
  "alignment_heads": [
3
  [
4
+ 2,
 
 
 
 
5
  2
6
  ],
7
  [
8
+ 3,
9
+ 0
 
 
 
 
10
  ],
11
  [
12
+ 3,
13
+ 2
14
  ],
15
  [
16
+ 3,
17
+ 3
18
  ],
19
  [
20
+ 3,
21
  4
22
  ],
23
  [
24
+ 3,
25
+ 5
26
  ]
27
  ],
28
  "begin_suppress_tokens": [