Dataset Viewer
Auto-converted to Parquet
epoch
int64
1
300
seed
int64
0
9
training_loss
float64
0
1.72
method
stringclasses
8 values
validation_loss
float64
0.22
1.9
validation_accuracy
float64
0.36
0.94
1
0
1.060839
Adam Fixed
1.231932
0.553711
2
0
0.934124
Adam Fixed
1.025473
0.649414
3
0
0.962286
Adam Fixed
0.925444
0.675781
4
0
0.598814
Adam Fixed
0.800371
0.71582
5
0
0.963103
Adam Fixed
0.743018
0.732422
6
0
0.699911
Adam Fixed
0.615396
0.788086
7
0
0.557853
Adam Fixed
0.626414
0.783203
8
0
0.635469
Adam Fixed
0.556866
0.80957
9
0
0.521673
Adam Fixed
0.615788
0.806641
10
0
0.905352
Adam Fixed
0.629823
0.785156
11
0
0.730249
Adam Fixed
0.53874
0.826172
12
0
0.43294
Adam Fixed
0.543053
0.800781
13
0
0.433066
Adam Fixed
0.470882
0.830078
14
0
0.619368
Adam Fixed
0.528702
0.825195
15
0
0.361452
Adam Fixed
0.450491
0.852539
16
0
0.699469
Adam Fixed
0.431405
0.849609
17
0
0.367467
Adam Fixed
0.407496
0.861328
18
0
0.273346
Adam Fixed
0.412982
0.864258
19
0
0.405836
Adam Fixed
0.408571
0.868164
20
0
0.354008
Adam Fixed
0.426654
0.851563
21
0
0.364337
Adam Fixed
0.383216
0.873047
22
0
0.323138
Adam Fixed
0.457647
0.837891
23
0
0.324045
Adam Fixed
0.407975
0.862305
24
0
0.26485
Adam Fixed
0.409704
0.849609
25
0
0.313273
Adam Fixed
0.428886
0.852539
26
0
0.342745
Adam Fixed
0.347467
0.886719
27
0
0.298248
Adam Fixed
0.389165
0.873047
28
0
0.247408
Adam Fixed
0.409747
0.870117
29
0
0.275738
Adam Fixed
0.305378
0.904297
30
0
0.446122
Adam Fixed
0.393545
0.865234
31
0
0.283534
Adam Fixed
0.341506
0.887695
32
0
0.394179
Adam Fixed
0.387961
0.871094
33
0
0.391788
Adam Fixed
0.327261
0.887695
34
0
0.107159
Adam Fixed
0.329453
0.881836
35
0
0.368637
Adam Fixed
0.380992
0.874023
36
0
0.165132
Adam Fixed
0.306967
0.899414
37
0
0.149204
Adam Fixed
0.370087
0.883789
38
0
0.335835
Adam Fixed
0.338101
0.879883
39
0
0.307245
Adam Fixed
0.378345
0.878906
40
0
0.247295
Adam Fixed
0.283358
0.90625
41
0
0.218452
Adam Fixed
0.372131
0.875
42
0
0.244655
Adam Fixed
0.359819
0.890625
43
0
0.231642
Adam Fixed
0.420418
0.863281
44
0
0.221624
Adam Fixed
0.314134
0.895508
45
0
0.247116
Adam Fixed
0.368665
0.884766
46
0
0.277582
Adam Fixed
0.351798
0.876953
47
0
0.148128
Adam Fixed
0.284749
0.905273
48
0
0.19405
Adam Fixed
0.358768
0.886719
49
0
0.243741
Adam Fixed
0.33671
0.896484
50
0
0.175937
Adam Fixed
0.346661
0.893555
51
0
0.158626
Adam Fixed
0.331068
0.888672
52
0
0.205239
Adam Fixed
0.367347
0.876953
53
0
0.225249
Adam Fixed
0.352468
0.885742
54
0
0.304709
Adam Fixed
0.29149
0.900391
55
0
0.365029
Adam Fixed
0.322387
0.893555
56
0
0.282952
Adam Fixed
0.284809
0.905273
57
0
0.392364
Adam Fixed
0.313094
0.90332
58
0
0.306716
Adam Fixed
0.394194
0.870117
59
0
0.256588
Adam Fixed
0.337306
0.893555
60
0
0.256237
Adam Fixed
0.343428
0.886719
61
0
0.29227
Adam Fixed
0.283777
0.907227
62
0
0.187209
Adam Fixed
0.333399
0.898438
63
0
0.16652
Adam Fixed
0.32008
0.889648
64
0
0.295499
Adam Fixed
0.297316
0.911133
65
0
0.221606
Adam Fixed
0.350899
0.887695
66
0
0.319684
Adam Fixed
0.296899
0.90918
67
0
0.074719
Adam Fixed
0.309055
0.901367
68
0
0.265801
Adam Fixed
0.332461
0.894531
69
0
0.225861
Adam Fixed
0.263283
0.915039
70
0
0.206479
Adam Fixed
0.294499
0.897461
71
0
0.222238
Adam Fixed
0.292614
0.898438
72
0
0.276114
Adam Fixed
0.299806
0.90625
73
0
0.244449
Adam Fixed
0.354685
0.895508
74
0
0.111932
Adam Fixed
0.295025
0.902344
75
0
0.143299
Adam Fixed
0.304515
0.908203
76
0
0.235577
Adam Fixed
0.316015
0.904297
77
0
0.140315
Adam Fixed
0.298997
0.916016
78
0
0.141991
Adam Fixed
0.286924
0.90625
79
0
0.411342
Adam Fixed
0.350639
0.904297
80
0
0.140426
Adam Fixed
0.307172
0.911133
81
0
0.304566
Adam Fixed
0.368815
0.887695
82
0
0.358366
Adam Fixed
0.346962
0.887695
83
0
0.280611
Adam Fixed
0.29772
0.905273
84
0
0.172546
Adam Fixed
0.359301
0.894531
85
0
0.183693
Adam Fixed
0.283173
0.905273
86
0
0.031684
Adam Fixed
0.296431
0.90625
87
0
0.065116
Adam Fixed
0.309091
0.908203
88
0
0.260357
Adam Fixed
0.387571
0.889648
89
0
0.107506
Adam Fixed
0.311045
0.916016
90
0
0.082933
Adam Fixed
0.342492
0.897461
91
0
0.070986
Adam Fixed
0.297057
0.917969
92
0
0.182342
Adam Fixed
0.310448
0.907227
93
0
0.042091
Adam Fixed
0.335996
0.901367
94
0
0.177517
Adam Fixed
0.288132
0.915039
95
0
0.065147
Adam Fixed
0.284933
0.913086
96
0
0.118742
Adam Fixed
0.294809
0.914063
97
0
0.163111
Adam Fixed
0.421157
0.881836
98
0
0.058927
Adam Fixed
0.345744
0.895508
99
0
0.072214
Adam Fixed
0.370492
0.896484
100
0
0.19312
Adam Fixed
0.352235
0.885742
End of preview. Expand in Data Studio

Revisiting Learning Rate Control

This dataset represents the experimental data collected from the paper Revisiting Learning Rate Control. We provide splits for computer_vision, libsvm and roberta experiments. The details for each split are as follows:

computer_vision

epoch, seed, methodtraining_loss, validation_loss, validation_accuracy \text{epoch, seed, method} \mapsto \text{training\_loss, validation\_loss, validation\_accuracy}

libsvm

epoch, seed, method, datasettraining_loss, validation_loss, validation_accuracy\text{epoch, seed, method, dataset} \mapsto \text{training\_loss, validation\_loss, validation\_accuracy}

roberta

step, seed, methodtraining_perplexity, validation_perplexity\text{step, seed, method} \mapsto \text{training\_perplexity, validation\_perplexity}

Downloads last month
0