|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9938347718865597, |
|
"eval_steps": 500, |
|
"global_step": 606, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.004932182490752158, |
|
"grad_norm": 1.1257907152175903, |
|
"learning_rate": 5.0000000000000004e-08, |
|
"loss": 1.0986, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.009864364981504316, |
|
"grad_norm": 1.5788811445236206, |
|
"learning_rate": 1.0000000000000001e-07, |
|
"loss": 1.1152, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.014796547472256474, |
|
"grad_norm": 1.7341364622116089, |
|
"learning_rate": 1.5000000000000002e-07, |
|
"loss": 1.1392, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01972872996300863, |
|
"grad_norm": 1.3568042516708374, |
|
"learning_rate": 2.0000000000000002e-07, |
|
"loss": 1.1046, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.02466091245376079, |
|
"grad_norm": 1.1873835325241089, |
|
"learning_rate": 2.5000000000000004e-07, |
|
"loss": 1.1123, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.029593094944512947, |
|
"grad_norm": 1.0728861093521118, |
|
"learning_rate": 3.0000000000000004e-07, |
|
"loss": 1.1024, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0345252774352651, |
|
"grad_norm": 1.991363525390625, |
|
"learning_rate": 3.5000000000000004e-07, |
|
"loss": 1.1045, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.03945745992601726, |
|
"grad_norm": 1.3626810312271118, |
|
"learning_rate": 4.0000000000000003e-07, |
|
"loss": 1.1414, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.04438964241676942, |
|
"grad_norm": 1.2338333129882812, |
|
"learning_rate": 4.5000000000000003e-07, |
|
"loss": 1.1312, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.04932182490752158, |
|
"grad_norm": 1.2466716766357422, |
|
"learning_rate": 5.000000000000001e-07, |
|
"loss": 1.0895, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.05425400739827373, |
|
"grad_norm": 1.360613465309143, |
|
"learning_rate": 5.5e-07, |
|
"loss": 1.1285, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.059186189889025895, |
|
"grad_norm": 2.145904541015625, |
|
"learning_rate": 6.000000000000001e-07, |
|
"loss": 1.079, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.06411837237977805, |
|
"grad_norm": 1.5354256629943848, |
|
"learning_rate": 6.5e-07, |
|
"loss": 1.0897, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0690505548705302, |
|
"grad_norm": 1.2002555131912231, |
|
"learning_rate": 7.000000000000001e-07, |
|
"loss": 1.0474, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.07398273736128237, |
|
"grad_norm": 1.213218331336975, |
|
"learning_rate": 7.5e-07, |
|
"loss": 1.0336, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.07891491985203453, |
|
"grad_norm": 1.0160884857177734, |
|
"learning_rate": 8.000000000000001e-07, |
|
"loss": 1.0396, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.08384710234278668, |
|
"grad_norm": 1.3291572332382202, |
|
"learning_rate": 8.500000000000001e-07, |
|
"loss": 1.0111, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.08877928483353884, |
|
"grad_norm": 9.560614585876465, |
|
"learning_rate": 9.000000000000001e-07, |
|
"loss": 1.0343, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.093711467324291, |
|
"grad_norm": 0.8536161184310913, |
|
"learning_rate": 9.500000000000001e-07, |
|
"loss": 1.0288, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.09864364981504316, |
|
"grad_norm": 0.8779953122138977, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 0.9296, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.10357583230579531, |
|
"grad_norm": 0.9102424383163452, |
|
"learning_rate": 1.0500000000000001e-06, |
|
"loss": 0.9348, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.10850801479654747, |
|
"grad_norm": 0.5936668515205383, |
|
"learning_rate": 1.1e-06, |
|
"loss": 0.9103, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.11344019728729964, |
|
"grad_norm": 0.38109758496284485, |
|
"learning_rate": 1.1500000000000002e-06, |
|
"loss": 0.9412, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.11837237977805179, |
|
"grad_norm": 0.6523936986923218, |
|
"learning_rate": 1.2000000000000002e-06, |
|
"loss": 0.8968, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.12330456226880394, |
|
"grad_norm": 0.4067033529281616, |
|
"learning_rate": 1.25e-06, |
|
"loss": 0.9031, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.1282367447595561, |
|
"grad_norm": 0.4689585566520691, |
|
"learning_rate": 1.3e-06, |
|
"loss": 0.8724, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.13316892725030827, |
|
"grad_norm": 0.5714285373687744, |
|
"learning_rate": 1.3500000000000002e-06, |
|
"loss": 0.9195, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.1381011097410604, |
|
"grad_norm": 0.43251562118530273, |
|
"learning_rate": 1.4000000000000001e-06, |
|
"loss": 0.8937, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.14303329223181258, |
|
"grad_norm": 0.3626736104488373, |
|
"learning_rate": 1.45e-06, |
|
"loss": 0.8685, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.14796547472256474, |
|
"grad_norm": 0.3419496715068817, |
|
"learning_rate": 1.5e-06, |
|
"loss": 0.8833, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.15289765721331688, |
|
"grad_norm": 0.4578935503959656, |
|
"learning_rate": 1.5500000000000002e-06, |
|
"loss": 0.8287, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.15782983970406905, |
|
"grad_norm": 0.4360993206501007, |
|
"learning_rate": 1.6000000000000001e-06, |
|
"loss": 0.8452, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.16276202219482122, |
|
"grad_norm": 0.5611799359321594, |
|
"learning_rate": 1.6500000000000003e-06, |
|
"loss": 0.872, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.16769420468557336, |
|
"grad_norm": 0.2829279601573944, |
|
"learning_rate": 1.7000000000000002e-06, |
|
"loss": 0.87, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.17262638717632553, |
|
"grad_norm": 0.38662174344062805, |
|
"learning_rate": 1.75e-06, |
|
"loss": 0.816, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.17755856966707767, |
|
"grad_norm": 0.4239096939563751, |
|
"learning_rate": 1.8000000000000001e-06, |
|
"loss": 0.8599, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.18249075215782984, |
|
"grad_norm": 0.653641939163208, |
|
"learning_rate": 1.85e-06, |
|
"loss": 0.8456, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.187422934648582, |
|
"grad_norm": 0.21455056965351105, |
|
"learning_rate": 1.9000000000000002e-06, |
|
"loss": 0.795, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.19235511713933415, |
|
"grad_norm": 0.26148131489753723, |
|
"learning_rate": 1.9500000000000004e-06, |
|
"loss": 0.8405, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.19728729963008632, |
|
"grad_norm": 0.31602978706359863, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 0.8569, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.20221948212083848, |
|
"grad_norm": 0.20896084606647491, |
|
"learning_rate": 2.05e-06, |
|
"loss": 0.801, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.20715166461159062, |
|
"grad_norm": 0.21203891932964325, |
|
"learning_rate": 2.1000000000000002e-06, |
|
"loss": 0.8202, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.2120838471023428, |
|
"grad_norm": 0.23695039749145508, |
|
"learning_rate": 2.15e-06, |
|
"loss": 0.8088, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.21701602959309493, |
|
"grad_norm": 0.1736220270395279, |
|
"learning_rate": 2.2e-06, |
|
"loss": 0.8031, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.2219482120838471, |
|
"grad_norm": 0.23390308022499084, |
|
"learning_rate": 2.25e-06, |
|
"loss": 0.7914, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.22688039457459927, |
|
"grad_norm": 0.23420561850070953, |
|
"learning_rate": 2.3000000000000004e-06, |
|
"loss": 0.8092, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.2318125770653514, |
|
"grad_norm": 0.19289737939834595, |
|
"learning_rate": 2.35e-06, |
|
"loss": 0.7767, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.23674475955610358, |
|
"grad_norm": 0.23044224083423615, |
|
"learning_rate": 2.4000000000000003e-06, |
|
"loss": 0.8185, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.24167694204685575, |
|
"grad_norm": 0.2052794247865677, |
|
"learning_rate": 2.4500000000000003e-06, |
|
"loss": 0.781, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.2466091245376079, |
|
"grad_norm": 0.19429263472557068, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.8125, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.25154130702836003, |
|
"grad_norm": 0.18123206496238708, |
|
"learning_rate": 2.55e-06, |
|
"loss": 0.7775, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.2564734895191122, |
|
"grad_norm": 1.8012878894805908, |
|
"learning_rate": 2.6e-06, |
|
"loss": 0.7692, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.26140567200986436, |
|
"grad_norm": 0.2761130928993225, |
|
"learning_rate": 2.6500000000000005e-06, |
|
"loss": 0.7996, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.26633785450061653, |
|
"grad_norm": 0.1579173356294632, |
|
"learning_rate": 2.7000000000000004e-06, |
|
"loss": 0.7936, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.2712700369913687, |
|
"grad_norm": 0.16990424692630768, |
|
"learning_rate": 2.7500000000000004e-06, |
|
"loss": 0.8067, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.2762022194821208, |
|
"grad_norm": 1.0547306537628174, |
|
"learning_rate": 2.8000000000000003e-06, |
|
"loss": 0.7894, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.281134401972873, |
|
"grad_norm": 0.18402548134326935, |
|
"learning_rate": 2.85e-06, |
|
"loss": 0.7564, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.28606658446362515, |
|
"grad_norm": 0.19761355221271515, |
|
"learning_rate": 2.9e-06, |
|
"loss": 0.7754, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.2909987669543773, |
|
"grad_norm": 0.26926949620246887, |
|
"learning_rate": 2.95e-06, |
|
"loss": 0.7698, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.2959309494451295, |
|
"grad_norm": 0.19200852513313293, |
|
"learning_rate": 3e-06, |
|
"loss": 0.757, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.3008631319358816, |
|
"grad_norm": 0.14297647774219513, |
|
"learning_rate": 3.05e-06, |
|
"loss": 0.7372, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.30579531442663377, |
|
"grad_norm": 0.1689106523990631, |
|
"learning_rate": 3.1000000000000004e-06, |
|
"loss": 0.7511, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.31072749691738594, |
|
"grad_norm": 0.3330126106739044, |
|
"learning_rate": 3.1500000000000003e-06, |
|
"loss": 0.747, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.3156596794081381, |
|
"grad_norm": 0.2532431483268738, |
|
"learning_rate": 3.2000000000000003e-06, |
|
"loss": 0.798, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.3205918618988903, |
|
"grad_norm": 0.14446307718753815, |
|
"learning_rate": 3.2500000000000002e-06, |
|
"loss": 0.7631, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.32552404438964244, |
|
"grad_norm": 0.24439038336277008, |
|
"learning_rate": 3.3000000000000006e-06, |
|
"loss": 0.7405, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.33045622688039455, |
|
"grad_norm": 0.15191350877285004, |
|
"learning_rate": 3.3500000000000005e-06, |
|
"loss": 0.7559, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.3353884093711467, |
|
"grad_norm": 0.16212493181228638, |
|
"learning_rate": 3.4000000000000005e-06, |
|
"loss": 0.7452, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.3403205918618989, |
|
"grad_norm": 0.1210586428642273, |
|
"learning_rate": 3.45e-06, |
|
"loss": 0.7817, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.34525277435265106, |
|
"grad_norm": 0.35743728280067444, |
|
"learning_rate": 3.5e-06, |
|
"loss": 0.7811, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.35018495684340323, |
|
"grad_norm": 0.12270136177539825, |
|
"learning_rate": 3.5500000000000003e-06, |
|
"loss": 0.7679, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.35511713933415534, |
|
"grad_norm": 0.14078450202941895, |
|
"learning_rate": 3.6000000000000003e-06, |
|
"loss": 0.7479, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.3600493218249075, |
|
"grad_norm": 0.11881807446479797, |
|
"learning_rate": 3.65e-06, |
|
"loss": 0.7475, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.3649815043156597, |
|
"grad_norm": 0.11201618611812592, |
|
"learning_rate": 3.7e-06, |
|
"loss": 0.7396, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.36991368680641185, |
|
"grad_norm": 0.14292244613170624, |
|
"learning_rate": 3.7500000000000005e-06, |
|
"loss": 0.7117, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.374845869297164, |
|
"grad_norm": 0.18425297737121582, |
|
"learning_rate": 3.8000000000000005e-06, |
|
"loss": 0.718, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.3797780517879161, |
|
"grad_norm": 0.16399289667606354, |
|
"learning_rate": 3.85e-06, |
|
"loss": 0.7314, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.3847102342786683, |
|
"grad_norm": 0.1419249176979065, |
|
"learning_rate": 3.900000000000001e-06, |
|
"loss": 0.7996, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.38964241676942046, |
|
"grad_norm": 0.154827281832695, |
|
"learning_rate": 3.95e-06, |
|
"loss": 0.7308, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.39457459926017263, |
|
"grad_norm": 0.17852246761322021, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.7161, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.3995067817509248, |
|
"grad_norm": 0.12098474055528641, |
|
"learning_rate": 4.05e-06, |
|
"loss": 0.7581, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.40443896424167697, |
|
"grad_norm": 0.15087738633155823, |
|
"learning_rate": 4.1e-06, |
|
"loss": 0.728, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.4093711467324291, |
|
"grad_norm": 0.23446552455425262, |
|
"learning_rate": 4.15e-06, |
|
"loss": 0.7089, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.41430332922318125, |
|
"grad_norm": 0.12758222222328186, |
|
"learning_rate": 4.2000000000000004e-06, |
|
"loss": 0.7513, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.4192355117139334, |
|
"grad_norm": 0.18490324914455414, |
|
"learning_rate": 4.25e-06, |
|
"loss": 0.7232, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.4241676942046856, |
|
"grad_norm": 3.9814538955688477, |
|
"learning_rate": 4.3e-06, |
|
"loss": 0.7071, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.42909987669543775, |
|
"grad_norm": 0.13754811882972717, |
|
"learning_rate": 4.350000000000001e-06, |
|
"loss": 0.7218, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.43403205918618987, |
|
"grad_norm": 0.15279600024223328, |
|
"learning_rate": 4.4e-06, |
|
"loss": 0.7531, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.43896424167694204, |
|
"grad_norm": 0.13091601431369781, |
|
"learning_rate": 4.450000000000001e-06, |
|
"loss": 0.6954, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.4438964241676942, |
|
"grad_norm": 0.14220909774303436, |
|
"learning_rate": 4.5e-06, |
|
"loss": 0.7227, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.44882860665844637, |
|
"grad_norm": 0.11386663466691971, |
|
"learning_rate": 4.5500000000000005e-06, |
|
"loss": 0.7196, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.45376078914919854, |
|
"grad_norm": 0.16582414507865906, |
|
"learning_rate": 4.600000000000001e-06, |
|
"loss": 0.7155, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.45869297163995065, |
|
"grad_norm": 0.2046297937631607, |
|
"learning_rate": 4.65e-06, |
|
"loss": 0.7201, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.4636251541307028, |
|
"grad_norm": 0.11329102516174316, |
|
"learning_rate": 4.7e-06, |
|
"loss": 0.7233, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.468557336621455, |
|
"grad_norm": 0.1256910264492035, |
|
"learning_rate": 4.75e-06, |
|
"loss": 0.7294, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.47348951911220716, |
|
"grad_norm": 0.16878068447113037, |
|
"learning_rate": 4.800000000000001e-06, |
|
"loss": 0.7292, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.4784217016029593, |
|
"grad_norm": 0.1140855923295021, |
|
"learning_rate": 4.85e-06, |
|
"loss": 0.7215, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.4833538840937115, |
|
"grad_norm": 0.13400927186012268, |
|
"learning_rate": 4.9000000000000005e-06, |
|
"loss": 0.7446, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.4882860665844636, |
|
"grad_norm": 0.13537898659706116, |
|
"learning_rate": 4.95e-06, |
|
"loss": 0.7083, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.4932182490752158, |
|
"grad_norm": 0.23342657089233398, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7172, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.49815043156596794, |
|
"grad_norm": 2.2964117527008057, |
|
"learning_rate": 4.999951815503011e-06, |
|
"loss": 0.7564, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.5030826140567201, |
|
"grad_norm": 0.20878185331821442, |
|
"learning_rate": 4.999807263869441e-06, |
|
"loss": 0.7255, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.5080147965474723, |
|
"grad_norm": 0.26244527101516724, |
|
"learning_rate": 4.999566350671405e-06, |
|
"loss": 0.707, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.5129469790382244, |
|
"grad_norm": 0.12257402390241623, |
|
"learning_rate": 4.999229085195532e-06, |
|
"loss": 0.7147, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.5178791615289766, |
|
"grad_norm": 0.18218959867954254, |
|
"learning_rate": 4.998795480442595e-06, |
|
"loss": 0.7113, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.5228113440197287, |
|
"grad_norm": 0.1487857848405838, |
|
"learning_rate": 4.998265553127013e-06, |
|
"loss": 0.7532, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.5277435265104808, |
|
"grad_norm": 0.18883858621120453, |
|
"learning_rate": 4.997639323676214e-06, |
|
"loss": 0.6822, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.5326757090012331, |
|
"grad_norm": 0.13591569662094116, |
|
"learning_rate": 4.996916816229838e-06, |
|
"loss": 0.7569, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.5376078914919852, |
|
"grad_norm": 0.9631307721138, |
|
"learning_rate": 4.99609805863881e-06, |
|
"loss": 0.745, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.5425400739827374, |
|
"grad_norm": 0.13628794252872467, |
|
"learning_rate": 4.995183082464269e-06, |
|
"loss": 0.732, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.5474722564734895, |
|
"grad_norm": 2.4701926708221436, |
|
"learning_rate": 4.994171922976349e-06, |
|
"loss": 0.756, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.5524044389642416, |
|
"grad_norm": 1.1472599506378174, |
|
"learning_rate": 4.993064619152818e-06, |
|
"loss": 0.7427, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.5573366214549939, |
|
"grad_norm": 0.11852557212114334, |
|
"learning_rate": 4.991861213677578e-06, |
|
"loss": 0.7124, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.562268803945746, |
|
"grad_norm": 0.09930302202701569, |
|
"learning_rate": 4.99056175293902e-06, |
|
"loss": 0.7065, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.5672009864364982, |
|
"grad_norm": 0.12781856954097748, |
|
"learning_rate": 4.989166287028234e-06, |
|
"loss": 0.7298, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.5721331689272503, |
|
"grad_norm": 0.39002031087875366, |
|
"learning_rate": 4.987674869737078e-06, |
|
"loss": 0.6996, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.5770653514180024, |
|
"grad_norm": 0.1187821626663208, |
|
"learning_rate": 4.986087558556104e-06, |
|
"loss": 0.7333, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.5819975339087546, |
|
"grad_norm": 0.11041084676980972, |
|
"learning_rate": 4.984404414672346e-06, |
|
"loss": 0.7101, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.5869297163995068, |
|
"grad_norm": 0.1762731671333313, |
|
"learning_rate": 4.9826255029669575e-06, |
|
"loss": 0.7063, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.591861898890259, |
|
"grad_norm": 0.12481699883937836, |
|
"learning_rate": 4.980750892012711e-06, |
|
"loss": 0.725, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.5967940813810111, |
|
"grad_norm": 0.7090242505073547, |
|
"learning_rate": 4.978780654071355e-06, |
|
"loss": 0.7105, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.6017262638717632, |
|
"grad_norm": 0.11730094999074936, |
|
"learning_rate": 4.976714865090827e-06, |
|
"loss": 0.7091, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.6066584463625154, |
|
"grad_norm": 0.12327645719051361, |
|
"learning_rate": 4.974553604702332e-06, |
|
"loss": 0.739, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.6115906288532675, |
|
"grad_norm": 0.22779831290245056, |
|
"learning_rate": 4.972296956217265e-06, |
|
"loss": 0.6924, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.6165228113440198, |
|
"grad_norm": 0.3117700517177582, |
|
"learning_rate": 4.969945006624003e-06, |
|
"loss": 0.7166, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.6214549938347719, |
|
"grad_norm": 0.1463162750005722, |
|
"learning_rate": 4.967497846584552e-06, |
|
"loss": 0.7236, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.6263871763255241, |
|
"grad_norm": 0.11583642661571503, |
|
"learning_rate": 4.9649555704310545e-06, |
|
"loss": 0.6957, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.6313193588162762, |
|
"grad_norm": 0.1344325989484787, |
|
"learning_rate": 4.962318276162148e-06, |
|
"loss": 0.7031, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.6362515413070283, |
|
"grad_norm": 0.13892586529254913, |
|
"learning_rate": 4.959586065439189e-06, |
|
"loss": 0.6925, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.6411837237977805, |
|
"grad_norm": 0.19953861832618713, |
|
"learning_rate": 4.956759043582339e-06, |
|
"loss": 0.72, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.6461159062885327, |
|
"grad_norm": 0.2914772629737854, |
|
"learning_rate": 4.953837319566498e-06, |
|
"loss": 0.7175, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.6510480887792849, |
|
"grad_norm": 0.109347403049469, |
|
"learning_rate": 4.950821006017107e-06, |
|
"loss": 0.6923, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.655980271270037, |
|
"grad_norm": 0.2301577925682068, |
|
"learning_rate": 4.947710219205808e-06, |
|
"loss": 0.7093, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.6609124537607891, |
|
"grad_norm": 0.1368824988603592, |
|
"learning_rate": 4.9445050790459585e-06, |
|
"loss": 0.6746, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.6658446362515413, |
|
"grad_norm": 0.13323882222175598, |
|
"learning_rate": 4.9412057090880115e-06, |
|
"loss": 0.6919, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.6707768187422934, |
|
"grad_norm": 0.14521068334579468, |
|
"learning_rate": 4.937812236514754e-06, |
|
"loss": 0.7352, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.6757090012330457, |
|
"grad_norm": 0.11869832128286362, |
|
"learning_rate": 4.9343247921364e-06, |
|
"loss": 0.6759, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.6806411837237978, |
|
"grad_norm": 0.20703694224357605, |
|
"learning_rate": 4.930743510385551e-06, |
|
"loss": 0.6984, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.6855733662145499, |
|
"grad_norm": 0.13110846281051636, |
|
"learning_rate": 4.927068529312017e-06, |
|
"loss": 0.7143, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.6905055487053021, |
|
"grad_norm": 0.13985510170459747, |
|
"learning_rate": 4.923299990577488e-06, |
|
"loss": 0.6899, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.6954377311960542, |
|
"grad_norm": 0.1306031197309494, |
|
"learning_rate": 4.919438039450079e-06, |
|
"loss": 0.6801, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.7003699136868065, |
|
"grad_norm": 0.10599593818187714, |
|
"learning_rate": 4.915482824798728e-06, |
|
"loss": 0.7191, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.7053020961775586, |
|
"grad_norm": 0.1460665911436081, |
|
"learning_rate": 4.911434499087457e-06, |
|
"loss": 0.7134, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.7102342786683107, |
|
"grad_norm": 0.1587439626455307, |
|
"learning_rate": 4.907293218369499e-06, |
|
"loss": 0.6987, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.7151664611590629, |
|
"grad_norm": 0.22658978402614594, |
|
"learning_rate": 4.903059142281273e-06, |
|
"loss": 0.7049, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.720098643649815, |
|
"grad_norm": 0.6049544215202332, |
|
"learning_rate": 4.8987324340362445e-06, |
|
"loss": 0.709, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.7250308261405672, |
|
"grad_norm": 0.1274397075176239, |
|
"learning_rate": 4.894313260418617e-06, |
|
"loss": 0.6762, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.7299630086313194, |
|
"grad_norm": 0.1254238486289978, |
|
"learning_rate": 4.889801791776921e-06, |
|
"loss": 0.6598, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.7348951911220715, |
|
"grad_norm": 0.12932169437408447, |
|
"learning_rate": 4.885198202017431e-06, |
|
"loss": 0.6905, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.7398273736128237, |
|
"grad_norm": 0.1857711523771286, |
|
"learning_rate": 4.880502668597475e-06, |
|
"loss": 0.7186, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.7447595561035758, |
|
"grad_norm": 0.11666952073574066, |
|
"learning_rate": 4.875715372518585e-06, |
|
"loss": 0.6971, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.749691738594328, |
|
"grad_norm": 0.11182554066181183, |
|
"learning_rate": 4.870836498319523e-06, |
|
"loss": 0.6876, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.7546239210850801, |
|
"grad_norm": 0.11282117664813995, |
|
"learning_rate": 4.865866234069169e-06, |
|
"loss": 0.7039, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.7595561035758323, |
|
"grad_norm": 0.11031971871852875, |
|
"learning_rate": 4.86080477135927e-06, |
|
"loss": 0.6703, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.7644882860665845, |
|
"grad_norm": 0.1600004881620407, |
|
"learning_rate": 4.855652305297052e-06, |
|
"loss": 0.6776, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.7694204685573366, |
|
"grad_norm": 0.13014693558216095, |
|
"learning_rate": 4.8504090344977036e-06, |
|
"loss": 0.6926, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.7743526510480888, |
|
"grad_norm": 0.17829261720180511, |
|
"learning_rate": 4.84507516107672e-06, |
|
"loss": 0.7036, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.7792848335388409, |
|
"grad_norm": 0.13715577125549316, |
|
"learning_rate": 4.839650890642104e-06, |
|
"loss": 0.6788, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.7842170160295932, |
|
"grad_norm": 0.16660194098949432, |
|
"learning_rate": 4.834136432286452e-06, |
|
"loss": 0.7052, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.7891491985203453, |
|
"grad_norm": 0.10915841907262802, |
|
"learning_rate": 4.828531998578885e-06, |
|
"loss": 0.6484, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.7940813810110974, |
|
"grad_norm": 0.11565055698156357, |
|
"learning_rate": 4.822837805556858e-06, |
|
"loss": 0.6853, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.7990135635018496, |
|
"grad_norm": 0.11839542537927628, |
|
"learning_rate": 4.817054072717833e-06, |
|
"loss": 0.6647, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.8039457459926017, |
|
"grad_norm": 0.11023162305355072, |
|
"learning_rate": 4.811181023010815e-06, |
|
"loss": 0.7049, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.8088779284833539, |
|
"grad_norm": 0.21431177854537964, |
|
"learning_rate": 4.805218882827761e-06, |
|
"loss": 0.6942, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.813810110974106, |
|
"grad_norm": 0.11522458493709564, |
|
"learning_rate": 4.799167881994852e-06, |
|
"loss": 0.6929, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.8187422934648582, |
|
"grad_norm": 0.12389165163040161, |
|
"learning_rate": 4.793028253763633e-06, |
|
"loss": 0.6994, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.8236744759556104, |
|
"grad_norm": 0.11536920070648193, |
|
"learning_rate": 4.786800234802022e-06, |
|
"loss": 0.6959, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.8286066584463625, |
|
"grad_norm": 0.1272544115781784, |
|
"learning_rate": 4.780484065185188e-06, |
|
"loss": 0.6626, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.8335388409371147, |
|
"grad_norm": 0.16449759900569916, |
|
"learning_rate": 4.7740799883862966e-06, |
|
"loss": 0.6925, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.8384710234278668, |
|
"grad_norm": 0.2522273361682892, |
|
"learning_rate": 4.767588251267121e-06, |
|
"loss": 0.7052, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.843403205918619, |
|
"grad_norm": 0.09476125985383987, |
|
"learning_rate": 4.761009104068533e-06, |
|
"loss": 0.6722, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.8483353884093712, |
|
"grad_norm": 0.23170353472232819, |
|
"learning_rate": 4.754342800400852e-06, |
|
"loss": 0.6794, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.8532675709001233, |
|
"grad_norm": 0.14337614178657532, |
|
"learning_rate": 4.747589597234068e-06, |
|
"loss": 0.6564, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.8581997533908755, |
|
"grad_norm": 0.17019444704055786, |
|
"learning_rate": 4.740749754887939e-06, |
|
"loss": 0.6688, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.8631319358816276, |
|
"grad_norm": 0.1297682523727417, |
|
"learning_rate": 4.7338235370219556e-06, |
|
"loss": 0.7185, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.8680641183723797, |
|
"grad_norm": 0.17522914707660675, |
|
"learning_rate": 4.726811210625176e-06, |
|
"loss": 0.7061, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.872996300863132, |
|
"grad_norm": 0.3105694651603699, |
|
"learning_rate": 4.7197130460059385e-06, |
|
"loss": 0.688, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.8779284833538841, |
|
"grad_norm": 0.10563701391220093, |
|
"learning_rate": 4.712529316781435e-06, |
|
"loss": 0.6695, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.8828606658446363, |
|
"grad_norm": 0.16776345670223236, |
|
"learning_rate": 4.705260299867169e-06, |
|
"loss": 0.6855, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.8877928483353884, |
|
"grad_norm": 0.10242436826229095, |
|
"learning_rate": 4.697906275466279e-06, |
|
"loss": 0.6768, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.8927250308261405, |
|
"grad_norm": 0.1279619336128235, |
|
"learning_rate": 4.69046752705874e-06, |
|
"loss": 0.6854, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.8976572133168927, |
|
"grad_norm": 0.190452441573143, |
|
"learning_rate": 4.682944341390431e-06, |
|
"loss": 0.7199, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.9025893958076449, |
|
"grad_norm": 0.2059255987405777, |
|
"learning_rate": 4.675337008462085e-06, |
|
"loss": 0.6694, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.9075215782983971, |
|
"grad_norm": 0.10989531874656677, |
|
"learning_rate": 4.667645821518111e-06, |
|
"loss": 0.6707, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.9124537607891492, |
|
"grad_norm": 0.11381805688142776, |
|
"learning_rate": 4.659871077035289e-06, |
|
"loss": 0.6617, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.9173859432799013, |
|
"grad_norm": 0.18315713107585907, |
|
"learning_rate": 4.65201307471134e-06, |
|
"loss": 0.6721, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.9223181257706535, |
|
"grad_norm": 0.13762731850147247, |
|
"learning_rate": 4.644072117453377e-06, |
|
"loss": 0.6691, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.9272503082614056, |
|
"grad_norm": 0.13186247646808624, |
|
"learning_rate": 4.636048511366222e-06, |
|
"loss": 0.6907, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.9321824907521579, |
|
"grad_norm": 0.23819328844547272, |
|
"learning_rate": 4.6279425657406154e-06, |
|
"loss": 0.7147, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.93711467324291, |
|
"grad_norm": 0.3640376329421997, |
|
"learning_rate": 4.619754593041287e-06, |
|
"loss": 0.6981, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.9420468557336621, |
|
"grad_norm": 0.19634462893009186, |
|
"learning_rate": 4.6114849088949146e-06, |
|
"loss": 0.6704, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.9469790382244143, |
|
"grad_norm": 0.11775479465723038, |
|
"learning_rate": 4.603133832077953e-06, |
|
"loss": 0.6909, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.9519112207151664, |
|
"grad_norm": 0.10143059492111206, |
|
"learning_rate": 4.594701684504352e-06, |
|
"loss": 0.6625, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.9568434032059187, |
|
"grad_norm": 0.11726385354995728, |
|
"learning_rate": 4.586188791213143e-06, |
|
"loss": 0.6692, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.9617755856966708, |
|
"grad_norm": 0.11105194687843323, |
|
"learning_rate": 4.577595480355911e-06, |
|
"loss": 0.6963, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.966707768187423, |
|
"grad_norm": 0.10140436887741089, |
|
"learning_rate": 4.568922083184144e-06, |
|
"loss": 0.6969, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.9716399506781751, |
|
"grad_norm": 0.2534874677658081, |
|
"learning_rate": 4.560168934036467e-06, |
|
"loss": 0.6852, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.9765721331689272, |
|
"grad_norm": 0.10877111554145813, |
|
"learning_rate": 4.55133637032575e-06, |
|
"loss": 0.6998, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.9815043156596794, |
|
"grad_norm": 0.14067795872688293, |
|
"learning_rate": 4.542424732526105e-06, |
|
"loss": 0.6685, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.9864364981504316, |
|
"grad_norm": 0.7147830128669739, |
|
"learning_rate": 4.533434364159761e-06, |
|
"loss": 0.6982, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.9913686806411838, |
|
"grad_norm": 0.15026246011257172, |
|
"learning_rate": 4.524365611783818e-06, |
|
"loss": 0.6915, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.9963008631319359, |
|
"grad_norm": 0.12991268932819366, |
|
"learning_rate": 4.515218824976895e-06, |
|
"loss": 0.6539, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 1.0036991368680641, |
|
"grad_norm": 0.11360019445419312, |
|
"learning_rate": 4.505994356325648e-06, |
|
"loss": 0.6479, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 1.0086313193588163, |
|
"grad_norm": 0.1208319365978241, |
|
"learning_rate": 4.496692561411182e-06, |
|
"loss": 0.6704, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 1.0135635018495683, |
|
"grad_norm": 0.15796445310115814, |
|
"learning_rate": 4.487313798795347e-06, |
|
"loss": 0.6585, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 1.0184956843403206, |
|
"grad_norm": 0.2212684154510498, |
|
"learning_rate": 4.477858430006906e-06, |
|
"loss": 0.6748, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 1.0234278668310728, |
|
"grad_norm": 0.14605554938316345, |
|
"learning_rate": 4.468326819527613e-06, |
|
"loss": 0.6671, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 1.028360049321825, |
|
"grad_norm": 0.11415625363588333, |
|
"learning_rate": 4.458719334778153e-06, |
|
"loss": 0.681, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 1.033292231812577, |
|
"grad_norm": 0.113845095038414, |
|
"learning_rate": 4.449036346103982e-06, |
|
"loss": 0.6803, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 1.0382244143033292, |
|
"grad_norm": 0.10725957155227661, |
|
"learning_rate": 4.43927822676105e-06, |
|
"loss": 0.648, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.0431565967940815, |
|
"grad_norm": 0.12335722148418427, |
|
"learning_rate": 4.429445352901415e-06, |
|
"loss": 0.695, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 1.0480887792848335, |
|
"grad_norm": 0.10324176400899887, |
|
"learning_rate": 4.419538103558742e-06, |
|
"loss": 0.6529, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 1.0530209617755857, |
|
"grad_norm": 0.10406983643770218, |
|
"learning_rate": 4.409556860633692e-06, |
|
"loss": 0.6599, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 1.057953144266338, |
|
"grad_norm": 0.11518678814172745, |
|
"learning_rate": 4.3995020088792e-06, |
|
"loss": 0.6601, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 1.06288532675709, |
|
"grad_norm": 0.14593787491321564, |
|
"learning_rate": 4.3893739358856465e-06, |
|
"loss": 0.6714, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 1.0678175092478421, |
|
"grad_norm": 0.17964190244674683, |
|
"learning_rate": 4.379173032065912e-06, |
|
"loss": 0.6441, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 1.0727496917385944, |
|
"grad_norm": 0.12014143913984299, |
|
"learning_rate": 4.368899690640333e-06, |
|
"loss": 0.6526, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 1.0776818742293466, |
|
"grad_norm": 0.2840045690536499, |
|
"learning_rate": 4.3585543076215405e-06, |
|
"loss": 0.6688, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 1.0826140567200986, |
|
"grad_norm": 0.23248234391212463, |
|
"learning_rate": 4.3481372817991976e-06, |
|
"loss": 0.6798, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 1.0875462392108508, |
|
"grad_norm": 0.11116907000541687, |
|
"learning_rate": 4.3376490147246205e-06, |
|
"loss": 0.6676, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.092478421701603, |
|
"grad_norm": 0.10024593770503998, |
|
"learning_rate": 4.32708991069531e-06, |
|
"loss": 0.6718, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 1.097410604192355, |
|
"grad_norm": 0.11960247159004211, |
|
"learning_rate": 4.31646037673936e-06, |
|
"loss": 0.665, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 1.1023427866831073, |
|
"grad_norm": 0.2154683619737625, |
|
"learning_rate": 4.305760822599766e-06, |
|
"loss": 0.6782, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 1.1072749691738595, |
|
"grad_norm": 0.13609148561954498, |
|
"learning_rate": 4.294991660718636e-06, |
|
"loss": 0.6715, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.1122071516646117, |
|
"grad_norm": 0.12187358736991882, |
|
"learning_rate": 4.284153306221289e-06, |
|
"loss": 0.6785, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.1171393341553637, |
|
"grad_norm": 0.10929637402296066, |
|
"learning_rate": 4.273246176900252e-06, |
|
"loss": 0.6757, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 1.122071516646116, |
|
"grad_norm": 0.12787918746471405, |
|
"learning_rate": 4.262270693199159e-06, |
|
"loss": 0.6895, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 1.1270036991368682, |
|
"grad_norm": 0.11202108860015869, |
|
"learning_rate": 4.2512272781965355e-06, |
|
"loss": 0.6823, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.1319358816276202, |
|
"grad_norm": 0.5738335251808167, |
|
"learning_rate": 4.240116357589502e-06, |
|
"loss": 0.6724, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 1.1368680641183724, |
|
"grad_norm": 0.13368858397006989, |
|
"learning_rate": 4.228938359677354e-06, |
|
"loss": 0.6583, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.1418002466091246, |
|
"grad_norm": 0.10556632280349731, |
|
"learning_rate": 4.217693715345057e-06, |
|
"loss": 0.6444, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 1.1467324290998766, |
|
"grad_norm": 0.28452828526496887, |
|
"learning_rate": 4.206382858046636e-06, |
|
"loss": 0.6638, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.1516646115906288, |
|
"grad_norm": 0.19593405723571777, |
|
"learning_rate": 4.195006223788466e-06, |
|
"loss": 0.6537, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 1.156596794081381, |
|
"grad_norm": 0.11627080291509628, |
|
"learning_rate": 4.183564251112466e-06, |
|
"loss": 0.6699, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 1.161528976572133, |
|
"grad_norm": 0.15152066946029663, |
|
"learning_rate": 4.172057381079196e-06, |
|
"loss": 0.6685, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.1664611590628853, |
|
"grad_norm": 0.20805396139621735, |
|
"learning_rate": 4.160486057250849e-06, |
|
"loss": 0.6605, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 1.1713933415536375, |
|
"grad_norm": 0.16027909517288208, |
|
"learning_rate": 4.148850725674162e-06, |
|
"loss": 0.6959, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 1.1763255240443897, |
|
"grad_norm": 0.12061706185340881, |
|
"learning_rate": 4.137151834863213e-06, |
|
"loss": 0.6706, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 1.1812577065351417, |
|
"grad_norm": 0.09794213622808456, |
|
"learning_rate": 4.125389835782138e-06, |
|
"loss": 0.6846, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 1.186189889025894, |
|
"grad_norm": 0.13422715663909912, |
|
"learning_rate": 4.113565181827745e-06, |
|
"loss": 0.6611, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.1911220715166462, |
|
"grad_norm": 0.11418966203927994, |
|
"learning_rate": 4.101678328812035e-06, |
|
"loss": 0.6687, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 1.1960542540073984, |
|
"grad_norm": 0.2175075113773346, |
|
"learning_rate": 4.0897297349446345e-06, |
|
"loss": 0.6933, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 1.2009864364981504, |
|
"grad_norm": 0.1245436891913414, |
|
"learning_rate": 4.077719860815132e-06, |
|
"loss": 0.6538, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 1.2059186189889026, |
|
"grad_norm": 0.12700359523296356, |
|
"learning_rate": 4.065649169375324e-06, |
|
"loss": 0.6555, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.2108508014796548, |
|
"grad_norm": 0.12477461248636246, |
|
"learning_rate": 4.053518125921365e-06, |
|
"loss": 0.6664, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 1.2157829839704068, |
|
"grad_norm": 0.16408082842826843, |
|
"learning_rate": 4.041327198075838e-06, |
|
"loss": 0.6482, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 1.220715166461159, |
|
"grad_norm": 0.11284561455249786, |
|
"learning_rate": 4.029076855769722e-06, |
|
"loss": 0.6557, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 1.2256473489519113, |
|
"grad_norm": 0.11041481792926788, |
|
"learning_rate": 4.016767571224285e-06, |
|
"loss": 0.6445, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.2305795314426633, |
|
"grad_norm": 0.1686171293258667, |
|
"learning_rate": 4.0043998189328705e-06, |
|
"loss": 0.641, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 1.2355117139334155, |
|
"grad_norm": 0.10102049261331558, |
|
"learning_rate": 3.991974075642621e-06, |
|
"loss": 0.6522, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.2404438964241677, |
|
"grad_norm": 0.22298377752304077, |
|
"learning_rate": 3.9794908203360865e-06, |
|
"loss": 0.6942, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 1.2453760789149197, |
|
"grad_norm": 0.11277928948402405, |
|
"learning_rate": 3.966950534212769e-06, |
|
"loss": 0.647, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 1.250308261405672, |
|
"grad_norm": 0.12690778076648712, |
|
"learning_rate": 3.954353700670573e-06, |
|
"loss": 0.6682, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 1.2552404438964242, |
|
"grad_norm": 0.1330857127904892, |
|
"learning_rate": 3.941700805287169e-06, |
|
"loss": 0.6902, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 1.2601726263871762, |
|
"grad_norm": 0.10677850991487503, |
|
"learning_rate": 3.9289923358012735e-06, |
|
"loss": 0.6523, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 1.2651048088779284, |
|
"grad_norm": 0.09997207671403885, |
|
"learning_rate": 3.9162287820938575e-06, |
|
"loss": 0.6396, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 1.2700369913686806, |
|
"grad_norm": 0.10573034733533859, |
|
"learning_rate": 3.903410636169252e-06, |
|
"loss": 0.6631, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 1.2749691738594329, |
|
"grad_norm": 0.11500944197177887, |
|
"learning_rate": 3.890538392136188e-06, |
|
"loss": 0.6485, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 1.279901356350185, |
|
"grad_norm": 0.11942487210035324, |
|
"learning_rate": 3.877612546188749e-06, |
|
"loss": 0.6783, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 1.284833538840937, |
|
"grad_norm": 0.22677327692508698, |
|
"learning_rate": 3.864633596587242e-06, |
|
"loss": 0.6761, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.2897657213316893, |
|
"grad_norm": 0.10834953933954239, |
|
"learning_rate": 3.8516020436389945e-06, |
|
"loss": 0.6667, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 1.2946979038224415, |
|
"grad_norm": 0.10524627566337585, |
|
"learning_rate": 3.838518389679065e-06, |
|
"loss": 0.6722, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 1.2996300863131935, |
|
"grad_norm": 0.20889033377170563, |
|
"learning_rate": 3.825383139050881e-06, |
|
"loss": 0.6471, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 1.3045622688039458, |
|
"grad_norm": 0.09844981133937836, |
|
"learning_rate": 3.812196798086799e-06, |
|
"loss": 0.6847, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 1.309494451294698, |
|
"grad_norm": 0.10606922209262848, |
|
"learning_rate": 3.798959875088584e-06, |
|
"loss": 0.666, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 1.31442663378545, |
|
"grad_norm": 0.1338571161031723, |
|
"learning_rate": 3.7856728803078168e-06, |
|
"loss": 0.6438, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 1.3193588162762022, |
|
"grad_norm": 0.09691350907087326, |
|
"learning_rate": 3.7723363259262253e-06, |
|
"loss": 0.6567, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 1.3242909987669544, |
|
"grad_norm": 0.12978605926036835, |
|
"learning_rate": 3.75895072603594e-06, |
|
"loss": 0.654, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 1.3292231812577064, |
|
"grad_norm": 0.11295681446790695, |
|
"learning_rate": 3.7455165966196817e-06, |
|
"loss": 0.6981, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 1.3341553637484587, |
|
"grad_norm": 0.11601907014846802, |
|
"learning_rate": 3.732034455530863e-06, |
|
"loss": 0.6891, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.339087546239211, |
|
"grad_norm": 0.10687986761331558, |
|
"learning_rate": 3.718504822473634e-06, |
|
"loss": 0.6606, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 1.344019728729963, |
|
"grad_norm": 0.11528074741363525, |
|
"learning_rate": 3.704928218982845e-06, |
|
"loss": 0.6481, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 1.3489519112207151, |
|
"grad_norm": 0.18344081938266754, |
|
"learning_rate": 3.6913051684039435e-06, |
|
"loss": 0.6645, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 1.3538840937114673, |
|
"grad_norm": 0.2689914405345917, |
|
"learning_rate": 3.6776361958728025e-06, |
|
"loss": 0.6648, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 1.3588162762022196, |
|
"grad_norm": 0.3526904582977295, |
|
"learning_rate": 3.663921828295474e-06, |
|
"loss": 0.6679, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 1.3637484586929716, |
|
"grad_norm": 0.12253173440694809, |
|
"learning_rate": 3.650162594327881e-06, |
|
"loss": 0.649, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 1.3686806411837238, |
|
"grad_norm": 0.11910586804151535, |
|
"learning_rate": 3.6363590243554362e-06, |
|
"loss": 0.6372, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 1.373612823674476, |
|
"grad_norm": 0.2642151117324829, |
|
"learning_rate": 3.6225116504726014e-06, |
|
"loss": 0.6465, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 1.3785450061652282, |
|
"grad_norm": 0.12019569426774979, |
|
"learning_rate": 3.6086210064623735e-06, |
|
"loss": 0.6815, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 1.3834771886559802, |
|
"grad_norm": 0.12397979199886322, |
|
"learning_rate": 3.5946876277757066e-06, |
|
"loss": 0.6836, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.3884093711467325, |
|
"grad_norm": 0.12133345752954483, |
|
"learning_rate": 3.580712051510876e-06, |
|
"loss": 0.6514, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 1.3933415536374847, |
|
"grad_norm": 0.12849684059619904, |
|
"learning_rate": 3.5666948163927716e-06, |
|
"loss": 0.6651, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 1.3982737361282367, |
|
"grad_norm": 0.10554145276546478, |
|
"learning_rate": 3.5526364627521322e-06, |
|
"loss": 0.6784, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 1.403205918618989, |
|
"grad_norm": 0.12648458778858185, |
|
"learning_rate": 3.5385375325047167e-06, |
|
"loss": 0.6738, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 1.4081381011097411, |
|
"grad_norm": 0.10506008565425873, |
|
"learning_rate": 3.5243985691304146e-06, |
|
"loss": 0.6478, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 1.4130702836004931, |
|
"grad_norm": 0.1196879968047142, |
|
"learning_rate": 3.5102201176522966e-06, |
|
"loss": 0.6516, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 1.4180024660912454, |
|
"grad_norm": 0.10076049715280533, |
|
"learning_rate": 3.4960027246156043e-06, |
|
"loss": 0.6363, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 1.4229346485819976, |
|
"grad_norm": 0.11161559075117111, |
|
"learning_rate": 3.4817469380666834e-06, |
|
"loss": 0.6501, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 1.4278668310727496, |
|
"grad_norm": 0.10083605349063873, |
|
"learning_rate": 3.467453307531858e-06, |
|
"loss": 0.6715, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 1.4327990135635018, |
|
"grad_norm": 0.16506797075271606, |
|
"learning_rate": 3.453122383996245e-06, |
|
"loss": 0.6404, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.437731196054254, |
|
"grad_norm": 0.12411224842071533, |
|
"learning_rate": 3.4387547198825187e-06, |
|
"loss": 0.668, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 1.442663378545006, |
|
"grad_norm": 0.1273510903120041, |
|
"learning_rate": 3.4243508690296133e-06, |
|
"loss": 0.6571, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 1.4475955610357583, |
|
"grad_norm": 0.10576938837766647, |
|
"learning_rate": 3.409911386671375e-06, |
|
"loss": 0.656, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 1.4525277435265105, |
|
"grad_norm": 0.10712239891290665, |
|
"learning_rate": 3.3954368294151603e-06, |
|
"loss": 0.6474, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 1.4574599260172627, |
|
"grad_norm": 0.1035386323928833, |
|
"learning_rate": 3.380927755220376e-06, |
|
"loss": 0.6512, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 1.462392108508015, |
|
"grad_norm": 0.1390061378479004, |
|
"learning_rate": 3.366384723376977e-06, |
|
"loss": 0.6538, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 1.467324290998767, |
|
"grad_norm": 0.12152580916881561, |
|
"learning_rate": 3.351808294483902e-06, |
|
"loss": 0.6414, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 1.4722564734895192, |
|
"grad_norm": 0.11257496476173401, |
|
"learning_rate": 3.3371990304274654e-06, |
|
"loss": 0.6793, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 1.4771886559802714, |
|
"grad_norm": 0.10283617675304413, |
|
"learning_rate": 3.3225574943597005e-06, |
|
"loss": 0.6444, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 1.4821208384710234, |
|
"grad_norm": 0.12117356061935425, |
|
"learning_rate": 3.3078842506766484e-06, |
|
"loss": 0.631, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.4870530209617756, |
|
"grad_norm": 0.14655297994613647, |
|
"learning_rate": 3.2931798649966e-06, |
|
"loss": 0.6614, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 1.4919852034525278, |
|
"grad_norm": 0.1442922204732895, |
|
"learning_rate": 3.2784449041382973e-06, |
|
"loss": 0.6602, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 1.4969173859432798, |
|
"grad_norm": 0.11084005236625671, |
|
"learning_rate": 3.263679936099083e-06, |
|
"loss": 0.6616, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 1.501849568434032, |
|
"grad_norm": 0.1568291336297989, |
|
"learning_rate": 3.248885530033004e-06, |
|
"loss": 0.6745, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 1.5067817509247843, |
|
"grad_norm": 0.11827261000871658, |
|
"learning_rate": 3.2340622562288717e-06, |
|
"loss": 0.666, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 1.5117139334155363, |
|
"grad_norm": 0.13850663602352142, |
|
"learning_rate": 3.2192106860882782e-06, |
|
"loss": 0.6425, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 1.5166461159062885, |
|
"grad_norm": 0.11242467910051346, |
|
"learning_rate": 3.2043313921035747e-06, |
|
"loss": 0.6366, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 1.5215782983970407, |
|
"grad_norm": 0.11277095973491669, |
|
"learning_rate": 3.189424947835797e-06, |
|
"loss": 0.6423, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 1.5265104808877927, |
|
"grad_norm": 0.1134696900844574, |
|
"learning_rate": 3.174491927892561e-06, |
|
"loss": 0.6637, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 1.531442663378545, |
|
"grad_norm": 0.10201112180948257, |
|
"learning_rate": 3.1595329079059102e-06, |
|
"loss": 0.6458, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.5363748458692972, |
|
"grad_norm": 0.22865180671215057, |
|
"learning_rate": 3.144548464510127e-06, |
|
"loss": 0.6693, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 1.5413070283600492, |
|
"grad_norm": 0.12015289813280106, |
|
"learning_rate": 3.129539175319505e-06, |
|
"loss": 0.6463, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 1.5462392108508016, |
|
"grad_norm": 0.14742949604988098, |
|
"learning_rate": 3.114505618906086e-06, |
|
"loss": 0.6398, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 1.5511713933415536, |
|
"grad_norm": 0.12067458778619766, |
|
"learning_rate": 3.0994483747773508e-06, |
|
"loss": 0.6618, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 1.5561035758323056, |
|
"grad_norm": 0.5674145817756653, |
|
"learning_rate": 3.0843680233538885e-06, |
|
"loss": 0.6552, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 1.561035758323058, |
|
"grad_norm": 0.1083153635263443, |
|
"learning_rate": 3.0692651459470164e-06, |
|
"loss": 0.6283, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 1.56596794081381, |
|
"grad_norm": 0.177827849984169, |
|
"learning_rate": 3.0541403247363756e-06, |
|
"loss": 0.6423, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 1.5709001233045623, |
|
"grad_norm": 0.10957730561494827, |
|
"learning_rate": 3.0389941427474873e-06, |
|
"loss": 0.6569, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 1.5758323057953145, |
|
"grad_norm": 0.14152103662490845, |
|
"learning_rate": 3.0238271838292815e-06, |
|
"loss": 0.6607, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 1.5807644882860665, |
|
"grad_norm": 0.2304203361272812, |
|
"learning_rate": 3.0086400326315853e-06, |
|
"loss": 0.6798, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.5856966707768188, |
|
"grad_norm": 0.11011958122253418, |
|
"learning_rate": 2.9934332745825924e-06, |
|
"loss": 0.6308, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 1.590628853267571, |
|
"grad_norm": 0.12152457237243652, |
|
"learning_rate": 2.9782074958662923e-06, |
|
"loss": 0.6423, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 1.595561035758323, |
|
"grad_norm": 0.10855505615472794, |
|
"learning_rate": 2.962963283399877e-06, |
|
"loss": 0.6759, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 1.6004932182490752, |
|
"grad_norm": 0.12242516875267029, |
|
"learning_rate": 2.947701224811113e-06, |
|
"loss": 0.6489, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 1.6054254007398274, |
|
"grad_norm": 0.11491915583610535, |
|
"learning_rate": 2.932421908415694e-06, |
|
"loss": 0.6476, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.6103575832305794, |
|
"grad_norm": 0.10774843394756317, |
|
"learning_rate": 2.9171259231945598e-06, |
|
"loss": 0.6535, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 1.6152897657213316, |
|
"grad_norm": 0.10615874826908112, |
|
"learning_rate": 2.901813858771193e-06, |
|
"loss": 0.6477, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 1.6202219482120839, |
|
"grad_norm": 0.09981298446655273, |
|
"learning_rate": 2.8864863053888927e-06, |
|
"loss": 0.6525, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 1.6251541307028359, |
|
"grad_norm": 0.1059933453798294, |
|
"learning_rate": 2.871143853888017e-06, |
|
"loss": 0.6609, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 1.6300863131935883, |
|
"grad_norm": 0.28401389718055725, |
|
"learning_rate": 2.8557870956832135e-06, |
|
"loss": 0.6738, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.6350184956843403, |
|
"grad_norm": 0.1126745268702507, |
|
"learning_rate": 2.840416622740617e-06, |
|
"loss": 0.6443, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 1.6399506781750923, |
|
"grad_norm": 0.10980167984962463, |
|
"learning_rate": 2.8250330275550337e-06, |
|
"loss": 0.6575, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 1.6448828606658448, |
|
"grad_norm": 0.20389322936534882, |
|
"learning_rate": 2.8096369031271e-06, |
|
"loss": 0.6855, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 1.6498150431565968, |
|
"grad_norm": 0.1709917038679123, |
|
"learning_rate": 2.7942288429404256e-06, |
|
"loss": 0.6752, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 1.654747225647349, |
|
"grad_norm": 0.12198632955551147, |
|
"learning_rate": 2.778809440938714e-06, |
|
"loss": 0.6398, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 1.6596794081381012, |
|
"grad_norm": 0.11568919569253922, |
|
"learning_rate": 2.763379291502868e-06, |
|
"loss": 0.6489, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.6646115906288532, |
|
"grad_norm": 0.15780635178089142, |
|
"learning_rate": 2.7479389894280793e-06, |
|
"loss": 0.6489, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 1.6695437731196054, |
|
"grad_norm": 0.1101425290107727, |
|
"learning_rate": 2.7324891299008987e-06, |
|
"loss": 0.6726, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 1.6744759556103577, |
|
"grad_norm": 0.10121767222881317, |
|
"learning_rate": 2.7170303084762958e-06, |
|
"loss": 0.6726, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 1.6794081381011097, |
|
"grad_norm": 0.1087949350476265, |
|
"learning_rate": 2.701563121054695e-06, |
|
"loss": 0.6175, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.684340320591862, |
|
"grad_norm": 0.11207877844572067, |
|
"learning_rate": 2.6860881638590146e-06, |
|
"loss": 0.6565, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 1.6892725030826141, |
|
"grad_norm": 0.11743203550577164, |
|
"learning_rate": 2.670606033411678e-06, |
|
"loss": 0.677, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 1.6942046855733661, |
|
"grad_norm": 0.10707394033670425, |
|
"learning_rate": 2.6551173265116182e-06, |
|
"loss": 0.6601, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 1.6991368680641183, |
|
"grad_norm": 0.10522118955850601, |
|
"learning_rate": 2.6396226402112768e-06, |
|
"loss": 0.6494, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 1.7040690505548706, |
|
"grad_norm": 0.19015160202980042, |
|
"learning_rate": 2.624122571793586e-06, |
|
"loss": 0.6494, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.7090012330456226, |
|
"grad_norm": 0.11446730047464371, |
|
"learning_rate": 2.6086177187489453e-06, |
|
"loss": 0.6183, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 1.7139334155363748, |
|
"grad_norm": 0.17297939956188202, |
|
"learning_rate": 2.593108678752191e-06, |
|
"loss": 0.6572, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 1.718865598027127, |
|
"grad_norm": 0.12407731264829636, |
|
"learning_rate": 2.5775960496395565e-06, |
|
"loss": 0.6496, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 1.723797780517879, |
|
"grad_norm": 0.36749541759490967, |
|
"learning_rate": 2.562080429385626e-06, |
|
"loss": 0.6611, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 1.7287299630086315, |
|
"grad_norm": 0.10345156490802765, |
|
"learning_rate": 2.5465624160802847e-06, |
|
"loss": 0.6716, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.7336621454993835, |
|
"grad_norm": 0.11017563939094543, |
|
"learning_rate": 2.531042607905665e-06, |
|
"loss": 0.6499, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 1.7385943279901355, |
|
"grad_norm": 0.09504040330648422, |
|
"learning_rate": 2.5155216031130882e-06, |
|
"loss": 0.6657, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 1.743526510480888, |
|
"grad_norm": 0.11456336081027985, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.6433, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 1.74845869297164, |
|
"grad_norm": 0.3017407953739166, |
|
"learning_rate": 2.4844783968869126e-06, |
|
"loss": 0.6678, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 1.7533908754623921, |
|
"grad_norm": 0.17654098570346832, |
|
"learning_rate": 2.4689573920943358e-06, |
|
"loss": 0.6407, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 1.7583230579531444, |
|
"grad_norm": 0.19192291796207428, |
|
"learning_rate": 2.4534375839197166e-06, |
|
"loss": 0.6538, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 1.7632552404438964, |
|
"grad_norm": 0.101350799202919, |
|
"learning_rate": 2.4379195706143755e-06, |
|
"loss": 0.6563, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 1.7681874229346486, |
|
"grad_norm": 0.09221342951059341, |
|
"learning_rate": 2.422403950360444e-06, |
|
"loss": 0.6465, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 1.7731196054254008, |
|
"grad_norm": 0.1029122844338417, |
|
"learning_rate": 2.406891321247809e-06, |
|
"loss": 0.6524, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 1.7780517879161528, |
|
"grad_norm": 0.10780887305736542, |
|
"learning_rate": 2.391382281251055e-06, |
|
"loss": 0.6577, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.782983970406905, |
|
"grad_norm": 0.11179229617118835, |
|
"learning_rate": 2.375877428206415e-06, |
|
"loss": 0.6281, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 1.7879161528976573, |
|
"grad_norm": 0.1713687628507614, |
|
"learning_rate": 2.360377359788724e-06, |
|
"loss": 0.6866, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 1.7928483353884093, |
|
"grad_norm": 0.20977120101451874, |
|
"learning_rate": 2.3448826734883826e-06, |
|
"loss": 0.6524, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 1.7977805178791615, |
|
"grad_norm": 0.10442214459180832, |
|
"learning_rate": 2.3293939665883233e-06, |
|
"loss": 0.6356, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 1.8027127003699137, |
|
"grad_norm": 0.11700151115655899, |
|
"learning_rate": 2.313911836140986e-06, |
|
"loss": 0.635, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.8076448828606657, |
|
"grad_norm": 0.20740574598312378, |
|
"learning_rate": 2.298436878945306e-06, |
|
"loss": 0.6407, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 1.8125770653514182, |
|
"grad_norm": 0.10487078875303268, |
|
"learning_rate": 2.2829696915237055e-06, |
|
"loss": 0.6606, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 1.8175092478421702, |
|
"grad_norm": 0.19342948496341705, |
|
"learning_rate": 2.267510870099101e-06, |
|
"loss": 0.6652, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 1.8224414303329222, |
|
"grad_norm": 0.12832309305667877, |
|
"learning_rate": 2.252061010571921e-06, |
|
"loss": 0.6699, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 1.8273736128236746, |
|
"grad_norm": 0.24777992069721222, |
|
"learning_rate": 2.2366207084971326e-06, |
|
"loss": 0.652, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.8323057953144266, |
|
"grad_norm": 0.2240799218416214, |
|
"learning_rate": 2.2211905590612864e-06, |
|
"loss": 0.6559, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 1.8372379778051788, |
|
"grad_norm": 0.21929948031902313, |
|
"learning_rate": 2.205771157059575e-06, |
|
"loss": 0.6524, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 1.842170160295931, |
|
"grad_norm": 0.09952554106712341, |
|
"learning_rate": 2.1903630968729003e-06, |
|
"loss": 0.663, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 1.847102342786683, |
|
"grad_norm": 0.13348402082920074, |
|
"learning_rate": 2.174966972444967e-06, |
|
"loss": 0.6594, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 1.8520345252774353, |
|
"grad_norm": 0.13318030536174774, |
|
"learning_rate": 2.159583377259384e-06, |
|
"loss": 0.6271, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.8569667077681875, |
|
"grad_norm": 0.10292787104845047, |
|
"learning_rate": 2.1442129043167877e-06, |
|
"loss": 0.6537, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 1.8618988902589395, |
|
"grad_norm": 0.1853366643190384, |
|
"learning_rate": 2.1288561461119837e-06, |
|
"loss": 0.6434, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 1.8668310727496917, |
|
"grad_norm": 0.3069738745689392, |
|
"learning_rate": 2.1135136946111077e-06, |
|
"loss": 0.6294, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 1.871763255240444, |
|
"grad_norm": 0.1591643989086151, |
|
"learning_rate": 2.098186141228807e-06, |
|
"loss": 0.6596, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 1.876695437731196, |
|
"grad_norm": 0.1068023145198822, |
|
"learning_rate": 2.0828740768054406e-06, |
|
"loss": 0.6559, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.8816276202219482, |
|
"grad_norm": 0.13300898671150208, |
|
"learning_rate": 2.0675780915843068e-06, |
|
"loss": 0.63, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 1.8865598027127004, |
|
"grad_norm": 0.13550280034542084, |
|
"learning_rate": 2.052298775188888e-06, |
|
"loss": 0.646, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 1.8914919852034524, |
|
"grad_norm": 0.19939783215522766, |
|
"learning_rate": 2.0370367166001245e-06, |
|
"loss": 0.6511, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 1.8964241676942046, |
|
"grad_norm": 0.13335835933685303, |
|
"learning_rate": 2.021792504133709e-06, |
|
"loss": 0.6584, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 1.9013563501849569, |
|
"grad_norm": 0.0967630073428154, |
|
"learning_rate": 2.006566725417409e-06, |
|
"loss": 0.656, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.9062885326757089, |
|
"grad_norm": 0.10085921734571457, |
|
"learning_rate": 1.991359967368416e-06, |
|
"loss": 0.6515, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 1.9112207151664613, |
|
"grad_norm": 0.1494772732257843, |
|
"learning_rate": 1.97617281617072e-06, |
|
"loss": 0.6625, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 1.9161528976572133, |
|
"grad_norm": 0.14001788198947906, |
|
"learning_rate": 1.9610058572525127e-06, |
|
"loss": 0.6265, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 1.9210850801479655, |
|
"grad_norm": 0.10699108988046646, |
|
"learning_rate": 1.945859675263625e-06, |
|
"loss": 0.6446, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 1.9260172626387178, |
|
"grad_norm": 0.10522957146167755, |
|
"learning_rate": 1.9307348540529845e-06, |
|
"loss": 0.6525, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.9309494451294698, |
|
"grad_norm": 0.0986817479133606, |
|
"learning_rate": 1.9156319766461124e-06, |
|
"loss": 0.6325, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 1.935881627620222, |
|
"grad_norm": 0.11651718616485596, |
|
"learning_rate": 1.90055162522265e-06, |
|
"loss": 0.6303, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 1.9408138101109742, |
|
"grad_norm": 0.10605504363775253, |
|
"learning_rate": 1.8854943810939152e-06, |
|
"loss": 0.6559, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 1.9457459926017262, |
|
"grad_norm": 0.09837733954191208, |
|
"learning_rate": 1.8704608246804956e-06, |
|
"loss": 0.6708, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 1.9506781750924784, |
|
"grad_norm": 0.08927737921476364, |
|
"learning_rate": 1.8554515354898744e-06, |
|
"loss": 0.654, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.9556103575832307, |
|
"grad_norm": 0.10312498360872269, |
|
"learning_rate": 1.840467092094091e-06, |
|
"loss": 0.6178, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.9605425400739827, |
|
"grad_norm": 0.12431120872497559, |
|
"learning_rate": 1.8255080721074391e-06, |
|
"loss": 0.6804, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 1.9654747225647349, |
|
"grad_norm": 0.17175759375095367, |
|
"learning_rate": 1.8105750521642035e-06, |
|
"loss": 0.6509, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 1.970406905055487, |
|
"grad_norm": 0.6188605427742004, |
|
"learning_rate": 1.7956686078964257e-06, |
|
"loss": 0.6311, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 1.975339087546239, |
|
"grad_norm": 0.15260924398899078, |
|
"learning_rate": 1.7807893139117222e-06, |
|
"loss": 0.6562, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.9802712700369913, |
|
"grad_norm": 0.10048322379589081, |
|
"learning_rate": 1.7659377437711294e-06, |
|
"loss": 0.6562, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 1.9852034525277436, |
|
"grad_norm": 0.14348655939102173, |
|
"learning_rate": 1.7511144699669967e-06, |
|
"loss": 0.6415, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 1.9901356350184956, |
|
"grad_norm": 0.2633483111858368, |
|
"learning_rate": 1.7363200639009176e-06, |
|
"loss": 0.6289, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 1.995067817509248, |
|
"grad_norm": 0.10529658943414688, |
|
"learning_rate": 1.7215550958617036e-06, |
|
"loss": 0.6494, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 2.0024660912453762, |
|
"grad_norm": 0.12062520533800125, |
|
"learning_rate": 1.7068201350034017e-06, |
|
"loss": 0.6449, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 2.0073982737361282, |
|
"grad_norm": 0.1379857212305069, |
|
"learning_rate": 1.692115749323353e-06, |
|
"loss": 0.6557, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 2.0123304562268802, |
|
"grad_norm": 0.10167136788368225, |
|
"learning_rate": 1.6774425056402993e-06, |
|
"loss": 0.6362, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 2.0172626387176327, |
|
"grad_norm": 0.1222464069724083, |
|
"learning_rate": 1.6628009695725348e-06, |
|
"loss": 0.6479, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 2.0221948212083847, |
|
"grad_norm": 0.19944354891777039, |
|
"learning_rate": 1.648191705516099e-06, |
|
"loss": 0.6606, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 2.0271270036991367, |
|
"grad_norm": 0.09848273545503616, |
|
"learning_rate": 1.6336152766230235e-06, |
|
"loss": 0.6408, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.032059186189889, |
|
"grad_norm": 0.09147990494966507, |
|
"learning_rate": 1.6190722447796242e-06, |
|
"loss": 0.6169, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 2.036991368680641, |
|
"grad_norm": 0.09940177202224731, |
|
"learning_rate": 1.6045631705848405e-06, |
|
"loss": 0.6397, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 2.041923551171393, |
|
"grad_norm": 0.10459251701831818, |
|
"learning_rate": 1.5900886133286254e-06, |
|
"loss": 0.6209, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 2.0468557336621456, |
|
"grad_norm": 0.10069490969181061, |
|
"learning_rate": 1.5756491309703875e-06, |
|
"loss": 0.6516, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 2.0517879161528976, |
|
"grad_norm": 0.10702253878116608, |
|
"learning_rate": 1.561245280117482e-06, |
|
"loss": 0.6491, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 2.05672009864365, |
|
"grad_norm": 0.10752403736114502, |
|
"learning_rate": 1.5468776160037558e-06, |
|
"loss": 0.6554, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 2.061652281134402, |
|
"grad_norm": 0.1126432865858078, |
|
"learning_rate": 1.5325466924681425e-06, |
|
"loss": 0.642, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 2.066584463625154, |
|
"grad_norm": 0.13967493176460266, |
|
"learning_rate": 1.5182530619333168e-06, |
|
"loss": 0.6271, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 2.0715166461159065, |
|
"grad_norm": 0.107243612408638, |
|
"learning_rate": 1.5039972753843966e-06, |
|
"loss": 0.6296, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 2.0764488286066585, |
|
"grad_norm": 0.12694260478019714, |
|
"learning_rate": 1.4897798823477045e-06, |
|
"loss": 0.6584, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.0813810110974105, |
|
"grad_norm": 0.11211492121219635, |
|
"learning_rate": 1.4756014308695865e-06, |
|
"loss": 0.6394, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 2.086313193588163, |
|
"grad_norm": 0.10561738908290863, |
|
"learning_rate": 1.4614624674952843e-06, |
|
"loss": 0.6333, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 2.091245376078915, |
|
"grad_norm": 0.11454541981220245, |
|
"learning_rate": 1.4473635372478692e-06, |
|
"loss": 0.6185, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 2.096177558569667, |
|
"grad_norm": 0.11180110275745392, |
|
"learning_rate": 1.4333051836072298e-06, |
|
"loss": 0.6266, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 2.1011097410604194, |
|
"grad_norm": 0.10139421373605728, |
|
"learning_rate": 1.4192879484891253e-06, |
|
"loss": 0.6519, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 2.1060419235511714, |
|
"grad_norm": 0.10878422111272812, |
|
"learning_rate": 1.405312372224294e-06, |
|
"loss": 0.6253, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 2.1109741060419234, |
|
"grad_norm": 0.09218919277191162, |
|
"learning_rate": 1.3913789935376271e-06, |
|
"loss": 0.6373, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 2.115906288532676, |
|
"grad_norm": 0.10704641044139862, |
|
"learning_rate": 1.3774883495273986e-06, |
|
"loss": 0.6498, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 2.120838471023428, |
|
"grad_norm": 0.09850191324949265, |
|
"learning_rate": 1.363640975644564e-06, |
|
"loss": 0.6595, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 2.12577065351418, |
|
"grad_norm": 0.13874687254428864, |
|
"learning_rate": 1.3498374056721198e-06, |
|
"loss": 0.6555, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.1307028360049323, |
|
"grad_norm": 0.11779513210058212, |
|
"learning_rate": 1.3360781717045266e-06, |
|
"loss": 0.6217, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 2.1356350184956843, |
|
"grad_norm": 0.22133827209472656, |
|
"learning_rate": 1.322363804127198e-06, |
|
"loss": 0.6624, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 2.1405672009864363, |
|
"grad_norm": 0.1059156209230423, |
|
"learning_rate": 1.3086948315960567e-06, |
|
"loss": 0.6412, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 2.1454993834771887, |
|
"grad_norm": 0.1276949644088745, |
|
"learning_rate": 1.295071781017156e-06, |
|
"loss": 0.6325, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 2.1504315659679407, |
|
"grad_norm": 0.09928814321756363, |
|
"learning_rate": 1.2814951775263671e-06, |
|
"loss": 0.6165, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 2.155363748458693, |
|
"grad_norm": 0.1885470151901245, |
|
"learning_rate": 1.267965544469137e-06, |
|
"loss": 0.6294, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 2.160295930949445, |
|
"grad_norm": 0.11126084625720978, |
|
"learning_rate": 1.2544834033803183e-06, |
|
"loss": 0.6392, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 2.165228113440197, |
|
"grad_norm": 0.09588748216629028, |
|
"learning_rate": 1.2410492739640592e-06, |
|
"loss": 0.6242, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 2.1701602959309496, |
|
"grad_norm": 0.3128255009651184, |
|
"learning_rate": 1.227663674073775e-06, |
|
"loss": 0.6191, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 2.1750924784217016, |
|
"grad_norm": 0.11006899923086166, |
|
"learning_rate": 1.2143271196921832e-06, |
|
"loss": 0.6418, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.1800246609124536, |
|
"grad_norm": 0.12083282321691513, |
|
"learning_rate": 1.2010401249114166e-06, |
|
"loss": 0.6192, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 2.184956843403206, |
|
"grad_norm": 0.13546785712242126, |
|
"learning_rate": 1.1878032019132016e-06, |
|
"loss": 0.6688, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 2.189889025893958, |
|
"grad_norm": 0.1263391375541687, |
|
"learning_rate": 1.1746168609491198e-06, |
|
"loss": 0.639, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 2.19482120838471, |
|
"grad_norm": 0.357994943857193, |
|
"learning_rate": 1.1614816103209363e-06, |
|
"loss": 0.6582, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 2.1997533908754625, |
|
"grad_norm": 0.09652648866176605, |
|
"learning_rate": 1.148397956361007e-06, |
|
"loss": 0.6323, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 2.2046855733662145, |
|
"grad_norm": 0.12993961572647095, |
|
"learning_rate": 1.1353664034127585e-06, |
|
"loss": 0.6461, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 2.2096177558569665, |
|
"grad_norm": 0.10628235340118408, |
|
"learning_rate": 1.122387453811252e-06, |
|
"loss": 0.6315, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 2.214549938347719, |
|
"grad_norm": 2.75348162651062, |
|
"learning_rate": 1.1094616078638123e-06, |
|
"loss": 0.6616, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 2.219482120838471, |
|
"grad_norm": 0.13206490874290466, |
|
"learning_rate": 1.0965893638307484e-06, |
|
"loss": 0.6227, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 2.2244143033292234, |
|
"grad_norm": 0.10024918615818024, |
|
"learning_rate": 1.083771217906143e-06, |
|
"loss": 0.6171, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.2293464858199754, |
|
"grad_norm": 0.1251702457666397, |
|
"learning_rate": 1.071007664198727e-06, |
|
"loss": 0.6592, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 2.2342786683107274, |
|
"grad_norm": 0.12185829877853394, |
|
"learning_rate": 1.0582991947128324e-06, |
|
"loss": 0.6235, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 2.23921085080148, |
|
"grad_norm": 0.09852628409862518, |
|
"learning_rate": 1.0456462993294273e-06, |
|
"loss": 0.6207, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 2.244143033292232, |
|
"grad_norm": 0.11288397014141083, |
|
"learning_rate": 1.0330494657872312e-06, |
|
"loss": 0.6497, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 2.249075215782984, |
|
"grad_norm": 0.1829065978527069, |
|
"learning_rate": 1.0205091796639143e-06, |
|
"loss": 0.6385, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 2.2540073982737363, |
|
"grad_norm": 0.13576146960258484, |
|
"learning_rate": 1.008025924357379e-06, |
|
"loss": 0.6214, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 2.2589395807644883, |
|
"grad_norm": 0.1026289239525795, |
|
"learning_rate": 9.95600181067129e-07, |
|
"loss": 0.6422, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 2.2638717632552403, |
|
"grad_norm": 0.11270825564861298, |
|
"learning_rate": 9.832324287757158e-07, |
|
"loss": 0.6441, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 2.2688039457459928, |
|
"grad_norm": 0.15193237364292145, |
|
"learning_rate": 9.709231442302777e-07, |
|
"loss": 0.6211, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 2.2737361282367448, |
|
"grad_norm": 0.09436095505952835, |
|
"learning_rate": 9.586728019241622e-07, |
|
"loss": 0.6341, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.2786683107274968, |
|
"grad_norm": 0.503972053527832, |
|
"learning_rate": 9.464818740786357e-07, |
|
"loss": 0.6092, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 2.283600493218249, |
|
"grad_norm": 0.1313806027173996, |
|
"learning_rate": 9.343508306246771e-07, |
|
"loss": 0.6338, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 2.288532675709001, |
|
"grad_norm": 0.21144555509090424, |
|
"learning_rate": 9.222801391848688e-07, |
|
"loss": 0.6227, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 2.293464858199753, |
|
"grad_norm": 0.09302227199077606, |
|
"learning_rate": 9.102702650553672e-07, |
|
"loss": 0.6538, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 2.2983970406905057, |
|
"grad_norm": 0.15463946759700775, |
|
"learning_rate": 8.983216711879663e-07, |
|
"loss": 0.6429, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 2.3033292231812577, |
|
"grad_norm": 0.09232427924871445, |
|
"learning_rate": 8.86434818172256e-07, |
|
"loss": 0.6414, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 2.3082614056720097, |
|
"grad_norm": 0.4079054296016693, |
|
"learning_rate": 8.746101642178623e-07, |
|
"loss": 0.6505, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 2.313193588162762, |
|
"grad_norm": 0.11529503017663956, |
|
"learning_rate": 8.628481651367876e-07, |
|
"loss": 0.6379, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 2.318125770653514, |
|
"grad_norm": 0.2939020097255707, |
|
"learning_rate": 8.51149274325839e-07, |
|
"loss": 0.6404, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 2.323057953144266, |
|
"grad_norm": 0.09394059330224991, |
|
"learning_rate": 8.395139427491517e-07, |
|
"loss": 0.6456, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.3279901356350186, |
|
"grad_norm": 0.12832114100456238, |
|
"learning_rate": 8.279426189208057e-07, |
|
"loss": 0.6372, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 2.3329223181257706, |
|
"grad_norm": 0.1322140395641327, |
|
"learning_rate": 8.164357488875349e-07, |
|
"loss": 0.6199, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 2.337854500616523, |
|
"grad_norm": 0.1532570719718933, |
|
"learning_rate": 8.049937762115354e-07, |
|
"loss": 0.6197, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 2.342786683107275, |
|
"grad_norm": 0.11840520799160004, |
|
"learning_rate": 7.936171419533653e-07, |
|
"loss": 0.6225, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 2.347718865598027, |
|
"grad_norm": 0.11926340311765671, |
|
"learning_rate": 7.823062846549432e-07, |
|
"loss": 0.6369, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 2.3526510480887795, |
|
"grad_norm": 0.10458600521087646, |
|
"learning_rate": 7.71061640322646e-07, |
|
"loss": 0.6192, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 2.3575832305795315, |
|
"grad_norm": 0.09384766221046448, |
|
"learning_rate": 7.59883642410498e-07, |
|
"loss": 0.6319, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 2.3625154130702835, |
|
"grad_norm": 0.13794849812984467, |
|
"learning_rate": 7.487727218034646e-07, |
|
"loss": 0.6627, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 2.367447595561036, |
|
"grad_norm": 0.09522448480129242, |
|
"learning_rate": 7.377293068008421e-07, |
|
"loss": 0.612, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 2.372379778051788, |
|
"grad_norm": 0.34490853548049927, |
|
"learning_rate": 7.267538230997487e-07, |
|
"loss": 0.6447, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.37731196054254, |
|
"grad_norm": 0.3206603229045868, |
|
"learning_rate": 7.15846693778712e-07, |
|
"loss": 0.6633, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 2.3822441430332923, |
|
"grad_norm": 0.10969394445419312, |
|
"learning_rate": 7.050083392813651e-07, |
|
"loss": 0.63, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 2.3871763255240444, |
|
"grad_norm": 0.09678292274475098, |
|
"learning_rate": 6.942391774002352e-07, |
|
"loss": 0.6102, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 2.392108508014797, |
|
"grad_norm": 0.12844984233379364, |
|
"learning_rate": 6.835396232606414e-07, |
|
"loss": 0.6065, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 2.397040690505549, |
|
"grad_norm": 0.10880632698535919, |
|
"learning_rate": 6.729100893046897e-07, |
|
"loss": 0.6388, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 2.401972872996301, |
|
"grad_norm": 0.15737678110599518, |
|
"learning_rate": 6.623509852753798e-07, |
|
"loss": 0.6288, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 2.4069050554870532, |
|
"grad_norm": 0.26250696182250977, |
|
"learning_rate": 6.518627182008034e-07, |
|
"loss": 0.6355, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 2.4118372379778052, |
|
"grad_norm": 0.1346411406993866, |
|
"learning_rate": 6.414456923784593e-07, |
|
"loss": 0.6637, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 2.4167694204685573, |
|
"grad_norm": 0.10590661317110062, |
|
"learning_rate": 6.311003093596674e-07, |
|
"loss": 0.6191, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 2.4217016029593097, |
|
"grad_norm": 0.10193908959627151, |
|
"learning_rate": 6.208269679340886e-07, |
|
"loss": 0.6224, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.4266337854500617, |
|
"grad_norm": 0.12438759952783585, |
|
"learning_rate": 6.106260641143547e-07, |
|
"loss": 0.6291, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 2.4315659679408137, |
|
"grad_norm": 0.11157315969467163, |
|
"learning_rate": 6.004979911208006e-07, |
|
"loss": 0.6265, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 2.436498150431566, |
|
"grad_norm": 0.10205821692943573, |
|
"learning_rate": 5.904431393663088e-07, |
|
"loss": 0.6219, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 2.441430332922318, |
|
"grad_norm": 0.0957137867808342, |
|
"learning_rate": 5.804618964412587e-07, |
|
"loss": 0.6376, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 2.44636251541307, |
|
"grad_norm": 0.14211417734622955, |
|
"learning_rate": 5.705546470985851e-07, |
|
"loss": 0.6255, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 2.4512946979038226, |
|
"grad_norm": 0.12571153044700623, |
|
"learning_rate": 5.607217732389503e-07, |
|
"loss": 0.6222, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 2.4562268803945746, |
|
"grad_norm": 0.18337282538414001, |
|
"learning_rate": 5.509636538960183e-07, |
|
"loss": 0.661, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 2.4611590628853266, |
|
"grad_norm": 0.10181237757205963, |
|
"learning_rate": 5.412806652218469e-07, |
|
"loss": 0.6097, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 2.466091245376079, |
|
"grad_norm": 0.10895208269357681, |
|
"learning_rate": 5.316731804723877e-07, |
|
"loss": 0.6544, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 2.471023427866831, |
|
"grad_norm": 0.11107916384935379, |
|
"learning_rate": 5.221415699930952e-07, |
|
"loss": 0.6326, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.475955610357583, |
|
"grad_norm": 0.09523724764585495, |
|
"learning_rate": 5.126862012046551e-07, |
|
"loss": 0.6295, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 2.4808877928483355, |
|
"grad_norm": 0.17275214195251465, |
|
"learning_rate": 5.03307438588819e-07, |
|
"loss": 0.647, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 2.4858199753390875, |
|
"grad_norm": 0.3728453516960144, |
|
"learning_rate": 4.940056436743534e-07, |
|
"loss": 0.6419, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 2.4907521578298395, |
|
"grad_norm": 0.09702899307012558, |
|
"learning_rate": 4.847811750231057e-07, |
|
"loss": 0.6374, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 2.495684340320592, |
|
"grad_norm": 0.12215188145637512, |
|
"learning_rate": 4.7563438821618236e-07, |
|
"loss": 0.6408, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 2.500616522811344, |
|
"grad_norm": 0.10780132561922073, |
|
"learning_rate": 4.6656563584023955e-07, |
|
"loss": 0.6191, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 2.505548705302096, |
|
"grad_norm": 0.14112994074821472, |
|
"learning_rate": 4.5757526747389506e-07, |
|
"loss": 0.6381, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 2.5104808877928484, |
|
"grad_norm": 0.09904135763645172, |
|
"learning_rate": 4.4866362967425054e-07, |
|
"loss": 0.6312, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 2.5154130702836004, |
|
"grad_norm": 0.24112163484096527, |
|
"learning_rate": 4.398310659635338e-07, |
|
"loss": 0.6271, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 2.5203452527743524, |
|
"grad_norm": 0.14225369691848755, |
|
"learning_rate": 4.310779168158566e-07, |
|
"loss": 0.6423, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.525277435265105, |
|
"grad_norm": 0.13167624175548553, |
|
"learning_rate": 4.2240451964408984e-07, |
|
"loss": 0.6338, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 2.530209617755857, |
|
"grad_norm": 0.10630662739276886, |
|
"learning_rate": 4.138112087868576e-07, |
|
"loss": 0.6255, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 2.5351418002466093, |
|
"grad_norm": 0.10379086434841156, |
|
"learning_rate": 4.052983154956483e-07, |
|
"loss": 0.6264, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 2.5400739827373613, |
|
"grad_norm": 0.09079194813966751, |
|
"learning_rate": 3.9686616792204677e-07, |
|
"loss": 0.5999, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 2.5450061652281133, |
|
"grad_norm": 0.09570661187171936, |
|
"learning_rate": 3.885150911050856e-07, |
|
"loss": 0.61, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 2.5499383477188657, |
|
"grad_norm": 0.09997903555631638, |
|
"learning_rate": 3.8024540695871275e-07, |
|
"loss": 0.6391, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 2.5548705302096177, |
|
"grad_norm": 0.10698223114013672, |
|
"learning_rate": 3.720574342593847e-07, |
|
"loss": 0.6197, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 2.55980271270037, |
|
"grad_norm": 0.13335859775543213, |
|
"learning_rate": 3.639514886337786e-07, |
|
"loss": 0.6507, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 2.564734895191122, |
|
"grad_norm": 0.2598697245121002, |
|
"learning_rate": 3.559278825466245e-07, |
|
"loss": 0.6284, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 2.569667077681874, |
|
"grad_norm": 0.19334684312343597, |
|
"learning_rate": 3.4798692528866057e-07, |
|
"loss": 0.631, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.5745992601726266, |
|
"grad_norm": 0.1008177101612091, |
|
"learning_rate": 3.4012892296471173e-07, |
|
"loss": 0.6651, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 2.5795314426633786, |
|
"grad_norm": 0.11722905933856964, |
|
"learning_rate": 3.3235417848188985e-07, |
|
"loss": 0.6524, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 2.5844636251541306, |
|
"grad_norm": 0.12759149074554443, |
|
"learning_rate": 3.2466299153791626e-07, |
|
"loss": 0.6508, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 2.589395807644883, |
|
"grad_norm": 0.12765510380268097, |
|
"learning_rate": 3.1705565860956994e-07, |
|
"loss": 0.636, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 2.594327990135635, |
|
"grad_norm": 0.10414116829633713, |
|
"learning_rate": 3.095324729412602e-07, |
|
"loss": 0.6192, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 2.599260172626387, |
|
"grad_norm": 0.28323763608932495, |
|
"learning_rate": 3.020937245337208e-07, |
|
"loss": 0.6614, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 2.6041923551171395, |
|
"grad_norm": 0.19360347092151642, |
|
"learning_rate": 2.947397001328314e-07, |
|
"loss": 0.6444, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 2.6091245376078915, |
|
"grad_norm": 0.10508602857589722, |
|
"learning_rate": 2.874706832185656e-07, |
|
"loss": 0.6113, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 2.6140567200986435, |
|
"grad_norm": 0.10985539853572845, |
|
"learning_rate": 2.80286953994062e-07, |
|
"loss": 0.6372, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 2.618988902589396, |
|
"grad_norm": 0.11580062657594681, |
|
"learning_rate": 2.731887893748242e-07, |
|
"loss": 0.6274, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.623921085080148, |
|
"grad_norm": 0.10763997584581375, |
|
"learning_rate": 2.6617646297804554e-07, |
|
"loss": 0.6345, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 2.6288532675709, |
|
"grad_norm": 0.10293387621641159, |
|
"learning_rate": 2.5925024511206207e-07, |
|
"loss": 0.6032, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 2.6337854500616524, |
|
"grad_norm": 0.1057887151837349, |
|
"learning_rate": 2.52410402765933e-07, |
|
"loss": 0.6594, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 2.6387176325524044, |
|
"grad_norm": 0.11511359363794327, |
|
"learning_rate": 2.45657199599148e-07, |
|
"loss": 0.6283, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 2.6436498150431564, |
|
"grad_norm": 0.10298167169094086, |
|
"learning_rate": 2.389908959314663e-07, |
|
"loss": 0.6194, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 2.648581997533909, |
|
"grad_norm": 0.13076511025428772, |
|
"learning_rate": 2.3241174873287892e-07, |
|
"loss": 0.6128, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 2.653514180024661, |
|
"grad_norm": 0.10795601457357407, |
|
"learning_rate": 2.2592001161370392e-07, |
|
"loss": 0.6604, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 2.658446362515413, |
|
"grad_norm": 0.08846652507781982, |
|
"learning_rate": 2.1951593481481236e-07, |
|
"loss": 0.5985, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 2.6633785450061653, |
|
"grad_norm": 0.14082390069961548, |
|
"learning_rate": 2.1319976519797862e-07, |
|
"loss": 0.6444, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 2.6683107274969173, |
|
"grad_norm": 0.1039619892835617, |
|
"learning_rate": 2.0697174623636795e-07, |
|
"loss": 0.6735, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.6732429099876693, |
|
"grad_norm": 0.09636316448450089, |
|
"learning_rate": 2.0083211800514868e-07, |
|
"loss": 0.6311, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 2.678175092478422, |
|
"grad_norm": 0.10067697614431381, |
|
"learning_rate": 1.9478111717223968e-07, |
|
"loss": 0.6237, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 2.683107274969174, |
|
"grad_norm": 0.11034560203552246, |
|
"learning_rate": 1.8881897698918544e-07, |
|
"loss": 0.6247, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 2.688039457459926, |
|
"grad_norm": 0.11042781919240952, |
|
"learning_rate": 1.8294592728216764e-07, |
|
"loss": 0.6413, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 2.6929716399506782, |
|
"grad_norm": 0.1032506600022316, |
|
"learning_rate": 1.7716219444314204e-07, |
|
"loss": 0.6604, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 2.6979038224414302, |
|
"grad_norm": 0.09196452796459198, |
|
"learning_rate": 1.7146800142111536e-07, |
|
"loss": 0.6503, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 2.7028360049321822, |
|
"grad_norm": 0.09848834574222565, |
|
"learning_rate": 1.658635677135484e-07, |
|
"loss": 0.6195, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 2.7077681874229347, |
|
"grad_norm": 0.12145579606294632, |
|
"learning_rate": 1.6034910935789628e-07, |
|
"loss": 0.6269, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 2.7127003699136867, |
|
"grad_norm": 0.10812917351722717, |
|
"learning_rate": 1.5492483892328104e-07, |
|
"loss": 0.6422, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 2.717632552404439, |
|
"grad_norm": 0.10012295842170715, |
|
"learning_rate": 1.4959096550229645e-07, |
|
"loss": 0.6266, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.722564734895191, |
|
"grad_norm": 0.24410749971866608, |
|
"learning_rate": 1.44347694702949e-07, |
|
"loss": 0.6406, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 2.727496917385943, |
|
"grad_norm": 0.11757177114486694, |
|
"learning_rate": 1.391952286407311e-07, |
|
"loss": 0.6446, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 2.7324290998766956, |
|
"grad_norm": 0.11057958751916885, |
|
"learning_rate": 1.341337659308309e-07, |
|
"loss": 0.6347, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 2.7373612823674476, |
|
"grad_norm": 0.11715658754110336, |
|
"learning_rate": 1.291635016804768e-07, |
|
"loss": 0.6508, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 2.7422934648582, |
|
"grad_norm": 0.10512126982212067, |
|
"learning_rate": 1.2428462748141523e-07, |
|
"loss": 0.6509, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 2.747225647348952, |
|
"grad_norm": 0.20344923436641693, |
|
"learning_rate": 1.1949733140252468e-07, |
|
"loss": 0.619, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 2.752157829839704, |
|
"grad_norm": 0.13411924242973328, |
|
"learning_rate": 1.1480179798256857e-07, |
|
"loss": 0.6254, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 2.7570900123304565, |
|
"grad_norm": 0.09810943156480789, |
|
"learning_rate": 1.1019820822307986e-07, |
|
"loss": 0.6301, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 2.7620221948212085, |
|
"grad_norm": 0.14971470832824707, |
|
"learning_rate": 1.056867395813832e-07, |
|
"loss": 0.6325, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 2.7669543773119605, |
|
"grad_norm": 0.11750344932079315, |
|
"learning_rate": 1.0126756596375687e-07, |
|
"loss": 0.6256, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.771886559802713, |
|
"grad_norm": 0.125830739736557, |
|
"learning_rate": 9.694085771872697e-08, |
|
"loss": 0.6014, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 2.776818742293465, |
|
"grad_norm": 0.16469189524650574, |
|
"learning_rate": 9.270678163050218e-08, |
|
"loss": 0.6491, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 2.781750924784217, |
|
"grad_norm": 0.12204600125551224, |
|
"learning_rate": 8.856550091254302e-08, |
|
"loss": 0.6472, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 2.7866831072749694, |
|
"grad_norm": 0.11357175558805466, |
|
"learning_rate": 8.451717520127272e-08, |
|
"loss": 0.6447, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 2.7916152897657214, |
|
"grad_norm": 0.2082013189792633, |
|
"learning_rate": 8.056196054992193e-08, |
|
"loss": 0.6686, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 2.7965474722564734, |
|
"grad_norm": 0.09912417829036713, |
|
"learning_rate": 7.670000942251288e-08, |
|
"loss": 0.6365, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 2.801479654747226, |
|
"grad_norm": 0.3408164381980896, |
|
"learning_rate": 7.293147068798384e-08, |
|
"loss": 0.7199, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 2.806411837237978, |
|
"grad_norm": 0.18314948678016663, |
|
"learning_rate": 6.92564896144493e-08, |
|
"loss": 0.6531, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 2.81134401972873, |
|
"grad_norm": 0.0905625969171524, |
|
"learning_rate": 6.56752078636011e-08, |
|
"loss": 0.6495, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 2.8162762022194823, |
|
"grad_norm": 0.10962024331092834, |
|
"learning_rate": 6.218776348524663e-08, |
|
"loss": 0.629, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.8212083847102343, |
|
"grad_norm": 0.1477546989917755, |
|
"learning_rate": 5.879429091198846e-08, |
|
"loss": 0.6415, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 2.8261405672009863, |
|
"grad_norm": 0.2878320515155792, |
|
"learning_rate": 5.549492095404202e-08, |
|
"loss": 0.6354, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 2.8310727496917387, |
|
"grad_norm": 0.09865111857652664, |
|
"learning_rate": 5.2289780794192726e-08, |
|
"loss": 0.6333, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 2.8360049321824907, |
|
"grad_norm": 0.11314819753170013, |
|
"learning_rate": 4.917899398289378e-08, |
|
"loss": 0.6353, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 2.8409371146732427, |
|
"grad_norm": 0.13539192080497742, |
|
"learning_rate": 4.6162680433503024e-08, |
|
"loss": 0.6449, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 2.845869297163995, |
|
"grad_norm": 0.11400008946657181, |
|
"learning_rate": 4.3240956417661685e-08, |
|
"loss": 0.5971, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 2.850801479654747, |
|
"grad_norm": 0.1335747390985489, |
|
"learning_rate": 4.0413934560811216e-08, |
|
"loss": 0.6234, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 2.855733662145499, |
|
"grad_norm": 1.7575569152832031, |
|
"learning_rate": 3.768172383785268e-08, |
|
"loss": 0.6393, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 2.8606658446362516, |
|
"grad_norm": 0.20131815969944, |
|
"learning_rate": 3.504442956894533e-08, |
|
"loss": 0.6389, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 2.8655980271270036, |
|
"grad_norm": 0.10285453498363495, |
|
"learning_rate": 3.250215341544766e-08, |
|
"loss": 0.6129, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.8705302096177556, |
|
"grad_norm": 0.11187023669481277, |
|
"learning_rate": 3.005499337599777e-08, |
|
"loss": 0.622, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 2.875462392108508, |
|
"grad_norm": 0.10102323442697525, |
|
"learning_rate": 2.7703043782735527e-08, |
|
"loss": 0.6667, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 2.88039457459926, |
|
"grad_norm": 0.16657452285289764, |
|
"learning_rate": 2.544639529766829e-08, |
|
"loss": 0.6302, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 2.885326757090012, |
|
"grad_norm": 0.09614276140928268, |
|
"learning_rate": 2.3285134909173113e-08, |
|
"loss": 0.614, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 2.8902589395807645, |
|
"grad_norm": 0.17049086093902588, |
|
"learning_rate": 2.1219345928646107e-08, |
|
"loss": 0.6179, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 2.8951911220715165, |
|
"grad_norm": 0.11277743428945541, |
|
"learning_rate": 1.924910798728946e-08, |
|
"loss": 0.6397, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 2.900123304562269, |
|
"grad_norm": 0.1780087947845459, |
|
"learning_rate": 1.7374497033042504e-08, |
|
"loss": 0.6358, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 2.905055487053021, |
|
"grad_norm": 0.33896899223327637, |
|
"learning_rate": 1.559558532765404e-08, |
|
"loss": 0.6364, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 2.909987669543773, |
|
"grad_norm": 0.11533108353614807, |
|
"learning_rate": 1.3912441443896529e-08, |
|
"loss": 0.6219, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 2.9149198520345254, |
|
"grad_norm": 0.2513286769390106, |
|
"learning_rate": 1.2325130262923202e-08, |
|
"loss": 0.6544, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.9198520345252774, |
|
"grad_norm": 0.10646089911460876, |
|
"learning_rate": 1.0833712971766442e-08, |
|
"loss": 0.6314, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 2.92478421701603, |
|
"grad_norm": 0.18473948538303375, |
|
"learning_rate": 9.438247060979954e-09, |
|
"loss": 0.6272, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 2.929716399506782, |
|
"grad_norm": 0.17810285091400146, |
|
"learning_rate": 8.13878632242221e-09, |
|
"loss": 0.6551, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 2.934648581997534, |
|
"grad_norm": 0.10676706582307816, |
|
"learning_rate": 6.935380847182815e-09, |
|
"loss": 0.6505, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 2.9395807644882863, |
|
"grad_norm": 0.1609431952238083, |
|
"learning_rate": 5.828077023651846e-09, |
|
"loss": 0.6403, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 2.9445129469790383, |
|
"grad_norm": 0.3682437539100647, |
|
"learning_rate": 4.816917535731547e-09, |
|
"loss": 0.6535, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 2.9494451294697903, |
|
"grad_norm": 0.08937019854784012, |
|
"learning_rate": 3.9019413611907084e-09, |
|
"loss": 0.6422, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 2.9543773119605428, |
|
"grad_norm": 0.09343031048774719, |
|
"learning_rate": 3.083183770162812e-09, |
|
"loss": 0.6454, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 2.9593094944512948, |
|
"grad_norm": 0.11115273088216782, |
|
"learning_rate": 2.360676323786282e-09, |
|
"loss": 0.6167, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 2.9642416769420468, |
|
"grad_norm": 0.09774809330701828, |
|
"learning_rate": 1.7344468729868502e-09, |
|
"loss": 0.6422, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.969173859432799, |
|
"grad_norm": 0.10831815004348755, |
|
"learning_rate": 1.2045195574056367e-09, |
|
"loss": 0.6387, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 2.974106041923551, |
|
"grad_norm": 0.22296251356601715, |
|
"learning_rate": 7.709148044679482e-10, |
|
"loss": 0.6415, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 2.979038224414303, |
|
"grad_norm": 0.10190446674823761, |
|
"learning_rate": 4.3364932859474293e-10, |
|
"loss": 0.6286, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 2.9839704069050557, |
|
"grad_norm": 0.15854156017303467, |
|
"learning_rate": 1.9273613056008945e-10, |
|
"loss": 0.6051, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 2.9889025893958077, |
|
"grad_norm": 0.09298679232597351, |
|
"learning_rate": 4.8184496989067684e-11, |
|
"loss": 0.6213, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 2.9938347718865597, |
|
"grad_norm": 0.2940804958343506, |
|
"learning_rate": 0.0, |
|
"loss": 0.6347, |
|
"step": 606 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 606, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 101, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.765503957851418e+20, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|