dannh's picture
Upload folder using huggingface_hub
b325709 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 11.928792569659443,
"eval_steps": 500,
"global_step": 1932,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006191950464396285,
"grad_norm": 583.2729636731592,
"learning_rate": 0.0,
"loss": 16.1607,
"step": 1
},
{
"epoch": 0.01238390092879257,
"grad_norm": 510.1711028527366,
"learning_rate": 1.724137931034483e-06,
"loss": 15.6126,
"step": 2
},
{
"epoch": 0.018575851393188854,
"grad_norm": 432.5760222925412,
"learning_rate": 3.448275862068966e-06,
"loss": 15.1224,
"step": 3
},
{
"epoch": 0.02476780185758514,
"grad_norm": 227.61889624860476,
"learning_rate": 5.172413793103448e-06,
"loss": 13.282,
"step": 4
},
{
"epoch": 0.030959752321981424,
"grad_norm": 158.07575641446655,
"learning_rate": 6.896551724137932e-06,
"loss": 12.5074,
"step": 5
},
{
"epoch": 0.03715170278637771,
"grad_norm": 154.3189118557477,
"learning_rate": 8.620689655172414e-06,
"loss": 12.0762,
"step": 6
},
{
"epoch": 0.043343653250773995,
"grad_norm": 76.42575226139347,
"learning_rate": 1.0344827586206897e-05,
"loss": 10.4845,
"step": 7
},
{
"epoch": 0.04953560371517028,
"grad_norm": 74.16335719258294,
"learning_rate": 1.206896551724138e-05,
"loss": 9.0803,
"step": 8
},
{
"epoch": 0.05572755417956656,
"grad_norm": 74.93338825694016,
"learning_rate": 1.3793103448275863e-05,
"loss": 8.8984,
"step": 9
},
{
"epoch": 0.06191950464396285,
"grad_norm": 72.98575970833309,
"learning_rate": 1.5517241379310346e-05,
"loss": 8.4824,
"step": 10
},
{
"epoch": 0.06811145510835913,
"grad_norm": 60.95311182317466,
"learning_rate": 1.7241379310344828e-05,
"loss": 8.0288,
"step": 11
},
{
"epoch": 0.07430340557275542,
"grad_norm": 25.949986997829786,
"learning_rate": 1.896551724137931e-05,
"loss": 7.4985,
"step": 12
},
{
"epoch": 0.0804953560371517,
"grad_norm": 48.812709487118596,
"learning_rate": 2.0689655172413793e-05,
"loss": 7.1892,
"step": 13
},
{
"epoch": 0.08668730650154799,
"grad_norm": 17.975953609206936,
"learning_rate": 2.2413793103448276e-05,
"loss": 7.2859,
"step": 14
},
{
"epoch": 0.09287925696594428,
"grad_norm": 13.444400027966196,
"learning_rate": 2.413793103448276e-05,
"loss": 6.8293,
"step": 15
},
{
"epoch": 0.09907120743034056,
"grad_norm": 16.906524456262332,
"learning_rate": 2.5862068965517244e-05,
"loss": 6.5689,
"step": 16
},
{
"epoch": 0.10526315789473684,
"grad_norm": 14.96615376509466,
"learning_rate": 2.7586206896551727e-05,
"loss": 6.477,
"step": 17
},
{
"epoch": 0.11145510835913312,
"grad_norm": 33.3089382115632,
"learning_rate": 2.9310344827586206e-05,
"loss": 6.4738,
"step": 18
},
{
"epoch": 0.11764705882352941,
"grad_norm": 132.30119742848726,
"learning_rate": 3.103448275862069e-05,
"loss": 6.1718,
"step": 19
},
{
"epoch": 0.1238390092879257,
"grad_norm": 14.885009490288567,
"learning_rate": 3.275862068965517e-05,
"loss": 5.9924,
"step": 20
},
{
"epoch": 0.13003095975232198,
"grad_norm": 14.20583957199301,
"learning_rate": 3.4482758620689657e-05,
"loss": 5.8076,
"step": 21
},
{
"epoch": 0.13622291021671826,
"grad_norm": 14.052226006442716,
"learning_rate": 3.620689655172414e-05,
"loss": 5.013,
"step": 22
},
{
"epoch": 0.14241486068111456,
"grad_norm": 13.91506279716363,
"learning_rate": 3.793103448275862e-05,
"loss": 5.3075,
"step": 23
},
{
"epoch": 0.14860681114551083,
"grad_norm": 12.62823248730767,
"learning_rate": 3.965517241379311e-05,
"loss": 5.2177,
"step": 24
},
{
"epoch": 0.15479876160990713,
"grad_norm": 19.04525031939675,
"learning_rate": 4.1379310344827587e-05,
"loss": 4.9121,
"step": 25
},
{
"epoch": 0.1609907120743034,
"grad_norm": 10.728631362957593,
"learning_rate": 4.3103448275862066e-05,
"loss": 4.6596,
"step": 26
},
{
"epoch": 0.16718266253869968,
"grad_norm": 13.956262307265314,
"learning_rate": 4.482758620689655e-05,
"loss": 4.7653,
"step": 27
},
{
"epoch": 0.17337461300309598,
"grad_norm": 10.083505632824219,
"learning_rate": 4.655172413793104e-05,
"loss": 4.4974,
"step": 28
},
{
"epoch": 0.17956656346749225,
"grad_norm": 9.936223155180423,
"learning_rate": 4.827586206896552e-05,
"loss": 4.7099,
"step": 29
},
{
"epoch": 0.18575851393188855,
"grad_norm": 13.494395241748343,
"learning_rate": 5e-05,
"loss": 4.4682,
"step": 30
},
{
"epoch": 0.19195046439628483,
"grad_norm": 9.751210714554487,
"learning_rate": 5.172413793103449e-05,
"loss": 4.6888,
"step": 31
},
{
"epoch": 0.19814241486068113,
"grad_norm": 8.53427701687458,
"learning_rate": 5.344827586206896e-05,
"loss": 4.3002,
"step": 32
},
{
"epoch": 0.2043343653250774,
"grad_norm": 18.845223437943833,
"learning_rate": 5.517241379310345e-05,
"loss": 4.5649,
"step": 33
},
{
"epoch": 0.21052631578947367,
"grad_norm": 14.037468185845768,
"learning_rate": 5.689655172413794e-05,
"loss": 4.1247,
"step": 34
},
{
"epoch": 0.21671826625386997,
"grad_norm": 10.461800393764017,
"learning_rate": 5.862068965517241e-05,
"loss": 4.173,
"step": 35
},
{
"epoch": 0.22291021671826625,
"grad_norm": 13.442613201138604,
"learning_rate": 6.03448275862069e-05,
"loss": 4.5809,
"step": 36
},
{
"epoch": 0.22910216718266255,
"grad_norm": 9.464361097873487,
"learning_rate": 6.206896551724138e-05,
"loss": 4.1074,
"step": 37
},
{
"epoch": 0.23529411764705882,
"grad_norm": 9.665048525162632,
"learning_rate": 6.379310344827587e-05,
"loss": 4.1199,
"step": 38
},
{
"epoch": 0.24148606811145512,
"grad_norm": 9.659469642364447,
"learning_rate": 6.551724137931034e-05,
"loss": 4.1702,
"step": 39
},
{
"epoch": 0.2476780185758514,
"grad_norm": 8.152823124328645,
"learning_rate": 6.724137931034483e-05,
"loss": 3.9314,
"step": 40
},
{
"epoch": 0.25386996904024767,
"grad_norm": 8.452741968127802,
"learning_rate": 6.896551724137931e-05,
"loss": 3.8782,
"step": 41
},
{
"epoch": 0.26006191950464397,
"grad_norm": 10.437508209318183,
"learning_rate": 7.06896551724138e-05,
"loss": 4.2638,
"step": 42
},
{
"epoch": 0.26625386996904027,
"grad_norm": 9.474156238032947,
"learning_rate": 7.241379310344828e-05,
"loss": 3.9874,
"step": 43
},
{
"epoch": 0.2724458204334365,
"grad_norm": 7.638324788974118,
"learning_rate": 7.413793103448277e-05,
"loss": 3.8656,
"step": 44
},
{
"epoch": 0.2786377708978328,
"grad_norm": 8.32912958344573,
"learning_rate": 7.586206896551724e-05,
"loss": 3.812,
"step": 45
},
{
"epoch": 0.2848297213622291,
"grad_norm": 9.374949463316568,
"learning_rate": 7.758620689655173e-05,
"loss": 3.7579,
"step": 46
},
{
"epoch": 0.29102167182662536,
"grad_norm": 11.53047562375587,
"learning_rate": 7.931034482758621e-05,
"loss": 3.8599,
"step": 47
},
{
"epoch": 0.29721362229102166,
"grad_norm": 7.0175519843629,
"learning_rate": 8.103448275862069e-05,
"loss": 3.6109,
"step": 48
},
{
"epoch": 0.30340557275541796,
"grad_norm": 6.559697059667157,
"learning_rate": 8.275862068965517e-05,
"loss": 3.6014,
"step": 49
},
{
"epoch": 0.30959752321981426,
"grad_norm": 6.717777404862978,
"learning_rate": 8.448275862068966e-05,
"loss": 3.7019,
"step": 50
},
{
"epoch": 0.3157894736842105,
"grad_norm": 5.982207393646023,
"learning_rate": 8.620689655172413e-05,
"loss": 3.4398,
"step": 51
},
{
"epoch": 0.3219814241486068,
"grad_norm": 5.945506225475662,
"learning_rate": 8.793103448275862e-05,
"loss": 3.4582,
"step": 52
},
{
"epoch": 0.3281733746130031,
"grad_norm": 5.704596615160976,
"learning_rate": 8.96551724137931e-05,
"loss": 3.6077,
"step": 53
},
{
"epoch": 0.33436532507739936,
"grad_norm": 5.973311803401044,
"learning_rate": 9.137931034482759e-05,
"loss": 3.5219,
"step": 54
},
{
"epoch": 0.34055727554179566,
"grad_norm": 6.814299767123095,
"learning_rate": 9.310344827586207e-05,
"loss": 3.7694,
"step": 55
},
{
"epoch": 0.34674922600619196,
"grad_norm": 7.052917863988468,
"learning_rate": 9.482758620689656e-05,
"loss": 3.6753,
"step": 56
},
{
"epoch": 0.35294117647058826,
"grad_norm": 5.525773749757123,
"learning_rate": 9.655172413793105e-05,
"loss": 3.5596,
"step": 57
},
{
"epoch": 0.3591331269349845,
"grad_norm": 5.3240485466656295,
"learning_rate": 9.827586206896552e-05,
"loss": 3.3443,
"step": 58
},
{
"epoch": 0.3653250773993808,
"grad_norm": 5.475602381517988,
"learning_rate": 0.0001,
"loss": 3.5009,
"step": 59
},
{
"epoch": 0.3715170278637771,
"grad_norm": 5.892096476556127,
"learning_rate": 9.999992974124023e-05,
"loss": 3.5758,
"step": 60
},
{
"epoch": 0.37770897832817335,
"grad_norm": 4.838516017994781,
"learning_rate": 9.999971896515837e-05,
"loss": 3.3336,
"step": 61
},
{
"epoch": 0.38390092879256965,
"grad_norm": 5.246620602818011,
"learning_rate": 9.999936767234674e-05,
"loss": 3.4909,
"step": 62
},
{
"epoch": 0.39009287925696595,
"grad_norm": 6.760983626210936,
"learning_rate": 9.999887586379264e-05,
"loss": 3.6338,
"step": 63
},
{
"epoch": 0.39628482972136225,
"grad_norm": 6.012748270095294,
"learning_rate": 9.99982435408782e-05,
"loss": 3.586,
"step": 64
},
{
"epoch": 0.4024767801857585,
"grad_norm": 6.5822760683818515,
"learning_rate": 9.999747070538049e-05,
"loss": 3.8235,
"step": 65
},
{
"epoch": 0.4086687306501548,
"grad_norm": 4.95793646000932,
"learning_rate": 9.999655735947144e-05,
"loss": 3.5195,
"step": 66
},
{
"epoch": 0.4148606811145511,
"grad_norm": 5.096298290945092,
"learning_rate": 9.999550350571785e-05,
"loss": 3.5608,
"step": 67
},
{
"epoch": 0.42105263157894735,
"grad_norm": 5.023897068332846,
"learning_rate": 9.999430914708143e-05,
"loss": 3.4636,
"step": 68
},
{
"epoch": 0.42724458204334365,
"grad_norm": 5.0162625600997695,
"learning_rate": 9.999297428691878e-05,
"loss": 3.4252,
"step": 69
},
{
"epoch": 0.43343653250773995,
"grad_norm": 5.347697718620594,
"learning_rate": 9.999149892898127e-05,
"loss": 3.4712,
"step": 70
},
{
"epoch": 0.43962848297213625,
"grad_norm": 5.277743291050418,
"learning_rate": 9.998988307741521e-05,
"loss": 3.3179,
"step": 71
},
{
"epoch": 0.4458204334365325,
"grad_norm": 5.092008006469729,
"learning_rate": 9.99881267367617e-05,
"loss": 3.2156,
"step": 72
},
{
"epoch": 0.4520123839009288,
"grad_norm": 5.602962681853112,
"learning_rate": 9.998622991195668e-05,
"loss": 3.6088,
"step": 73
},
{
"epoch": 0.4582043343653251,
"grad_norm": 4.618945934546075,
"learning_rate": 9.998419260833087e-05,
"loss": 3.401,
"step": 74
},
{
"epoch": 0.46439628482972134,
"grad_norm": 4.407828092310794,
"learning_rate": 9.998201483160981e-05,
"loss": 3.3034,
"step": 75
},
{
"epoch": 0.47058823529411764,
"grad_norm": 5.060837334792514,
"learning_rate": 9.997969658791384e-05,
"loss": 3.4433,
"step": 76
},
{
"epoch": 0.47678018575851394,
"grad_norm": 4.8844393650400475,
"learning_rate": 9.997723788375803e-05,
"loss": 3.4946,
"step": 77
},
{
"epoch": 0.48297213622291024,
"grad_norm": 4.739079596692814,
"learning_rate": 9.997463872605217e-05,
"loss": 3.5946,
"step": 78
},
{
"epoch": 0.4891640866873065,
"grad_norm": 4.180305626117479,
"learning_rate": 9.997189912210085e-05,
"loss": 3.3336,
"step": 79
},
{
"epoch": 0.4953560371517028,
"grad_norm": 4.4496272009926665,
"learning_rate": 9.996901907960329e-05,
"loss": 3.3069,
"step": 80
},
{
"epoch": 0.5015479876160991,
"grad_norm": 4.839247704934278,
"learning_rate": 9.99659986066534e-05,
"loss": 3.456,
"step": 81
},
{
"epoch": 0.5077399380804953,
"grad_norm": 4.096642605457276,
"learning_rate": 9.996283771173982e-05,
"loss": 3.3263,
"step": 82
},
{
"epoch": 0.5139318885448917,
"grad_norm": 5.436181697917328,
"learning_rate": 9.995953640374573e-05,
"loss": 3.4265,
"step": 83
},
{
"epoch": 0.5201238390092879,
"grad_norm": 4.030576988739356,
"learning_rate": 9.995609469194897e-05,
"loss": 3.2658,
"step": 84
},
{
"epoch": 0.5263157894736842,
"grad_norm": 4.030663123263123,
"learning_rate": 9.995251258602199e-05,
"loss": 3.2758,
"step": 85
},
{
"epoch": 0.5325077399380805,
"grad_norm": 4.482470637813161,
"learning_rate": 9.99487900960317e-05,
"loss": 3.2483,
"step": 86
},
{
"epoch": 0.5386996904024768,
"grad_norm": 4.652325115807756,
"learning_rate": 9.994492723243965e-05,
"loss": 3.3789,
"step": 87
},
{
"epoch": 0.544891640866873,
"grad_norm": 4.9472490461802145,
"learning_rate": 9.994092400610182e-05,
"loss": 3.4134,
"step": 88
},
{
"epoch": 0.5510835913312694,
"grad_norm": 5.586407428655971,
"learning_rate": 9.993678042826868e-05,
"loss": 3.4805,
"step": 89
},
{
"epoch": 0.5572755417956656,
"grad_norm": 4.060091954707022,
"learning_rate": 9.993249651058513e-05,
"loss": 3.2706,
"step": 90
},
{
"epoch": 0.5634674922600619,
"grad_norm": 4.208423705318374,
"learning_rate": 9.992807226509051e-05,
"loss": 3.2509,
"step": 91
},
{
"epoch": 0.5696594427244582,
"grad_norm": 4.259213911444956,
"learning_rate": 9.992350770421848e-05,
"loss": 3.3747,
"step": 92
},
{
"epoch": 0.5758513931888545,
"grad_norm": 4.0529186511539885,
"learning_rate": 9.991880284079704e-05,
"loss": 3.3591,
"step": 93
},
{
"epoch": 0.5820433436532507,
"grad_norm": 4.989718273873512,
"learning_rate": 9.991395768804852e-05,
"loss": 3.2976,
"step": 94
},
{
"epoch": 0.5882352941176471,
"grad_norm": 3.7531790113696353,
"learning_rate": 9.990897225958951e-05,
"loss": 3.0671,
"step": 95
},
{
"epoch": 0.5944272445820433,
"grad_norm": 4.578086243141991,
"learning_rate": 9.990384656943079e-05,
"loss": 3.3136,
"step": 96
},
{
"epoch": 0.6006191950464397,
"grad_norm": 4.239806465153399,
"learning_rate": 9.989858063197735e-05,
"loss": 3.1344,
"step": 97
},
{
"epoch": 0.6068111455108359,
"grad_norm": 5.296536247781439,
"learning_rate": 9.989317446202832e-05,
"loss": 3.3383,
"step": 98
},
{
"epoch": 0.6130030959752322,
"grad_norm": 4.833348660562072,
"learning_rate": 9.988762807477693e-05,
"loss": 3.244,
"step": 99
},
{
"epoch": 0.6191950464396285,
"grad_norm": 4.148795233753714,
"learning_rate": 9.988194148581048e-05,
"loss": 3.2728,
"step": 100
},
{
"epoch": 0.6253869969040248,
"grad_norm": 4.269419850788365,
"learning_rate": 9.987611471111027e-05,
"loss": 3.1576,
"step": 101
},
{
"epoch": 0.631578947368421,
"grad_norm": 4.494909466750201,
"learning_rate": 9.98701477670516e-05,
"loss": 3.3776,
"step": 102
},
{
"epoch": 0.6377708978328174,
"grad_norm": 4.12371607549373,
"learning_rate": 9.986404067040363e-05,
"loss": 3.2136,
"step": 103
},
{
"epoch": 0.6439628482972136,
"grad_norm": 4.1066986326075545,
"learning_rate": 9.985779343832947e-05,
"loss": 3.342,
"step": 104
},
{
"epoch": 0.6501547987616099,
"grad_norm": 3.411792974866172,
"learning_rate": 9.985140608838604e-05,
"loss": 3.0947,
"step": 105
},
{
"epoch": 0.6563467492260062,
"grad_norm": 3.784426944839016,
"learning_rate": 9.984487863852401e-05,
"loss": 3.0899,
"step": 106
},
{
"epoch": 0.6625386996904025,
"grad_norm": 4.337957539267078,
"learning_rate": 9.98382111070878e-05,
"loss": 3.1851,
"step": 107
},
{
"epoch": 0.6687306501547987,
"grad_norm": 4.6165406721174955,
"learning_rate": 9.983140351281554e-05,
"loss": 3.2591,
"step": 108
},
{
"epoch": 0.6749226006191951,
"grad_norm": 3.895886556884958,
"learning_rate": 9.982445587483892e-05,
"loss": 3.074,
"step": 109
},
{
"epoch": 0.6811145510835913,
"grad_norm": 5.790495625569918,
"learning_rate": 9.981736821268325e-05,
"loss": 3.4478,
"step": 110
},
{
"epoch": 0.6873065015479877,
"grad_norm": 4.040001891109945,
"learning_rate": 9.981014054626735e-05,
"loss": 3.1273,
"step": 111
},
{
"epoch": 0.6934984520123839,
"grad_norm": 4.9533963229924725,
"learning_rate": 9.980277289590349e-05,
"loss": 3.2547,
"step": 112
},
{
"epoch": 0.6996904024767802,
"grad_norm": 3.5329595802639893,
"learning_rate": 9.979526528229737e-05,
"loss": 3.1815,
"step": 113
},
{
"epoch": 0.7058823529411765,
"grad_norm": 4.953624964196446,
"learning_rate": 9.978761772654797e-05,
"loss": 3.3241,
"step": 114
},
{
"epoch": 0.7120743034055728,
"grad_norm": 3.538087255596596,
"learning_rate": 9.977983025014764e-05,
"loss": 3.1501,
"step": 115
},
{
"epoch": 0.718266253869969,
"grad_norm": 4.243934280107101,
"learning_rate": 9.977190287498191e-05,
"loss": 3.1266,
"step": 116
},
{
"epoch": 0.7244582043343654,
"grad_norm": 4.2764939685444325,
"learning_rate": 9.976383562332944e-05,
"loss": 3.273,
"step": 117
},
{
"epoch": 0.7306501547987616,
"grad_norm": 4.97238240181323,
"learning_rate": 9.975562851786211e-05,
"loss": 3.2663,
"step": 118
},
{
"epoch": 0.7368421052631579,
"grad_norm": 3.887695247258602,
"learning_rate": 9.97472815816447e-05,
"loss": 3.2012,
"step": 119
},
{
"epoch": 0.7430340557275542,
"grad_norm": 3.9487406631272832,
"learning_rate": 9.973879483813506e-05,
"loss": 3.1826,
"step": 120
},
{
"epoch": 0.7492260061919505,
"grad_norm": 3.8144073397429756,
"learning_rate": 9.973016831118389e-05,
"loss": 3.2853,
"step": 121
},
{
"epoch": 0.7554179566563467,
"grad_norm": 4.337511589444347,
"learning_rate": 9.972140202503477e-05,
"loss": 3.122,
"step": 122
},
{
"epoch": 0.7616099071207431,
"grad_norm": 3.993475854912506,
"learning_rate": 9.971249600432402e-05,
"loss": 3.1977,
"step": 123
},
{
"epoch": 0.7678018575851393,
"grad_norm": 3.6866484704267104,
"learning_rate": 9.97034502740807e-05,
"loss": 3.1031,
"step": 124
},
{
"epoch": 0.7739938080495357,
"grad_norm": 3.709528741959154,
"learning_rate": 9.969426485972645e-05,
"loss": 3.0934,
"step": 125
},
{
"epoch": 0.7801857585139319,
"grad_norm": 3.9467902947943148,
"learning_rate": 9.968493978707554e-05,
"loss": 3.2937,
"step": 126
},
{
"epoch": 0.7863777089783281,
"grad_norm": 4.339424645551262,
"learning_rate": 9.967547508233466e-05,
"loss": 3.2433,
"step": 127
},
{
"epoch": 0.7925696594427245,
"grad_norm": 3.8363601911667766,
"learning_rate": 9.966587077210297e-05,
"loss": 3.1673,
"step": 128
},
{
"epoch": 0.7987616099071208,
"grad_norm": 3.750919154659581,
"learning_rate": 9.965612688337194e-05,
"loss": 3.0407,
"step": 129
},
{
"epoch": 0.804953560371517,
"grad_norm": 3.7822587436662016,
"learning_rate": 9.96462434435253e-05,
"loss": 3.1473,
"step": 130
},
{
"epoch": 0.8111455108359134,
"grad_norm": 4.247847819904635,
"learning_rate": 9.963622048033898e-05,
"loss": 3.19,
"step": 131
},
{
"epoch": 0.8173374613003096,
"grad_norm": 3.9735446888529964,
"learning_rate": 9.962605802198104e-05,
"loss": 3.1126,
"step": 132
},
{
"epoch": 0.8235294117647058,
"grad_norm": 3.9880407535617883,
"learning_rate": 9.961575609701154e-05,
"loss": 3.102,
"step": 133
},
{
"epoch": 0.8297213622291022,
"grad_norm": 3.644066398077279,
"learning_rate": 9.960531473438248e-05,
"loss": 3.1088,
"step": 134
},
{
"epoch": 0.8359133126934984,
"grad_norm": 3.772634387387643,
"learning_rate": 9.959473396343777e-05,
"loss": 3.1237,
"step": 135
},
{
"epoch": 0.8421052631578947,
"grad_norm": 3.620434973837521,
"learning_rate": 9.958401381391307e-05,
"loss": 3.1037,
"step": 136
},
{
"epoch": 0.848297213622291,
"grad_norm": 3.3883607787252497,
"learning_rate": 9.957315431593577e-05,
"loss": 3.0548,
"step": 137
},
{
"epoch": 0.8544891640866873,
"grad_norm": 3.525654083685345,
"learning_rate": 9.956215550002485e-05,
"loss": 3.0744,
"step": 138
},
{
"epoch": 0.8606811145510835,
"grad_norm": 3.276172822856916,
"learning_rate": 9.955101739709086e-05,
"loss": 3.107,
"step": 139
},
{
"epoch": 0.8668730650154799,
"grad_norm": 3.7505406680079796,
"learning_rate": 9.953974003843572e-05,
"loss": 3.0649,
"step": 140
},
{
"epoch": 0.8730650154798761,
"grad_norm": 3.4245906115032128,
"learning_rate": 9.952832345575282e-05,
"loss": 3.1105,
"step": 141
},
{
"epoch": 0.8792569659442725,
"grad_norm": 3.999576346323058,
"learning_rate": 9.951676768112672e-05,
"loss": 3.2317,
"step": 142
},
{
"epoch": 0.8854489164086687,
"grad_norm": 6.72936741544915,
"learning_rate": 9.950507274703323e-05,
"loss": 3.2017,
"step": 143
},
{
"epoch": 0.891640866873065,
"grad_norm": 4.253970968740358,
"learning_rate": 9.949323868633916e-05,
"loss": 3.1011,
"step": 144
},
{
"epoch": 0.8978328173374613,
"grad_norm": 6.551804513164246,
"learning_rate": 9.948126553230241e-05,
"loss": 3.1527,
"step": 145
},
{
"epoch": 0.9040247678018576,
"grad_norm": 4.6411829687147526,
"learning_rate": 9.946915331857171e-05,
"loss": 3.1914,
"step": 146
},
{
"epoch": 0.9102167182662538,
"grad_norm": 3.6707162993101052,
"learning_rate": 9.945690207918666e-05,
"loss": 3.0628,
"step": 147
},
{
"epoch": 0.9164086687306502,
"grad_norm": 4.333487880078733,
"learning_rate": 9.944451184857751e-05,
"loss": 3.0723,
"step": 148
},
{
"epoch": 0.9226006191950464,
"grad_norm": 4.643228355107828,
"learning_rate": 9.943198266156516e-05,
"loss": 3.2188,
"step": 149
},
{
"epoch": 0.9287925696594427,
"grad_norm": 3.81900951272679,
"learning_rate": 9.9419314553361e-05,
"loss": 3.1535,
"step": 150
},
{
"epoch": 0.934984520123839,
"grad_norm": 3.483225515541042,
"learning_rate": 9.940650755956686e-05,
"loss": 3.0807,
"step": 151
},
{
"epoch": 0.9411764705882353,
"grad_norm": 4.054952951582781,
"learning_rate": 9.93935617161749e-05,
"loss": 3.2731,
"step": 152
},
{
"epoch": 0.9473684210526315,
"grad_norm": 4.045010851938665,
"learning_rate": 9.938047705956746e-05,
"loss": 3.2919,
"step": 153
},
{
"epoch": 0.9535603715170279,
"grad_norm": 3.880953002409461,
"learning_rate": 9.936725362651699e-05,
"loss": 2.9833,
"step": 154
},
{
"epoch": 0.9597523219814241,
"grad_norm": 3.8222791710005684,
"learning_rate": 9.935389145418599e-05,
"loss": 3.1347,
"step": 155
},
{
"epoch": 0.9659442724458205,
"grad_norm": 3.6318649766832016,
"learning_rate": 9.934039058012686e-05,
"loss": 3.0258,
"step": 156
},
{
"epoch": 0.9721362229102167,
"grad_norm": 4.207757115469884,
"learning_rate": 9.932675104228177e-05,
"loss": 3.189,
"step": 157
},
{
"epoch": 0.978328173374613,
"grad_norm": 3.650959223126377,
"learning_rate": 9.931297287898259e-05,
"loss": 3.1595,
"step": 158
},
{
"epoch": 0.9845201238390093,
"grad_norm": 3.5893527928395668,
"learning_rate": 9.929905612895081e-05,
"loss": 3.1518,
"step": 159
},
{
"epoch": 0.9907120743034056,
"grad_norm": 4.052866492865934,
"learning_rate": 9.928500083129736e-05,
"loss": 2.9796,
"step": 160
},
{
"epoch": 0.9969040247678018,
"grad_norm": 3.4284326492586814,
"learning_rate": 9.927080702552254e-05,
"loss": 2.988,
"step": 161
},
{
"epoch": 1.0,
"grad_norm": 3.4284326492586814,
"learning_rate": 9.925647475151596e-05,
"loss": 1.6511,
"step": 162
},
{
"epoch": 1.0061919504643964,
"grad_norm": 3.6814653270432887,
"learning_rate": 9.924200404955629e-05,
"loss": 2.7978,
"step": 163
},
{
"epoch": 1.0123839009287925,
"grad_norm": 3.1067979115563875,
"learning_rate": 9.922739496031129e-05,
"loss": 2.8784,
"step": 164
},
{
"epoch": 1.0185758513931888,
"grad_norm": 3.811812480430689,
"learning_rate": 9.921264752483761e-05,
"loss": 3.0437,
"step": 165
},
{
"epoch": 1.0247678018575852,
"grad_norm": 3.662249859659687,
"learning_rate": 9.919776178458071e-05,
"loss": 3.0534,
"step": 166
},
{
"epoch": 1.0309597523219813,
"grad_norm": 3.5933991920665718,
"learning_rate": 9.918273778137477e-05,
"loss": 2.9006,
"step": 167
},
{
"epoch": 1.0371517027863777,
"grad_norm": 3.809378521805632,
"learning_rate": 9.916757555744245e-05,
"loss": 2.9678,
"step": 168
},
{
"epoch": 1.043343653250774,
"grad_norm": 3.403865267525689,
"learning_rate": 9.915227515539496e-05,
"loss": 2.9319,
"step": 169
},
{
"epoch": 1.0495356037151702,
"grad_norm": 3.745099532944618,
"learning_rate": 9.913683661823176e-05,
"loss": 3.0997,
"step": 170
},
{
"epoch": 1.0557275541795665,
"grad_norm": 3.545813633080689,
"learning_rate": 9.912125998934055e-05,
"loss": 2.9005,
"step": 171
},
{
"epoch": 1.061919504643963,
"grad_norm": 3.44006824427855,
"learning_rate": 9.910554531249714e-05,
"loss": 3.0533,
"step": 172
},
{
"epoch": 1.068111455108359,
"grad_norm": 4.183934094926584,
"learning_rate": 9.908969263186525e-05,
"loss": 3.0094,
"step": 173
},
{
"epoch": 1.0743034055727554,
"grad_norm": 3.988230695923216,
"learning_rate": 9.907370199199648e-05,
"loss": 2.9598,
"step": 174
},
{
"epoch": 1.0804953560371517,
"grad_norm": 4.058788474570464,
"learning_rate": 9.905757343783014e-05,
"loss": 2.9081,
"step": 175
},
{
"epoch": 1.086687306501548,
"grad_norm": 3.721071325513958,
"learning_rate": 9.904130701469309e-05,
"loss": 3.0445,
"step": 176
},
{
"epoch": 1.0928792569659442,
"grad_norm": 3.507314113135694,
"learning_rate": 9.902490276829971e-05,
"loss": 2.9417,
"step": 177
},
{
"epoch": 1.0990712074303406,
"grad_norm": 3.659862506679075,
"learning_rate": 9.900836074475165e-05,
"loss": 3.0042,
"step": 178
},
{
"epoch": 1.1052631578947367,
"grad_norm": 4.343669608302679,
"learning_rate": 9.899168099053783e-05,
"loss": 2.9413,
"step": 179
},
{
"epoch": 1.111455108359133,
"grad_norm": 3.3781335101832872,
"learning_rate": 9.897486355253419e-05,
"loss": 2.9604,
"step": 180
},
{
"epoch": 1.1176470588235294,
"grad_norm": 3.469752850980135,
"learning_rate": 9.895790847800361e-05,
"loss": 2.9337,
"step": 181
},
{
"epoch": 1.1238390092879258,
"grad_norm": 3.2733115281041543,
"learning_rate": 9.894081581459578e-05,
"loss": 2.9247,
"step": 182
},
{
"epoch": 1.130030959752322,
"grad_norm": 3.2291083717016797,
"learning_rate": 9.892358561034711e-05,
"loss": 2.9189,
"step": 183
},
{
"epoch": 1.1362229102167183,
"grad_norm": 3.176371624508396,
"learning_rate": 9.890621791368049e-05,
"loss": 2.9176,
"step": 184
},
{
"epoch": 1.1424148606811146,
"grad_norm": 3.1779099737432275,
"learning_rate": 9.888871277340522e-05,
"loss": 2.9177,
"step": 185
},
{
"epoch": 1.1486068111455108,
"grad_norm": 3.4931730579286517,
"learning_rate": 9.887107023871691e-05,
"loss": 2.8821,
"step": 186
},
{
"epoch": 1.1547987616099071,
"grad_norm": 3.640445043121747,
"learning_rate": 9.885329035919724e-05,
"loss": 2.9101,
"step": 187
},
{
"epoch": 1.1609907120743035,
"grad_norm": 3.3296285577441322,
"learning_rate": 9.88353731848139e-05,
"loss": 2.8937,
"step": 188
},
{
"epoch": 1.1671826625386996,
"grad_norm": 3.6268853078486494,
"learning_rate": 9.881731876592045e-05,
"loss": 2.9918,
"step": 189
},
{
"epoch": 1.173374613003096,
"grad_norm": 3.8683456506715803,
"learning_rate": 9.879912715325612e-05,
"loss": 2.8969,
"step": 190
},
{
"epoch": 1.1795665634674923,
"grad_norm": 3.465926903871117,
"learning_rate": 9.878079839794571e-05,
"loss": 2.9277,
"step": 191
},
{
"epoch": 1.1857585139318885,
"grad_norm": 3.4135357034349245,
"learning_rate": 9.876233255149945e-05,
"loss": 2.855,
"step": 192
},
{
"epoch": 1.1919504643962848,
"grad_norm": 3.4138858890956487,
"learning_rate": 9.874372966581285e-05,
"loss": 2.8223,
"step": 193
},
{
"epoch": 1.1981424148606812,
"grad_norm": 3.1946174932663407,
"learning_rate": 9.87249897931665e-05,
"loss": 2.8626,
"step": 194
},
{
"epoch": 1.2043343653250773,
"grad_norm": 3.395109193915768,
"learning_rate": 9.870611298622605e-05,
"loss": 2.9117,
"step": 195
},
{
"epoch": 1.2105263157894737,
"grad_norm": 4.160726453183203,
"learning_rate": 9.868709929804193e-05,
"loss": 3.0261,
"step": 196
},
{
"epoch": 1.21671826625387,
"grad_norm": 3.7222325769945863,
"learning_rate": 9.866794878204926e-05,
"loss": 2.9805,
"step": 197
},
{
"epoch": 1.2229102167182662,
"grad_norm": 3.9212738228776742,
"learning_rate": 9.86486614920677e-05,
"loss": 2.8858,
"step": 198
},
{
"epoch": 1.2291021671826625,
"grad_norm": 3.6580281529557195,
"learning_rate": 9.862923748230129e-05,
"loss": 2.9386,
"step": 199
},
{
"epoch": 1.2352941176470589,
"grad_norm": 3.4226197107570133,
"learning_rate": 9.860967680733831e-05,
"loss": 2.9348,
"step": 200
},
{
"epoch": 1.2414860681114552,
"grad_norm": 4.0166432759377075,
"learning_rate": 9.858997952215112e-05,
"loss": 2.8854,
"step": 201
},
{
"epoch": 1.2476780185758514,
"grad_norm": 3.8146296663629933,
"learning_rate": 9.857014568209597e-05,
"loss": 2.971,
"step": 202
},
{
"epoch": 1.2538699690402477,
"grad_norm": 3.606205528393454,
"learning_rate": 9.855017534291292e-05,
"loss": 2.8111,
"step": 203
},
{
"epoch": 1.2600619195046439,
"grad_norm": 3.041777523006725,
"learning_rate": 9.853006856072561e-05,
"loss": 2.8318,
"step": 204
},
{
"epoch": 1.2662538699690402,
"grad_norm": 3.0864585116855596,
"learning_rate": 9.850982539204115e-05,
"loss": 2.8505,
"step": 205
},
{
"epoch": 1.2724458204334366,
"grad_norm": 3.321546800155687,
"learning_rate": 9.848944589374993e-05,
"loss": 2.94,
"step": 206
},
{
"epoch": 1.278637770897833,
"grad_norm": 3.5740731241406007,
"learning_rate": 9.846893012312549e-05,
"loss": 2.9864,
"step": 207
},
{
"epoch": 1.284829721362229,
"grad_norm": 3.8848590600257116,
"learning_rate": 9.844827813782431e-05,
"loss": 3.0241,
"step": 208
},
{
"epoch": 1.2910216718266254,
"grad_norm": 3.3954938267589907,
"learning_rate": 9.842748999588573e-05,
"loss": 2.8355,
"step": 209
},
{
"epoch": 1.2972136222910216,
"grad_norm": 4.997061770959796,
"learning_rate": 9.840656575573172e-05,
"loss": 2.938,
"step": 210
},
{
"epoch": 1.303405572755418,
"grad_norm": 3.0491464862772504,
"learning_rate": 9.83855054761667e-05,
"loss": 2.9129,
"step": 211
},
{
"epoch": 1.3095975232198143,
"grad_norm": 3.738292669236681,
"learning_rate": 9.836430921637745e-05,
"loss": 3.1255,
"step": 212
},
{
"epoch": 1.3157894736842106,
"grad_norm": 3.1688052745498445,
"learning_rate": 9.834297703593289e-05,
"loss": 2.9097,
"step": 213
},
{
"epoch": 1.3219814241486068,
"grad_norm": 3.285545746034363,
"learning_rate": 9.832150899478391e-05,
"loss": 2.9212,
"step": 214
},
{
"epoch": 1.328173374613003,
"grad_norm": 3.885090577145353,
"learning_rate": 9.829990515326324e-05,
"loss": 3.0205,
"step": 215
},
{
"epoch": 1.3343653250773992,
"grad_norm": 3.9975283980429697,
"learning_rate": 9.827816557208524e-05,
"loss": 2.8925,
"step": 216
},
{
"epoch": 1.3405572755417956,
"grad_norm": 3.0485953751430026,
"learning_rate": 9.825629031234574e-05,
"loss": 2.9117,
"step": 217
},
{
"epoch": 1.346749226006192,
"grad_norm": 3.315430153804413,
"learning_rate": 9.823427943552189e-05,
"loss": 2.8829,
"step": 218
},
{
"epoch": 1.3529411764705883,
"grad_norm": 3.5651809329580604,
"learning_rate": 9.821213300347198e-05,
"loss": 3.0057,
"step": 219
},
{
"epoch": 1.3591331269349844,
"grad_norm": 3.827599928349406,
"learning_rate": 9.818985107843523e-05,
"loss": 2.966,
"step": 220
},
{
"epoch": 1.3653250773993808,
"grad_norm": 3.7317415108791643,
"learning_rate": 9.816743372303165e-05,
"loss": 2.84,
"step": 221
},
{
"epoch": 1.3715170278637772,
"grad_norm": 3.205316348701027,
"learning_rate": 9.81448810002619e-05,
"loss": 3.0217,
"step": 222
},
{
"epoch": 1.3777089783281733,
"grad_norm": 3.3987871140372836,
"learning_rate": 9.812219297350697e-05,
"loss": 2.935,
"step": 223
},
{
"epoch": 1.3839009287925697,
"grad_norm": 3.3696442144844214,
"learning_rate": 9.809936970652823e-05,
"loss": 2.904,
"step": 224
},
{
"epoch": 1.390092879256966,
"grad_norm": 3.764260097136488,
"learning_rate": 9.807641126346702e-05,
"loss": 2.874,
"step": 225
},
{
"epoch": 1.3962848297213624,
"grad_norm": 3.367367046433176,
"learning_rate": 9.805331770884463e-05,
"loss": 2.8504,
"step": 226
},
{
"epoch": 1.4024767801857585,
"grad_norm": 3.107765697644355,
"learning_rate": 9.803008910756202e-05,
"loss": 2.9753,
"step": 227
},
{
"epoch": 1.4086687306501549,
"grad_norm": 3.2508192746442464,
"learning_rate": 9.800672552489972e-05,
"loss": 2.8143,
"step": 228
},
{
"epoch": 1.414860681114551,
"grad_norm": 3.792239883033678,
"learning_rate": 9.798322702651754e-05,
"loss": 2.978,
"step": 229
},
{
"epoch": 1.4210526315789473,
"grad_norm": 2.959609552632562,
"learning_rate": 9.795959367845456e-05,
"loss": 2.8268,
"step": 230
},
{
"epoch": 1.4272445820433437,
"grad_norm": 3.284598910008138,
"learning_rate": 9.793582554712872e-05,
"loss": 2.8506,
"step": 231
},
{
"epoch": 1.43343653250774,
"grad_norm": 3.3502742038031217,
"learning_rate": 9.79119226993368e-05,
"loss": 2.9007,
"step": 232
},
{
"epoch": 1.4396284829721362,
"grad_norm": 3.361315897122181,
"learning_rate": 9.788788520225421e-05,
"loss": 2.8859,
"step": 233
},
{
"epoch": 1.4458204334365325,
"grad_norm": 3.314532975859471,
"learning_rate": 9.786371312343471e-05,
"loss": 2.8042,
"step": 234
},
{
"epoch": 1.4520123839009287,
"grad_norm": 3.191152838747858,
"learning_rate": 9.783940653081031e-05,
"loss": 2.8412,
"step": 235
},
{
"epoch": 1.458204334365325,
"grad_norm": 2.991136323469225,
"learning_rate": 9.781496549269108e-05,
"loss": 2.8797,
"step": 236
},
{
"epoch": 1.4643962848297214,
"grad_norm": 3.038355309822365,
"learning_rate": 9.779039007776487e-05,
"loss": 2.8915,
"step": 237
},
{
"epoch": 1.4705882352941178,
"grad_norm": 2.909235326921926,
"learning_rate": 9.776568035509723e-05,
"loss": 2.9127,
"step": 238
},
{
"epoch": 1.4767801857585139,
"grad_norm": 2.91895773779436,
"learning_rate": 9.774083639413111e-05,
"loss": 2.799,
"step": 239
},
{
"epoch": 1.4829721362229102,
"grad_norm": 3.492761784355138,
"learning_rate": 9.771585826468679e-05,
"loss": 3.0023,
"step": 240
},
{
"epoch": 1.4891640866873064,
"grad_norm": 4.063292684306405,
"learning_rate": 9.769074603696153e-05,
"loss": 2.9405,
"step": 241
},
{
"epoch": 1.4953560371517027,
"grad_norm": 3.4157481247973918,
"learning_rate": 9.76654997815295e-05,
"loss": 2.8109,
"step": 242
},
{
"epoch": 1.501547987616099,
"grad_norm": 3.2243028297166583,
"learning_rate": 9.764011956934151e-05,
"loss": 2.9908,
"step": 243
},
{
"epoch": 1.5077399380804954,
"grad_norm": 3.84874713440287,
"learning_rate": 9.761460547172487e-05,
"loss": 2.7957,
"step": 244
},
{
"epoch": 1.5139318885448918,
"grad_norm": 3.2207069415479057,
"learning_rate": 9.758895756038313e-05,
"loss": 2.9172,
"step": 245
},
{
"epoch": 1.520123839009288,
"grad_norm": 3.6267787360150447,
"learning_rate": 9.756317590739591e-05,
"loss": 2.9436,
"step": 246
},
{
"epoch": 1.526315789473684,
"grad_norm": 3.257441826018387,
"learning_rate": 9.753726058521867e-05,
"loss": 2.9519,
"step": 247
},
{
"epoch": 1.5325077399380804,
"grad_norm": 3.333004086171816,
"learning_rate": 9.751121166668256e-05,
"loss": 2.9425,
"step": 248
},
{
"epoch": 1.5386996904024768,
"grad_norm": 3.4442648583827706,
"learning_rate": 9.748502922499418e-05,
"loss": 2.9805,
"step": 249
},
{
"epoch": 1.5448916408668731,
"grad_norm": 2.885763067404801,
"learning_rate": 9.745871333373533e-05,
"loss": 2.8287,
"step": 250
},
{
"epoch": 1.5510835913312695,
"grad_norm": 2.828880724852376,
"learning_rate": 9.743226406686292e-05,
"loss": 2.7098,
"step": 251
},
{
"epoch": 1.5572755417956656,
"grad_norm": 2.932748551447019,
"learning_rate": 9.740568149870864e-05,
"loss": 2.905,
"step": 252
},
{
"epoch": 1.5634674922600618,
"grad_norm": 3.663837129950405,
"learning_rate": 9.737896570397884e-05,
"loss": 2.8037,
"step": 253
},
{
"epoch": 1.5696594427244581,
"grad_norm": 3.2587738369006787,
"learning_rate": 9.735211675775423e-05,
"loss": 2.7836,
"step": 254
},
{
"epoch": 1.5758513931888545,
"grad_norm": 3.6383097347732267,
"learning_rate": 9.732513473548978e-05,
"loss": 2.823,
"step": 255
},
{
"epoch": 1.5820433436532508,
"grad_norm": 3.1219260643503786,
"learning_rate": 9.729801971301443e-05,
"loss": 2.8655,
"step": 256
},
{
"epoch": 1.5882352941176472,
"grad_norm": 3.194832439749176,
"learning_rate": 9.727077176653089e-05,
"loss": 2.8342,
"step": 257
},
{
"epoch": 1.5944272445820433,
"grad_norm": 3.278437494188074,
"learning_rate": 9.724339097261544e-05,
"loss": 2.8681,
"step": 258
},
{
"epoch": 1.6006191950464397,
"grad_norm": 3.631741144245325,
"learning_rate": 9.721587740821767e-05,
"loss": 2.8873,
"step": 259
},
{
"epoch": 1.6068111455108358,
"grad_norm": 3.6607227863330754,
"learning_rate": 9.718823115066039e-05,
"loss": 2.8828,
"step": 260
},
{
"epoch": 1.6130030959752322,
"grad_norm": 3.1235965092567777,
"learning_rate": 9.716045227763923e-05,
"loss": 2.7903,
"step": 261
},
{
"epoch": 1.6191950464396285,
"grad_norm": 3.5015393310113843,
"learning_rate": 9.713254086722258e-05,
"loss": 2.9677,
"step": 262
},
{
"epoch": 1.6253869969040249,
"grad_norm": 3.45142761429685,
"learning_rate": 9.710449699785129e-05,
"loss": 2.8914,
"step": 263
},
{
"epoch": 1.631578947368421,
"grad_norm": 3.4244592017420143,
"learning_rate": 9.707632074833843e-05,
"loss": 2.9314,
"step": 264
},
{
"epoch": 1.6377708978328174,
"grad_norm": 3.413531385719404,
"learning_rate": 9.704801219786915e-05,
"loss": 2.934,
"step": 265
},
{
"epoch": 1.6439628482972135,
"grad_norm": 3.09548146061664,
"learning_rate": 9.70195714260004e-05,
"loss": 2.8336,
"step": 266
},
{
"epoch": 1.6501547987616099,
"grad_norm": 3.0823694542298994,
"learning_rate": 9.69909985126607e-05,
"loss": 2.8104,
"step": 267
},
{
"epoch": 1.6563467492260062,
"grad_norm": 4.038090716591855,
"learning_rate": 9.696229353814996e-05,
"loss": 2.9018,
"step": 268
},
{
"epoch": 1.6625386996904026,
"grad_norm": 4.07074396660653,
"learning_rate": 9.693345658313923e-05,
"loss": 2.9215,
"step": 269
},
{
"epoch": 1.6687306501547987,
"grad_norm": 2.945420498745823,
"learning_rate": 9.690448772867042e-05,
"loss": 2.9206,
"step": 270
},
{
"epoch": 1.674922600619195,
"grad_norm": 2.9798751049881256,
"learning_rate": 9.687538705615619e-05,
"loss": 2.9972,
"step": 271
},
{
"epoch": 1.6811145510835912,
"grad_norm": 3.224093315814822,
"learning_rate": 9.684615464737963e-05,
"loss": 2.8949,
"step": 272
},
{
"epoch": 1.6873065015479876,
"grad_norm": 3.2093046728721455,
"learning_rate": 9.681679058449402e-05,
"loss": 2.8111,
"step": 273
},
{
"epoch": 1.693498452012384,
"grad_norm": 3.298118927005518,
"learning_rate": 9.67872949500227e-05,
"loss": 2.8876,
"step": 274
},
{
"epoch": 1.6996904024767803,
"grad_norm": 3.4137199531552667,
"learning_rate": 9.675766782685874e-05,
"loss": 2.7961,
"step": 275
},
{
"epoch": 1.7058823529411766,
"grad_norm": 2.872759585021391,
"learning_rate": 9.672790929826469e-05,
"loss": 2.8342,
"step": 276
},
{
"epoch": 1.7120743034055728,
"grad_norm": 2.8673153506122646,
"learning_rate": 9.669801944787249e-05,
"loss": 2.7848,
"step": 277
},
{
"epoch": 1.718266253869969,
"grad_norm": 3.0243844826232795,
"learning_rate": 9.666799835968308e-05,
"loss": 2.8087,
"step": 278
},
{
"epoch": 1.7244582043343653,
"grad_norm": 3.278008218842454,
"learning_rate": 9.663784611806624e-05,
"loss": 2.8108,
"step": 279
},
{
"epoch": 1.7306501547987616,
"grad_norm": 3.3406097276972475,
"learning_rate": 9.660756280776031e-05,
"loss": 2.8563,
"step": 280
},
{
"epoch": 1.736842105263158,
"grad_norm": 3.139644391198559,
"learning_rate": 9.657714851387202e-05,
"loss": 2.8115,
"step": 281
},
{
"epoch": 1.7430340557275543,
"grad_norm": 3.040332202758651,
"learning_rate": 9.65466033218762e-05,
"loss": 2.9067,
"step": 282
},
{
"epoch": 1.7492260061919505,
"grad_norm": 3.1365447433215956,
"learning_rate": 9.651592731761554e-05,
"loss": 2.8302,
"step": 283
},
{
"epoch": 1.7554179566563466,
"grad_norm": 2.83437264967015,
"learning_rate": 9.648512058730035e-05,
"loss": 2.8447,
"step": 284
},
{
"epoch": 1.761609907120743,
"grad_norm": 3.3452921626485446,
"learning_rate": 9.645418321750834e-05,
"loss": 2.7867,
"step": 285
},
{
"epoch": 1.7678018575851393,
"grad_norm": 3.0390803645621025,
"learning_rate": 9.642311529518438e-05,
"loss": 2.8429,
"step": 286
},
{
"epoch": 1.7739938080495357,
"grad_norm": 2.97311416389538,
"learning_rate": 9.639191690764018e-05,
"loss": 2.9274,
"step": 287
},
{
"epoch": 1.780185758513932,
"grad_norm": 3.3736768005646787,
"learning_rate": 9.636058814255418e-05,
"loss": 2.9686,
"step": 288
},
{
"epoch": 1.7863777089783281,
"grad_norm": 4.0454229812273965,
"learning_rate": 9.632912908797115e-05,
"loss": 2.6697,
"step": 289
},
{
"epoch": 1.7925696594427245,
"grad_norm": 2.932905575784552,
"learning_rate": 9.629753983230207e-05,
"loss": 2.8265,
"step": 290
},
{
"epoch": 1.7987616099071206,
"grad_norm": 3.4815911353329274,
"learning_rate": 9.626582046432384e-05,
"loss": 2.8889,
"step": 291
},
{
"epoch": 1.804953560371517,
"grad_norm": 3.304517485229584,
"learning_rate": 9.623397107317898e-05,
"loss": 2.8299,
"step": 292
},
{
"epoch": 1.8111455108359134,
"grad_norm": 3.4688423925072573,
"learning_rate": 9.620199174837541e-05,
"loss": 2.8008,
"step": 293
},
{
"epoch": 1.8173374613003097,
"grad_norm": 3.46294088440551,
"learning_rate": 9.616988257978628e-05,
"loss": 2.8924,
"step": 294
},
{
"epoch": 1.8235294117647058,
"grad_norm": 3.5088075010477504,
"learning_rate": 9.61376436576496e-05,
"loss": 2.8466,
"step": 295
},
{
"epoch": 1.8297213622291022,
"grad_norm": 3.5637222019380825,
"learning_rate": 9.610527507256802e-05,
"loss": 2.8358,
"step": 296
},
{
"epoch": 1.8359133126934983,
"grad_norm": 3.1688045200480155,
"learning_rate": 9.607277691550862e-05,
"loss": 2.8324,
"step": 297
},
{
"epoch": 1.8421052631578947,
"grad_norm": 3.0702348395736236,
"learning_rate": 9.60401492778026e-05,
"loss": 2.9719,
"step": 298
},
{
"epoch": 1.848297213622291,
"grad_norm": 3.588593649177979,
"learning_rate": 9.600739225114506e-05,
"loss": 2.826,
"step": 299
},
{
"epoch": 1.8544891640866874,
"grad_norm": 3.3305202346452294,
"learning_rate": 9.59745059275947e-05,
"loss": 3.0333,
"step": 300
},
{
"epoch": 1.8606811145510835,
"grad_norm": 3.1335924494846665,
"learning_rate": 9.594149039957365e-05,
"loss": 2.8217,
"step": 301
},
{
"epoch": 1.86687306501548,
"grad_norm": 2.596605310119535,
"learning_rate": 9.590834575986708e-05,
"loss": 2.8302,
"step": 302
},
{
"epoch": 1.873065015479876,
"grad_norm": 2.842224636487708,
"learning_rate": 9.587507210162307e-05,
"loss": 2.77,
"step": 303
},
{
"epoch": 1.8792569659442724,
"grad_norm": 3.375074913274891,
"learning_rate": 9.584166951835222e-05,
"loss": 2.914,
"step": 304
},
{
"epoch": 1.8854489164086687,
"grad_norm": 3.086672402067182,
"learning_rate": 9.580813810392755e-05,
"loss": 2.8741,
"step": 305
},
{
"epoch": 1.891640866873065,
"grad_norm": 3.412761519550504,
"learning_rate": 9.577447795258403e-05,
"loss": 2.713,
"step": 306
},
{
"epoch": 1.8978328173374615,
"grad_norm": 2.825161059955203,
"learning_rate": 9.574068915891849e-05,
"loss": 2.7335,
"step": 307
},
{
"epoch": 1.9040247678018576,
"grad_norm": 2.9720239351377775,
"learning_rate": 9.57067718178893e-05,
"loss": 2.7431,
"step": 308
},
{
"epoch": 1.9102167182662537,
"grad_norm": 3.2093709733529856,
"learning_rate": 9.567272602481606e-05,
"loss": 2.9308,
"step": 309
},
{
"epoch": 1.91640866873065,
"grad_norm": 3.048048026454268,
"learning_rate": 9.563855187537937e-05,
"loss": 2.8424,
"step": 310
},
{
"epoch": 1.9226006191950464,
"grad_norm": 3.0542567194175567,
"learning_rate": 9.56042494656206e-05,
"loss": 2.8537,
"step": 311
},
{
"epoch": 1.9287925696594428,
"grad_norm": 3.349906695410103,
"learning_rate": 9.556981889194149e-05,
"loss": 2.7549,
"step": 312
},
{
"epoch": 1.9349845201238391,
"grad_norm": 3.0260829521434807,
"learning_rate": 9.553526025110405e-05,
"loss": 2.962,
"step": 313
},
{
"epoch": 1.9411764705882353,
"grad_norm": 3.1632389516665778,
"learning_rate": 9.550057364023016e-05,
"loss": 2.8224,
"step": 314
},
{
"epoch": 1.9473684210526314,
"grad_norm": 3.1179495852636516,
"learning_rate": 9.546575915680134e-05,
"loss": 2.8394,
"step": 315
},
{
"epoch": 1.9535603715170278,
"grad_norm": 3.242102168782828,
"learning_rate": 9.54308168986585e-05,
"loss": 2.8167,
"step": 316
},
{
"epoch": 1.9597523219814241,
"grad_norm": 3.53462667333172,
"learning_rate": 9.539574696400164e-05,
"loss": 2.7766,
"step": 317
},
{
"epoch": 1.9659442724458205,
"grad_norm": 3.987768918752747,
"learning_rate": 9.536054945138953e-05,
"loss": 2.6656,
"step": 318
},
{
"epoch": 1.9721362229102168,
"grad_norm": 2.8942848523535947,
"learning_rate": 9.532522445973955e-05,
"loss": 2.798,
"step": 319
},
{
"epoch": 1.978328173374613,
"grad_norm": 2.938369919319297,
"learning_rate": 9.528977208832729e-05,
"loss": 2.8506,
"step": 320
},
{
"epoch": 1.9845201238390093,
"grad_norm": 3.286930320946629,
"learning_rate": 9.525419243678632e-05,
"loss": 2.9692,
"step": 321
},
{
"epoch": 1.9907120743034055,
"grad_norm": 2.900305427286331,
"learning_rate": 9.521848560510796e-05,
"loss": 2.8208,
"step": 322
},
{
"epoch": 1.9969040247678018,
"grad_norm": 3.183121159883855,
"learning_rate": 9.518265169364089e-05,
"loss": 2.761,
"step": 323
},
{
"epoch": 2.0,
"grad_norm": 2.768600331874623,
"learning_rate": 9.514669080309096e-05,
"loss": 1.4118,
"step": 324
},
{
"epoch": 2.0061919504643964,
"grad_norm": 2.7548583500669976,
"learning_rate": 9.51106030345209e-05,
"loss": 2.6787,
"step": 325
},
{
"epoch": 2.0123839009287927,
"grad_norm": 3.4443221488903926,
"learning_rate": 9.507438848934995e-05,
"loss": 2.6948,
"step": 326
},
{
"epoch": 2.018575851393189,
"grad_norm": 2.8726596638954938,
"learning_rate": 9.503804726935368e-05,
"loss": 2.6093,
"step": 327
},
{
"epoch": 2.024767801857585,
"grad_norm": 3.389044741091986,
"learning_rate": 9.500157947666367e-05,
"loss": 2.5705,
"step": 328
},
{
"epoch": 2.0309597523219813,
"grad_norm": 3.0580993903626306,
"learning_rate": 9.496498521376717e-05,
"loss": 2.5358,
"step": 329
},
{
"epoch": 2.0371517027863777,
"grad_norm": 3.468931786923507,
"learning_rate": 9.492826458350691e-05,
"loss": 2.5419,
"step": 330
},
{
"epoch": 2.043343653250774,
"grad_norm": 3.3444948196299427,
"learning_rate": 9.489141768908071e-05,
"loss": 2.7272,
"step": 331
},
{
"epoch": 2.0495356037151704,
"grad_norm": 3.1880613030560454,
"learning_rate": 9.485444463404125e-05,
"loss": 2.6179,
"step": 332
},
{
"epoch": 2.0557275541795668,
"grad_norm": 3.987073827921253,
"learning_rate": 9.481734552229578e-05,
"loss": 2.6154,
"step": 333
},
{
"epoch": 2.0619195046439627,
"grad_norm": 3.405296329527151,
"learning_rate": 9.47801204581058e-05,
"loss": 2.6811,
"step": 334
},
{
"epoch": 2.068111455108359,
"grad_norm": 3.1752299278728175,
"learning_rate": 9.474276954608677e-05,
"loss": 2.638,
"step": 335
},
{
"epoch": 2.0743034055727554,
"grad_norm": 3.4269857244773543,
"learning_rate": 9.470529289120786e-05,
"loss": 2.6532,
"step": 336
},
{
"epoch": 2.0804953560371517,
"grad_norm": 3.2676309514016495,
"learning_rate": 9.466769059879159e-05,
"loss": 2.6029,
"step": 337
},
{
"epoch": 2.086687306501548,
"grad_norm": 3.613838896557624,
"learning_rate": 9.462996277451359e-05,
"loss": 2.5344,
"step": 338
},
{
"epoch": 2.0928792569659445,
"grad_norm": 3.1931623478278928,
"learning_rate": 9.459210952440225e-05,
"loss": 2.5997,
"step": 339
},
{
"epoch": 2.0990712074303404,
"grad_norm": 3.5892129567202913,
"learning_rate": 9.455413095483849e-05,
"loss": 2.6291,
"step": 340
},
{
"epoch": 2.1052631578947367,
"grad_norm": 3.1425227101080955,
"learning_rate": 9.451602717255537e-05,
"loss": 2.5365,
"step": 341
},
{
"epoch": 2.111455108359133,
"grad_norm": 3.7512542147695407,
"learning_rate": 9.447779828463788e-05,
"loss": 2.7276,
"step": 342
},
{
"epoch": 2.1176470588235294,
"grad_norm": 3.8083642585550237,
"learning_rate": 9.443944439852259e-05,
"loss": 2.619,
"step": 343
},
{
"epoch": 2.123839009287926,
"grad_norm": 3.870821371956724,
"learning_rate": 9.440096562199736e-05,
"loss": 2.5995,
"step": 344
},
{
"epoch": 2.130030959752322,
"grad_norm": 3.004401568037884,
"learning_rate": 9.436236206320104e-05,
"loss": 2.5807,
"step": 345
},
{
"epoch": 2.136222910216718,
"grad_norm": 3.5811035323778624,
"learning_rate": 9.432363383062314e-05,
"loss": 2.5922,
"step": 346
},
{
"epoch": 2.1424148606811144,
"grad_norm": 3.1678609835071634,
"learning_rate": 9.428478103310357e-05,
"loss": 2.5559,
"step": 347
},
{
"epoch": 2.1486068111455108,
"grad_norm": 4.00863984446441,
"learning_rate": 9.424580377983231e-05,
"loss": 2.814,
"step": 348
},
{
"epoch": 2.154798761609907,
"grad_norm": 3.3069124426114636,
"learning_rate": 9.420670218034913e-05,
"loss": 2.5825,
"step": 349
},
{
"epoch": 2.1609907120743035,
"grad_norm": 3.227932310894545,
"learning_rate": 9.416747634454316e-05,
"loss": 2.7328,
"step": 350
},
{
"epoch": 2.1671826625387,
"grad_norm": 3.193488562980502,
"learning_rate": 9.412812638265279e-05,
"loss": 2.6004,
"step": 351
},
{
"epoch": 2.173374613003096,
"grad_norm": 3.138286190856697,
"learning_rate": 9.408865240526518e-05,
"loss": 2.6494,
"step": 352
},
{
"epoch": 2.179566563467492,
"grad_norm": 3.486032822685665,
"learning_rate": 9.404905452331604e-05,
"loss": 2.5935,
"step": 353
},
{
"epoch": 2.1857585139318885,
"grad_norm": 3.57049625554477,
"learning_rate": 9.400933284808933e-05,
"loss": 2.5879,
"step": 354
},
{
"epoch": 2.191950464396285,
"grad_norm": 3.0638448164434826,
"learning_rate": 9.396948749121681e-05,
"loss": 2.695,
"step": 355
},
{
"epoch": 2.198142414860681,
"grad_norm": 3.9956558838253056,
"learning_rate": 9.392951856467795e-05,
"loss": 2.6175,
"step": 356
},
{
"epoch": 2.2043343653250775,
"grad_norm": 3.8578672053862775,
"learning_rate": 9.388942618079941e-05,
"loss": 2.6658,
"step": 357
},
{
"epoch": 2.2105263157894735,
"grad_norm": 3.25975304104384,
"learning_rate": 9.384921045225483e-05,
"loss": 2.5892,
"step": 358
},
{
"epoch": 2.21671826625387,
"grad_norm": 2.5766286937173626,
"learning_rate": 9.380887149206452e-05,
"loss": 2.6644,
"step": 359
},
{
"epoch": 2.222910216718266,
"grad_norm": 3.047972701055752,
"learning_rate": 9.376840941359508e-05,
"loss": 2.5642,
"step": 360
},
{
"epoch": 2.2291021671826625,
"grad_norm": 2.9547625831600492,
"learning_rate": 9.372782433055914e-05,
"loss": 2.5834,
"step": 361
},
{
"epoch": 2.235294117647059,
"grad_norm": 3.2164908658189844,
"learning_rate": 9.368711635701499e-05,
"loss": 2.6811,
"step": 362
},
{
"epoch": 2.2414860681114552,
"grad_norm": 3.283475564049949,
"learning_rate": 9.364628560736631e-05,
"loss": 2.6111,
"step": 363
},
{
"epoch": 2.2476780185758516,
"grad_norm": 3.343599128743468,
"learning_rate": 9.36053321963618e-05,
"loss": 2.6123,
"step": 364
},
{
"epoch": 2.2538699690402475,
"grad_norm": 3.4502623336737424,
"learning_rate": 9.356425623909493e-05,
"loss": 2.7775,
"step": 365
},
{
"epoch": 2.260061919504644,
"grad_norm": 3.162272731188207,
"learning_rate": 9.352305785100348e-05,
"loss": 2.7052,
"step": 366
},
{
"epoch": 2.26625386996904,
"grad_norm": 3.3567857923317272,
"learning_rate": 9.348173714786939e-05,
"loss": 2.6605,
"step": 367
},
{
"epoch": 2.2724458204334366,
"grad_norm": 3.3134267019833974,
"learning_rate": 9.344029424581829e-05,
"loss": 2.6522,
"step": 368
},
{
"epoch": 2.278637770897833,
"grad_norm": 3.0966854245280353,
"learning_rate": 9.339872926131929e-05,
"loss": 2.5706,
"step": 369
},
{
"epoch": 2.2848297213622293,
"grad_norm": 2.9494878884102533,
"learning_rate": 9.335704231118454e-05,
"loss": 2.7847,
"step": 370
},
{
"epoch": 2.291021671826625,
"grad_norm": 3.0052075309306483,
"learning_rate": 9.331523351256896e-05,
"loss": 2.4242,
"step": 371
},
{
"epoch": 2.2972136222910216,
"grad_norm": 3.865779841352929,
"learning_rate": 9.327330298296997e-05,
"loss": 2.6202,
"step": 372
},
{
"epoch": 2.303405572755418,
"grad_norm": 3.413710786689101,
"learning_rate": 9.323125084022701e-05,
"loss": 2.6319,
"step": 373
},
{
"epoch": 2.3095975232198143,
"grad_norm": 3.9872687160051927,
"learning_rate": 9.318907720252136e-05,
"loss": 2.7452,
"step": 374
},
{
"epoch": 2.3157894736842106,
"grad_norm": 3.606954976832062,
"learning_rate": 9.31467821883757e-05,
"loss": 2.7013,
"step": 375
},
{
"epoch": 2.321981424148607,
"grad_norm": 3.0244580104851364,
"learning_rate": 9.310436591665384e-05,
"loss": 2.7322,
"step": 376
},
{
"epoch": 2.3281733746130033,
"grad_norm": 3.2847212713666702,
"learning_rate": 9.306182850656038e-05,
"loss": 2.6815,
"step": 377
},
{
"epoch": 2.3343653250773992,
"grad_norm": 3.253192934470869,
"learning_rate": 9.301917007764034e-05,
"loss": 2.7149,
"step": 378
},
{
"epoch": 2.3405572755417956,
"grad_norm": 3.384274414142735,
"learning_rate": 9.297639074977885e-05,
"loss": 2.6428,
"step": 379
},
{
"epoch": 2.346749226006192,
"grad_norm": 3.5615004713406666,
"learning_rate": 9.293349064320083e-05,
"loss": 2.6023,
"step": 380
},
{
"epoch": 2.3529411764705883,
"grad_norm": 3.153626899286148,
"learning_rate": 9.289046987847058e-05,
"loss": 2.5489,
"step": 381
},
{
"epoch": 2.3591331269349847,
"grad_norm": 4.008345134198719,
"learning_rate": 9.284732857649154e-05,
"loss": 2.6347,
"step": 382
},
{
"epoch": 2.3653250773993806,
"grad_norm": 3.605420355608457,
"learning_rate": 9.280406685850586e-05,
"loss": 2.6115,
"step": 383
},
{
"epoch": 2.371517027863777,
"grad_norm": 3.2090318428995896,
"learning_rate": 9.276068484609418e-05,
"loss": 2.5584,
"step": 384
},
{
"epoch": 2.3777089783281733,
"grad_norm": 3.7889425362871556,
"learning_rate": 9.271718266117511e-05,
"loss": 2.7685,
"step": 385
},
{
"epoch": 2.3839009287925697,
"grad_norm": 3.4617864132858056,
"learning_rate": 9.267356042600505e-05,
"loss": 2.6928,
"step": 386
},
{
"epoch": 2.390092879256966,
"grad_norm": 3.4918988126495414,
"learning_rate": 9.262981826317777e-05,
"loss": 2.5776,
"step": 387
},
{
"epoch": 2.3962848297213624,
"grad_norm": 3.2844017613022225,
"learning_rate": 9.258595629562405e-05,
"loss": 2.5817,
"step": 388
},
{
"epoch": 2.4024767801857587,
"grad_norm": 3.8704827773214006,
"learning_rate": 9.254197464661143e-05,
"loss": 2.6336,
"step": 389
},
{
"epoch": 2.4086687306501546,
"grad_norm": 3.605148120779814,
"learning_rate": 9.249787343974371e-05,
"loss": 2.5303,
"step": 390
},
{
"epoch": 2.414860681114551,
"grad_norm": 3.9953677588197496,
"learning_rate": 9.245365279896075e-05,
"loss": 2.5591,
"step": 391
},
{
"epoch": 2.4210526315789473,
"grad_norm": 3.2326130501278465,
"learning_rate": 9.240931284853807e-05,
"loss": 2.6319,
"step": 392
},
{
"epoch": 2.4272445820433437,
"grad_norm": 3.4297234946250135,
"learning_rate": 9.236485371308642e-05,
"loss": 2.6363,
"step": 393
},
{
"epoch": 2.43343653250774,
"grad_norm": 3.9434355381933837,
"learning_rate": 9.232027551755157e-05,
"loss": 2.5911,
"step": 394
},
{
"epoch": 2.4396284829721364,
"grad_norm": 3.416834504577997,
"learning_rate": 9.22755783872139e-05,
"loss": 2.7753,
"step": 395
},
{
"epoch": 2.4458204334365323,
"grad_norm": 2.798767439755245,
"learning_rate": 9.223076244768797e-05,
"loss": 2.5571,
"step": 396
},
{
"epoch": 2.4520123839009287,
"grad_norm": 3.312096781560587,
"learning_rate": 9.218582782492227e-05,
"loss": 2.7031,
"step": 397
},
{
"epoch": 2.458204334365325,
"grad_norm": 3.3106706550562772,
"learning_rate": 9.214077464519885e-05,
"loss": 2.5017,
"step": 398
},
{
"epoch": 2.4643962848297214,
"grad_norm": 2.882854108845148,
"learning_rate": 9.209560303513296e-05,
"loss": 2.779,
"step": 399
},
{
"epoch": 2.4705882352941178,
"grad_norm": 3.603742563496827,
"learning_rate": 9.20503131216726e-05,
"loss": 2.5953,
"step": 400
},
{
"epoch": 2.476780185758514,
"grad_norm": 3.056177277208107,
"learning_rate": 9.20049050320983e-05,
"loss": 2.6001,
"step": 401
},
{
"epoch": 2.4829721362229105,
"grad_norm": 3.494580488159622,
"learning_rate": 9.195937889402276e-05,
"loss": 2.6322,
"step": 402
},
{
"epoch": 2.4891640866873064,
"grad_norm": 3.360260834795966,
"learning_rate": 9.19137348353903e-05,
"loss": 2.63,
"step": 403
},
{
"epoch": 2.4953560371517027,
"grad_norm": 3.320590906602975,
"learning_rate": 9.186797298447678e-05,
"loss": 2.5977,
"step": 404
},
{
"epoch": 2.501547987616099,
"grad_norm": 3.4580929903970414,
"learning_rate": 9.182209346988902e-05,
"loss": 2.7052,
"step": 405
},
{
"epoch": 2.5077399380804954,
"grad_norm": 3.113537260619799,
"learning_rate": 9.177609642056451e-05,
"loss": 2.6589,
"step": 406
},
{
"epoch": 2.513931888544892,
"grad_norm": 2.7289651129651147,
"learning_rate": 9.17299819657711e-05,
"loss": 2.5995,
"step": 407
},
{
"epoch": 2.5201238390092877,
"grad_norm": 2.9890848188099066,
"learning_rate": 9.168375023510654e-05,
"loss": 2.7075,
"step": 408
},
{
"epoch": 2.526315789473684,
"grad_norm": 3.1251539583034615,
"learning_rate": 9.163740135849823e-05,
"loss": 2.6149,
"step": 409
},
{
"epoch": 2.5325077399380804,
"grad_norm": 3.4651982610153986,
"learning_rate": 9.159093546620272e-05,
"loss": 2.597,
"step": 410
},
{
"epoch": 2.538699690402477,
"grad_norm": 3.443446796385959,
"learning_rate": 9.154435268880546e-05,
"loss": 2.7328,
"step": 411
},
{
"epoch": 2.544891640866873,
"grad_norm": 3.5113324699487842,
"learning_rate": 9.149765315722039e-05,
"loss": 2.6076,
"step": 412
},
{
"epoch": 2.5510835913312695,
"grad_norm": 3.6507929025803088,
"learning_rate": 9.145083700268954e-05,
"loss": 2.5952,
"step": 413
},
{
"epoch": 2.557275541795666,
"grad_norm": 3.1280086170625503,
"learning_rate": 9.140390435678272e-05,
"loss": 2.6725,
"step": 414
},
{
"epoch": 2.5634674922600618,
"grad_norm": 3.4152513222205316,
"learning_rate": 9.135685535139709e-05,
"loss": 2.6314,
"step": 415
},
{
"epoch": 2.569659442724458,
"grad_norm": 3.278814596158421,
"learning_rate": 9.130969011875685e-05,
"loss": 2.7043,
"step": 416
},
{
"epoch": 2.5758513931888545,
"grad_norm": 3.4642520708537496,
"learning_rate": 9.126240879141285e-05,
"loss": 2.6676,
"step": 417
},
{
"epoch": 2.582043343653251,
"grad_norm": 3.2092964996317783,
"learning_rate": 9.121501150224217e-05,
"loss": 2.6003,
"step": 418
},
{
"epoch": 2.588235294117647,
"grad_norm": 3.983130708249009,
"learning_rate": 9.116749838444778e-05,
"loss": 2.8115,
"step": 419
},
{
"epoch": 2.594427244582043,
"grad_norm": 3.2303996809709057,
"learning_rate": 9.111986957155823e-05,
"loss": 2.7408,
"step": 420
},
{
"epoch": 2.6006191950464395,
"grad_norm": 3.021042071829146,
"learning_rate": 9.107212519742714e-05,
"loss": 2.6704,
"step": 421
},
{
"epoch": 2.606811145510836,
"grad_norm": 3.7177435258237534,
"learning_rate": 9.102426539623295e-05,
"loss": 2.6222,
"step": 422
},
{
"epoch": 2.613003095975232,
"grad_norm": 3.3065094436349667,
"learning_rate": 9.097629030247846e-05,
"loss": 2.6983,
"step": 423
},
{
"epoch": 2.6191950464396285,
"grad_norm": 3.3461082798054864,
"learning_rate": 9.09282000509905e-05,
"loss": 2.6419,
"step": 424
},
{
"epoch": 2.625386996904025,
"grad_norm": 3.3413036061467976,
"learning_rate": 9.087999477691952e-05,
"loss": 2.6388,
"step": 425
},
{
"epoch": 2.6315789473684212,
"grad_norm": 3.866138226599775,
"learning_rate": 9.083167461573925e-05,
"loss": 2.6888,
"step": 426
},
{
"epoch": 2.6377708978328176,
"grad_norm": 3.751760167233359,
"learning_rate": 9.078323970324625e-05,
"loss": 2.722,
"step": 427
},
{
"epoch": 2.6439628482972135,
"grad_norm": 3.2089862942574894,
"learning_rate": 9.07346901755596e-05,
"loss": 2.6437,
"step": 428
},
{
"epoch": 2.65015479876161,
"grad_norm": 3.0031534771552697,
"learning_rate": 9.068602616912049e-05,
"loss": 2.6457,
"step": 429
},
{
"epoch": 2.656346749226006,
"grad_norm": 2.998028740990232,
"learning_rate": 9.063724782069183e-05,
"loss": 2.5922,
"step": 430
},
{
"epoch": 2.6625386996904026,
"grad_norm": 3.2856296543776735,
"learning_rate": 9.058835526735787e-05,
"loss": 2.5476,
"step": 431
},
{
"epoch": 2.6687306501547985,
"grad_norm": 3.278728909273792,
"learning_rate": 9.053934864652382e-05,
"loss": 2.7063,
"step": 432
},
{
"epoch": 2.674922600619195,
"grad_norm": 3.203008246999124,
"learning_rate": 9.049022809591546e-05,
"loss": 2.7475,
"step": 433
},
{
"epoch": 2.681114551083591,
"grad_norm": 3.354224738161965,
"learning_rate": 9.04409937535787e-05,
"loss": 2.6243,
"step": 434
},
{
"epoch": 2.6873065015479876,
"grad_norm": 3.5804759078596056,
"learning_rate": 9.039164575787938e-05,
"loss": 2.5816,
"step": 435
},
{
"epoch": 2.693498452012384,
"grad_norm": 2.9452472701608543,
"learning_rate": 9.034218424750259e-05,
"loss": 2.6851,
"step": 436
},
{
"epoch": 2.6996904024767803,
"grad_norm": 3.196962334371782,
"learning_rate": 9.029260936145251e-05,
"loss": 2.5857,
"step": 437
},
{
"epoch": 2.7058823529411766,
"grad_norm": 2.905856705546016,
"learning_rate": 9.0242921239052e-05,
"loss": 2.6213,
"step": 438
},
{
"epoch": 2.712074303405573,
"grad_norm": 2.9055102854318506,
"learning_rate": 9.019312001994202e-05,
"loss": 2.535,
"step": 439
},
{
"epoch": 2.718266253869969,
"grad_norm": 3.1173866399821506,
"learning_rate": 9.014320584408148e-05,
"loss": 2.5721,
"step": 440
},
{
"epoch": 2.7244582043343653,
"grad_norm": 3.7789655683417287,
"learning_rate": 9.00931788517467e-05,
"loss": 2.6374,
"step": 441
},
{
"epoch": 2.7306501547987616,
"grad_norm": 3.4646023201290106,
"learning_rate": 9.004303918353107e-05,
"loss": 2.6499,
"step": 442
},
{
"epoch": 2.736842105263158,
"grad_norm": 3.9484421562322747,
"learning_rate": 8.999278698034461e-05,
"loss": 2.6948,
"step": 443
},
{
"epoch": 2.7430340557275543,
"grad_norm": 3.442553230839332,
"learning_rate": 8.994242238341362e-05,
"loss": 2.6144,
"step": 444
},
{
"epoch": 2.7492260061919502,
"grad_norm": 5.0516235581172895,
"learning_rate": 8.989194553428028e-05,
"loss": 2.6532,
"step": 445
},
{
"epoch": 2.7554179566563466,
"grad_norm": 3.68844887791005,
"learning_rate": 8.984135657480219e-05,
"loss": 2.5519,
"step": 446
},
{
"epoch": 2.761609907120743,
"grad_norm": 3.135950737705265,
"learning_rate": 8.97906556471521e-05,
"loss": 2.6005,
"step": 447
},
{
"epoch": 2.7678018575851393,
"grad_norm": 2.973619254046368,
"learning_rate": 8.973984289381733e-05,
"loss": 2.4656,
"step": 448
},
{
"epoch": 2.7739938080495357,
"grad_norm": 4.290896676580274,
"learning_rate": 8.968891845759955e-05,
"loss": 2.6052,
"step": 449
},
{
"epoch": 2.780185758513932,
"grad_norm": 3.8773305146838783,
"learning_rate": 8.963788248161428e-05,
"loss": 2.6004,
"step": 450
},
{
"epoch": 2.7863777089783284,
"grad_norm": 3.5246724411520507,
"learning_rate": 8.958673510929046e-05,
"loss": 2.7062,
"step": 451
},
{
"epoch": 2.7925696594427247,
"grad_norm": 3.68544856626367,
"learning_rate": 8.953547648437016e-05,
"loss": 2.7555,
"step": 452
},
{
"epoch": 2.7987616099071206,
"grad_norm": 4.282996144603401,
"learning_rate": 8.948410675090805e-05,
"loss": 2.7417,
"step": 453
},
{
"epoch": 2.804953560371517,
"grad_norm": 4.9262667481372135,
"learning_rate": 8.943262605327112e-05,
"loss": 2.6828,
"step": 454
},
{
"epoch": 2.8111455108359134,
"grad_norm": 3.51958635902412,
"learning_rate": 8.938103453613813e-05,
"loss": 2.7015,
"step": 455
},
{
"epoch": 2.8173374613003097,
"grad_norm": 3.003904355716457,
"learning_rate": 8.932933234449934e-05,
"loss": 2.6282,
"step": 456
},
{
"epoch": 2.8235294117647056,
"grad_norm": 2.8951986217352133,
"learning_rate": 8.927751962365603e-05,
"loss": 2.5677,
"step": 457
},
{
"epoch": 2.829721362229102,
"grad_norm": 2.973677750838087,
"learning_rate": 8.922559651922006e-05,
"loss": 2.6943,
"step": 458
},
{
"epoch": 2.8359133126934983,
"grad_norm": 3.923968948876088,
"learning_rate": 8.917356317711358e-05,
"loss": 2.8386,
"step": 459
},
{
"epoch": 2.8421052631578947,
"grad_norm": 3.748206431260409,
"learning_rate": 8.912141974356853e-05,
"loss": 2.5004,
"step": 460
},
{
"epoch": 2.848297213622291,
"grad_norm": 2.9589835275416165,
"learning_rate": 8.90691663651262e-05,
"loss": 2.7211,
"step": 461
},
{
"epoch": 2.8544891640866874,
"grad_norm": 4.413461022857995,
"learning_rate": 8.90168031886369e-05,
"loss": 2.7161,
"step": 462
},
{
"epoch": 2.8606811145510838,
"grad_norm": 3.8458500455119355,
"learning_rate": 8.896433036125949e-05,
"loss": 2.6386,
"step": 463
},
{
"epoch": 2.86687306501548,
"grad_norm": 3.5064727605205235,
"learning_rate": 8.891174803046102e-05,
"loss": 2.6769,
"step": 464
},
{
"epoch": 2.873065015479876,
"grad_norm": 3.3253407518893554,
"learning_rate": 8.885905634401628e-05,
"loss": 2.3765,
"step": 465
},
{
"epoch": 2.8792569659442724,
"grad_norm": 3.2901194346879232,
"learning_rate": 8.880625545000734e-05,
"loss": 2.6121,
"step": 466
},
{
"epoch": 2.8854489164086687,
"grad_norm": 3.8787003089594965,
"learning_rate": 8.875334549682322e-05,
"loss": 2.7162,
"step": 467
},
{
"epoch": 2.891640866873065,
"grad_norm": 3.1158601459237087,
"learning_rate": 8.870032663315943e-05,
"loss": 2.4268,
"step": 468
},
{
"epoch": 2.8978328173374615,
"grad_norm": 3.523227254158828,
"learning_rate": 8.864719900801755e-05,
"loss": 2.6866,
"step": 469
},
{
"epoch": 2.9040247678018574,
"grad_norm": 3.7947906426929294,
"learning_rate": 8.859396277070483e-05,
"loss": 2.5134,
"step": 470
},
{
"epoch": 2.9102167182662537,
"grad_norm": 3.5665946360922756,
"learning_rate": 8.854061807083376e-05,
"loss": 2.6598,
"step": 471
},
{
"epoch": 2.91640866873065,
"grad_norm": 3.6433600779737723,
"learning_rate": 8.848716505832162e-05,
"loss": 2.615,
"step": 472
},
{
"epoch": 2.9226006191950464,
"grad_norm": 3.280696174381033,
"learning_rate": 8.843360388339011e-05,
"loss": 2.6286,
"step": 473
},
{
"epoch": 2.928792569659443,
"grad_norm": 3.5271347403489295,
"learning_rate": 8.837993469656489e-05,
"loss": 2.6689,
"step": 474
},
{
"epoch": 2.934984520123839,
"grad_norm": 3.535085290280349,
"learning_rate": 8.83261576486752e-05,
"loss": 2.7654,
"step": 475
},
{
"epoch": 2.9411764705882355,
"grad_norm": 3.1024372707379664,
"learning_rate": 8.827227289085339e-05,
"loss": 2.5418,
"step": 476
},
{
"epoch": 2.9473684210526314,
"grad_norm": 3.0048339939826154,
"learning_rate": 8.821828057453447e-05,
"loss": 2.6055,
"step": 477
},
{
"epoch": 2.9535603715170278,
"grad_norm": 3.6042625802261625,
"learning_rate": 8.816418085145582e-05,
"loss": 2.8061,
"step": 478
},
{
"epoch": 2.959752321981424,
"grad_norm": 4.4277235133739605,
"learning_rate": 8.810997387365656e-05,
"loss": 2.8611,
"step": 479
},
{
"epoch": 2.9659442724458205,
"grad_norm": 3.672119841517933,
"learning_rate": 8.805565979347734e-05,
"loss": 2.781,
"step": 480
},
{
"epoch": 2.972136222910217,
"grad_norm": 4.2331486999864,
"learning_rate": 8.800123876355976e-05,
"loss": 2.6954,
"step": 481
},
{
"epoch": 2.9783281733746128,
"grad_norm": 3.6567653823539694,
"learning_rate": 8.794671093684595e-05,
"loss": 2.6517,
"step": 482
},
{
"epoch": 2.984520123839009,
"grad_norm": 3.4224822473176966,
"learning_rate": 8.789207646657822e-05,
"loss": 2.6759,
"step": 483
},
{
"epoch": 2.9907120743034055,
"grad_norm": 3.299869011062663,
"learning_rate": 8.783733550629857e-05,
"loss": 2.7011,
"step": 484
},
{
"epoch": 2.996904024767802,
"grad_norm": 3.2365085802605464,
"learning_rate": 8.778248820984829e-05,
"loss": 2.6494,
"step": 485
},
{
"epoch": 3.0,
"grad_norm": 3.2365085802605464,
"learning_rate": 8.77275347313675e-05,
"loss": 1.2508,
"step": 486
},
{
"epoch": 3.0061919504643964,
"grad_norm": 3.142785796812641,
"learning_rate": 8.767247522529474e-05,
"loss": 2.5231,
"step": 487
},
{
"epoch": 3.0123839009287927,
"grad_norm": 2.9912371857567117,
"learning_rate": 8.761730984636648e-05,
"loss": 2.4012,
"step": 488
},
{
"epoch": 3.018575851393189,
"grad_norm": 3.242699233358939,
"learning_rate": 8.75620387496168e-05,
"loss": 2.4209,
"step": 489
},
{
"epoch": 3.024767801857585,
"grad_norm": 3.4343945955583606,
"learning_rate": 8.750666209037684e-05,
"loss": 2.3436,
"step": 490
},
{
"epoch": 3.0309597523219813,
"grad_norm": 4.028979174625844,
"learning_rate": 8.745118002427439e-05,
"loss": 2.3537,
"step": 491
},
{
"epoch": 3.0371517027863777,
"grad_norm": 3.419844103554572,
"learning_rate": 8.739559270723353e-05,
"loss": 2.3517,
"step": 492
},
{
"epoch": 3.043343653250774,
"grad_norm": 4.079075872820777,
"learning_rate": 8.733990029547407e-05,
"loss": 2.4506,
"step": 493
},
{
"epoch": 3.0495356037151704,
"grad_norm": 4.453433315359627,
"learning_rate": 8.728410294551123e-05,
"loss": 2.4242,
"step": 494
},
{
"epoch": 3.0557275541795668,
"grad_norm": 4.185230535033415,
"learning_rate": 8.72282008141551e-05,
"loss": 2.4632,
"step": 495
},
{
"epoch": 3.0619195046439627,
"grad_norm": 4.434685158118041,
"learning_rate": 8.717219405851025e-05,
"loss": 2.3609,
"step": 496
},
{
"epoch": 3.068111455108359,
"grad_norm": 4.265164936477565,
"learning_rate": 8.71160828359753e-05,
"loss": 2.4868,
"step": 497
},
{
"epoch": 3.0743034055727554,
"grad_norm": 4.115197176895919,
"learning_rate": 8.705986730424243e-05,
"loss": 2.4306,
"step": 498
},
{
"epoch": 3.0804953560371517,
"grad_norm": 3.6219070451775246,
"learning_rate": 8.7003547621297e-05,
"loss": 2.3345,
"step": 499
},
{
"epoch": 3.086687306501548,
"grad_norm": 3.260204917718609,
"learning_rate": 8.694712394541705e-05,
"loss": 2.4696,
"step": 500
},
{
"epoch": 3.0928792569659445,
"grad_norm": 3.661355382479066,
"learning_rate": 8.689059643517286e-05,
"loss": 2.482,
"step": 501
},
{
"epoch": 3.0990712074303404,
"grad_norm": 4.094138838211809,
"learning_rate": 8.683396524942656e-05,
"loss": 2.3903,
"step": 502
},
{
"epoch": 3.1052631578947367,
"grad_norm": 3.970245486136833,
"learning_rate": 8.677723054733162e-05,
"loss": 2.3571,
"step": 503
},
{
"epoch": 3.111455108359133,
"grad_norm": 4.532439692668108,
"learning_rate": 8.672039248833244e-05,
"loss": 2.3878,
"step": 504
},
{
"epoch": 3.1176470588235294,
"grad_norm": 4.033683253050144,
"learning_rate": 8.666345123216386e-05,
"loss": 2.4167,
"step": 505
},
{
"epoch": 3.123839009287926,
"grad_norm": 3.975201751116555,
"learning_rate": 8.66064069388508e-05,
"loss": 2.3517,
"step": 506
},
{
"epoch": 3.130030959752322,
"grad_norm": 3.3788260641137975,
"learning_rate": 8.654925976870766e-05,
"loss": 2.3212,
"step": 507
},
{
"epoch": 3.136222910216718,
"grad_norm": 3.9584592514363166,
"learning_rate": 8.649200988233806e-05,
"loss": 2.4028,
"step": 508
},
{
"epoch": 3.1424148606811144,
"grad_norm": 3.689388657105196,
"learning_rate": 8.643465744063419e-05,
"loss": 2.4571,
"step": 509
},
{
"epoch": 3.1486068111455108,
"grad_norm": 3.605974543644076,
"learning_rate": 8.637720260477656e-05,
"loss": 2.4103,
"step": 510
},
{
"epoch": 3.154798761609907,
"grad_norm": 3.8794947630259613,
"learning_rate": 8.631964553623337e-05,
"loss": 2.3767,
"step": 511
},
{
"epoch": 3.1609907120743035,
"grad_norm": 3.4775368857027193,
"learning_rate": 8.626198639676014e-05,
"loss": 2.3674,
"step": 512
},
{
"epoch": 3.1671826625387,
"grad_norm": 3.6633952894093467,
"learning_rate": 8.620422534839925e-05,
"loss": 2.4016,
"step": 513
},
{
"epoch": 3.173374613003096,
"grad_norm": 3.8137471386364568,
"learning_rate": 8.614636255347952e-05,
"loss": 2.3726,
"step": 514
},
{
"epoch": 3.179566563467492,
"grad_norm": 3.7359331492842753,
"learning_rate": 8.608839817461564e-05,
"loss": 2.362,
"step": 515
},
{
"epoch": 3.1857585139318885,
"grad_norm": 3.7777692904063915,
"learning_rate": 8.603033237470784e-05,
"loss": 2.403,
"step": 516
},
{
"epoch": 3.191950464396285,
"grad_norm": 4.18451321680891,
"learning_rate": 8.597216531694137e-05,
"loss": 2.3073,
"step": 517
},
{
"epoch": 3.198142414860681,
"grad_norm": 4.067292793649438,
"learning_rate": 8.591389716478604e-05,
"loss": 2.3289,
"step": 518
},
{
"epoch": 3.2043343653250775,
"grad_norm": 3.934930338281878,
"learning_rate": 8.585552808199576e-05,
"loss": 2.4575,
"step": 519
},
{
"epoch": 3.2105263157894735,
"grad_norm": 4.657392240081009,
"learning_rate": 8.579705823260813e-05,
"loss": 2.4673,
"step": 520
},
{
"epoch": 3.21671826625387,
"grad_norm": 3.684852355859821,
"learning_rate": 8.573848778094389e-05,
"loss": 2.4688,
"step": 521
},
{
"epoch": 3.222910216718266,
"grad_norm": 4.011742312465073,
"learning_rate": 8.567981689160654e-05,
"loss": 2.3442,
"step": 522
},
{
"epoch": 3.2291021671826625,
"grad_norm": 3.661871017054601,
"learning_rate": 8.562104572948185e-05,
"loss": 2.3394,
"step": 523
},
{
"epoch": 3.235294117647059,
"grad_norm": 4.123772593484842,
"learning_rate": 8.556217445973735e-05,
"loss": 2.571,
"step": 524
},
{
"epoch": 3.2414860681114552,
"grad_norm": 3.938828313810375,
"learning_rate": 8.550320324782197e-05,
"loss": 2.3857,
"step": 525
},
{
"epoch": 3.2476780185758516,
"grad_norm": 3.625870077036559,
"learning_rate": 8.544413225946545e-05,
"loss": 2.3553,
"step": 526
},
{
"epoch": 3.2538699690402475,
"grad_norm": 3.7489633832482023,
"learning_rate": 8.538496166067798e-05,
"loss": 2.402,
"step": 527
},
{
"epoch": 3.260061919504644,
"grad_norm": 3.8589651067740354,
"learning_rate": 8.532569161774968e-05,
"loss": 2.3985,
"step": 528
},
{
"epoch": 3.26625386996904,
"grad_norm": 5.456327193741036,
"learning_rate": 8.52663222972501e-05,
"loss": 2.5039,
"step": 529
},
{
"epoch": 3.2724458204334366,
"grad_norm": 4.052545897678689,
"learning_rate": 8.520685386602792e-05,
"loss": 2.4497,
"step": 530
},
{
"epoch": 3.278637770897833,
"grad_norm": 3.5992235741605474,
"learning_rate": 8.514728649121018e-05,
"loss": 2.3914,
"step": 531
},
{
"epoch": 3.2848297213622293,
"grad_norm": 3.432155774584645,
"learning_rate": 8.50876203402021e-05,
"loss": 2.386,
"step": 532
},
{
"epoch": 3.291021671826625,
"grad_norm": 3.8557093863718346,
"learning_rate": 8.502785558068648e-05,
"loss": 2.3473,
"step": 533
},
{
"epoch": 3.2972136222910216,
"grad_norm": 4.9666307995715915,
"learning_rate": 8.496799238062325e-05,
"loss": 2.4944,
"step": 534
},
{
"epoch": 3.303405572755418,
"grad_norm": 3.917374781482906,
"learning_rate": 8.490803090824895e-05,
"loss": 2.4824,
"step": 535
},
{
"epoch": 3.3095975232198143,
"grad_norm": 3.969181260137565,
"learning_rate": 8.484797133207633e-05,
"loss": 2.374,
"step": 536
},
{
"epoch": 3.3157894736842106,
"grad_norm": 3.7224610182620217,
"learning_rate": 8.478781382089387e-05,
"loss": 2.5348,
"step": 537
},
{
"epoch": 3.321981424148607,
"grad_norm": 4.1424980914113165,
"learning_rate": 8.472755854376521e-05,
"loss": 2.3643,
"step": 538
},
{
"epoch": 3.3281733746130033,
"grad_norm": 3.722762538292435,
"learning_rate": 8.466720567002885e-05,
"loss": 2.384,
"step": 539
},
{
"epoch": 3.3343653250773992,
"grad_norm": 3.8280649398435034,
"learning_rate": 8.460675536929748e-05,
"loss": 2.3841,
"step": 540
},
{
"epoch": 3.3405572755417956,
"grad_norm": 4.837273639103312,
"learning_rate": 8.454620781145762e-05,
"loss": 2.5264,
"step": 541
},
{
"epoch": 3.346749226006192,
"grad_norm": 4.19293715559688,
"learning_rate": 8.448556316666912e-05,
"loss": 2.3774,
"step": 542
},
{
"epoch": 3.3529411764705883,
"grad_norm": 3.5839163084269927,
"learning_rate": 8.442482160536469e-05,
"loss": 2.3416,
"step": 543
},
{
"epoch": 3.3591331269349847,
"grad_norm": 4.333258685944541,
"learning_rate": 8.436398329824942e-05,
"loss": 2.4427,
"step": 544
},
{
"epoch": 3.3653250773993806,
"grad_norm": 3.478712498678592,
"learning_rate": 8.430304841630023e-05,
"loss": 2.4323,
"step": 545
},
{
"epoch": 3.371517027863777,
"grad_norm": 4.32087845434783,
"learning_rate": 8.424201713076553e-05,
"loss": 2.3134,
"step": 546
},
{
"epoch": 3.3777089783281733,
"grad_norm": 3.9274514954655815,
"learning_rate": 8.41808896131646e-05,
"loss": 2.398,
"step": 547
},
{
"epoch": 3.3839009287925697,
"grad_norm": 4.053745527455987,
"learning_rate": 8.411966603528716e-05,
"loss": 2.3076,
"step": 548
},
{
"epoch": 3.390092879256966,
"grad_norm": 4.460782410782793,
"learning_rate": 8.405834656919295e-05,
"loss": 2.3321,
"step": 549
},
{
"epoch": 3.3962848297213624,
"grad_norm": 3.9192879825788647,
"learning_rate": 8.399693138721116e-05,
"loss": 2.3456,
"step": 550
},
{
"epoch": 3.4024767801857587,
"grad_norm": 4.239580955809909,
"learning_rate": 8.393542066193994e-05,
"loss": 2.4856,
"step": 551
},
{
"epoch": 3.4086687306501546,
"grad_norm": 3.707684771229425,
"learning_rate": 8.387381456624599e-05,
"loss": 2.3784,
"step": 552
},
{
"epoch": 3.414860681114551,
"grad_norm": 4.002189613200214,
"learning_rate": 8.381211327326402e-05,
"loss": 2.3785,
"step": 553
},
{
"epoch": 3.4210526315789473,
"grad_norm": 4.0245735243415535,
"learning_rate": 8.375031695639631e-05,
"loss": 2.4096,
"step": 554
},
{
"epoch": 3.4272445820433437,
"grad_norm": 3.570204013129324,
"learning_rate": 8.368842578931214e-05,
"loss": 2.4111,
"step": 555
},
{
"epoch": 3.43343653250774,
"grad_norm": 4.453628574002878,
"learning_rate": 8.362643994594739e-05,
"loss": 2.5287,
"step": 556
},
{
"epoch": 3.4396284829721364,
"grad_norm": 3.739785978737487,
"learning_rate": 8.356435960050397e-05,
"loss": 2.3796,
"step": 557
},
{
"epoch": 3.4458204334365323,
"grad_norm": 3.9547310234690323,
"learning_rate": 8.350218492744944e-05,
"loss": 2.3646,
"step": 558
},
{
"epoch": 3.4520123839009287,
"grad_norm": 3.5013107085457618,
"learning_rate": 8.343991610151641e-05,
"loss": 2.4255,
"step": 559
},
{
"epoch": 3.458204334365325,
"grad_norm": 3.7666231303992412,
"learning_rate": 8.337755329770207e-05,
"loss": 2.5381,
"step": 560
},
{
"epoch": 3.4643962848297214,
"grad_norm": 4.382382117785664,
"learning_rate": 8.331509669126778e-05,
"loss": 2.4004,
"step": 561
},
{
"epoch": 3.4705882352941178,
"grad_norm": 3.7395619287804345,
"learning_rate": 8.325254645773848e-05,
"loss": 2.3416,
"step": 562
},
{
"epoch": 3.476780185758514,
"grad_norm": 4.3553154299552705,
"learning_rate": 8.318990277290224e-05,
"loss": 2.3566,
"step": 563
},
{
"epoch": 3.4829721362229105,
"grad_norm": 3.8059837871427935,
"learning_rate": 8.312716581280976e-05,
"loss": 2.517,
"step": 564
},
{
"epoch": 3.4891640866873064,
"grad_norm": 3.784476713847252,
"learning_rate": 8.306433575377388e-05,
"loss": 2.4462,
"step": 565
},
{
"epoch": 3.4953560371517027,
"grad_norm": 3.628508905744923,
"learning_rate": 8.30014127723691e-05,
"loss": 2.4279,
"step": 566
},
{
"epoch": 3.501547987616099,
"grad_norm": 3.677250997486939,
"learning_rate": 8.293839704543104e-05,
"loss": 2.3862,
"step": 567
},
{
"epoch": 3.5077399380804954,
"grad_norm": 3.2429432477815503,
"learning_rate": 8.287528875005595e-05,
"loss": 2.5449,
"step": 568
},
{
"epoch": 3.513931888544892,
"grad_norm": 3.7827903049802085,
"learning_rate": 8.281208806360027e-05,
"loss": 2.5051,
"step": 569
},
{
"epoch": 3.5201238390092877,
"grad_norm": 3.7378668516720706,
"learning_rate": 8.274879516368007e-05,
"loss": 2.4024,
"step": 570
},
{
"epoch": 3.526315789473684,
"grad_norm": 3.4374901729582525,
"learning_rate": 8.268541022817058e-05,
"loss": 2.4265,
"step": 571
},
{
"epoch": 3.5325077399380804,
"grad_norm": 3.294348302203455,
"learning_rate": 8.262193343520568e-05,
"loss": 2.3196,
"step": 572
},
{
"epoch": 3.538699690402477,
"grad_norm": 3.390087246250273,
"learning_rate": 8.255836496317739e-05,
"loss": 2.33,
"step": 573
},
{
"epoch": 3.544891640866873,
"grad_norm": 3.6843958177256524,
"learning_rate": 8.24947049907354e-05,
"loss": 2.4453,
"step": 574
},
{
"epoch": 3.5510835913312695,
"grad_norm": 3.7638544633311866,
"learning_rate": 8.243095369678652e-05,
"loss": 2.359,
"step": 575
},
{
"epoch": 3.557275541795666,
"grad_norm": 3.774344465762464,
"learning_rate": 8.236711126049427e-05,
"loss": 2.4222,
"step": 576
},
{
"epoch": 3.5634674922600618,
"grad_norm": 3.671051399878454,
"learning_rate": 8.230317786127821e-05,
"loss": 2.5869,
"step": 577
},
{
"epoch": 3.569659442724458,
"grad_norm": 3.2906514603638803,
"learning_rate": 8.223915367881364e-05,
"loss": 2.3977,
"step": 578
},
{
"epoch": 3.5758513931888545,
"grad_norm": 3.4347910639301538,
"learning_rate": 8.217503889303089e-05,
"loss": 2.5456,
"step": 579
},
{
"epoch": 3.582043343653251,
"grad_norm": 3.800848071745359,
"learning_rate": 8.211083368411503e-05,
"loss": 2.4824,
"step": 580
},
{
"epoch": 3.588235294117647,
"grad_norm": 3.6359215248511174,
"learning_rate": 8.204653823250516e-05,
"loss": 2.3716,
"step": 581
},
{
"epoch": 3.594427244582043,
"grad_norm": 4.226736841356074,
"learning_rate": 8.198215271889405e-05,
"loss": 2.4543,
"step": 582
},
{
"epoch": 3.6006191950464395,
"grad_norm": 3.6133621782855165,
"learning_rate": 8.191767732422753e-05,
"loss": 2.4124,
"step": 583
},
{
"epoch": 3.606811145510836,
"grad_norm": 3.8540886045466007,
"learning_rate": 8.185311222970408e-05,
"loss": 2.3871,
"step": 584
},
{
"epoch": 3.613003095975232,
"grad_norm": 3.8474432290665472,
"learning_rate": 8.178845761677421e-05,
"loss": 2.4103,
"step": 585
},
{
"epoch": 3.6191950464396285,
"grad_norm": 4.626307948711146,
"learning_rate": 8.172371366714005e-05,
"loss": 2.4615,
"step": 586
},
{
"epoch": 3.625386996904025,
"grad_norm": 4.37767462664029,
"learning_rate": 8.165888056275478e-05,
"loss": 2.2778,
"step": 587
},
{
"epoch": 3.6315789473684212,
"grad_norm": 3.2021961110763173,
"learning_rate": 8.159395848582214e-05,
"loss": 2.3953,
"step": 588
},
{
"epoch": 3.6377708978328176,
"grad_norm": 3.617124402696774,
"learning_rate": 8.152894761879593e-05,
"loss": 2.3027,
"step": 589
},
{
"epoch": 3.6439628482972135,
"grad_norm": 4.484054406918155,
"learning_rate": 8.146384814437942e-05,
"loss": 2.5096,
"step": 590
},
{
"epoch": 3.65015479876161,
"grad_norm": 4.108253074688598,
"learning_rate": 8.139866024552501e-05,
"loss": 2.3963,
"step": 591
},
{
"epoch": 3.656346749226006,
"grad_norm": 3.884001309713194,
"learning_rate": 8.13333841054335e-05,
"loss": 2.3743,
"step": 592
},
{
"epoch": 3.6625386996904026,
"grad_norm": 3.2469329314047286,
"learning_rate": 8.126801990755371e-05,
"loss": 2.3849,
"step": 593
},
{
"epoch": 3.6687306501547985,
"grad_norm": 3.680814463784443,
"learning_rate": 8.120256783558193e-05,
"loss": 2.4079,
"step": 594
},
{
"epoch": 3.674922600619195,
"grad_norm": 3.7527216066173783,
"learning_rate": 8.113702807346146e-05,
"loss": 2.4133,
"step": 595
},
{
"epoch": 3.681114551083591,
"grad_norm": 3.9183133821763816,
"learning_rate": 8.107140080538196e-05,
"loss": 2.3375,
"step": 596
},
{
"epoch": 3.6873065015479876,
"grad_norm": 3.5880106416119575,
"learning_rate": 8.100568621577907e-05,
"loss": 2.4574,
"step": 597
},
{
"epoch": 3.693498452012384,
"grad_norm": 3.8184548274334236,
"learning_rate": 8.093988448933379e-05,
"loss": 2.3915,
"step": 598
},
{
"epoch": 3.6996904024767803,
"grad_norm": 4.008539397400194,
"learning_rate": 8.087399581097204e-05,
"loss": 2.4552,
"step": 599
},
{
"epoch": 3.7058823529411766,
"grad_norm": 3.611810991037467,
"learning_rate": 8.080802036586408e-05,
"loss": 2.5122,
"step": 600
},
{
"epoch": 3.712074303405573,
"grad_norm": 3.6007336687193416,
"learning_rate": 8.074195833942404e-05,
"loss": 2.2904,
"step": 601
},
{
"epoch": 3.718266253869969,
"grad_norm": 3.661330589005197,
"learning_rate": 8.067580991730939e-05,
"loss": 2.5167,
"step": 602
},
{
"epoch": 3.7244582043343653,
"grad_norm": 3.701053256190786,
"learning_rate": 8.060957528542032e-05,
"loss": 2.3615,
"step": 603
},
{
"epoch": 3.7306501547987616,
"grad_norm": 3.560815000825057,
"learning_rate": 8.054325462989938e-05,
"loss": 2.4362,
"step": 604
},
{
"epoch": 3.736842105263158,
"grad_norm": 3.6833443377659254,
"learning_rate": 8.047684813713086e-05,
"loss": 2.3669,
"step": 605
},
{
"epoch": 3.7430340557275543,
"grad_norm": 3.6944003385921436,
"learning_rate": 8.041035599374026e-05,
"loss": 2.39,
"step": 606
},
{
"epoch": 3.7492260061919502,
"grad_norm": 3.722950982457078,
"learning_rate": 8.034377838659379e-05,
"loss": 2.4393,
"step": 607
},
{
"epoch": 3.7554179566563466,
"grad_norm": 3.391406440568772,
"learning_rate": 8.027711550279788e-05,
"loss": 2.4148,
"step": 608
},
{
"epoch": 3.761609907120743,
"grad_norm": 4.116410869786648,
"learning_rate": 8.021036752969859e-05,
"loss": 2.4822,
"step": 609
},
{
"epoch": 3.7678018575851393,
"grad_norm": 3.806836392232153,
"learning_rate": 8.014353465488109e-05,
"loss": 2.4317,
"step": 610
},
{
"epoch": 3.7739938080495357,
"grad_norm": 3.7907290625778396,
"learning_rate": 8.007661706616918e-05,
"loss": 2.5755,
"step": 611
},
{
"epoch": 3.780185758513932,
"grad_norm": 3.857390237638102,
"learning_rate": 8.000961495162475e-05,
"loss": 2.5528,
"step": 612
},
{
"epoch": 3.7863777089783284,
"grad_norm": 3.691317382868572,
"learning_rate": 7.99425284995472e-05,
"loss": 2.3678,
"step": 613
},
{
"epoch": 3.7925696594427247,
"grad_norm": 4.22434396093653,
"learning_rate": 7.987535789847297e-05,
"loss": 2.375,
"step": 614
},
{
"epoch": 3.7987616099071206,
"grad_norm": 4.203684424148445,
"learning_rate": 7.980810333717499e-05,
"loss": 2.3522,
"step": 615
},
{
"epoch": 3.804953560371517,
"grad_norm": 3.5184765286157664,
"learning_rate": 7.974076500466215e-05,
"loss": 2.498,
"step": 616
},
{
"epoch": 3.8111455108359134,
"grad_norm": 4.091552173859175,
"learning_rate": 7.967334309017875e-05,
"loss": 2.5066,
"step": 617
},
{
"epoch": 3.8173374613003097,
"grad_norm": 3.7460766300362502,
"learning_rate": 7.960583778320398e-05,
"loss": 2.4016,
"step": 618
},
{
"epoch": 3.8235294117647056,
"grad_norm": 3.7076259071585373,
"learning_rate": 7.953824927345145e-05,
"loss": 2.4761,
"step": 619
},
{
"epoch": 3.829721362229102,
"grad_norm": 3.627102527680507,
"learning_rate": 7.947057775086852e-05,
"loss": 2.418,
"step": 620
},
{
"epoch": 3.8359133126934983,
"grad_norm": 3.596902423154603,
"learning_rate": 7.940282340563585e-05,
"loss": 2.3954,
"step": 621
},
{
"epoch": 3.8421052631578947,
"grad_norm": 3.7747227866779784,
"learning_rate": 7.933498642816697e-05,
"loss": 2.4548,
"step": 622
},
{
"epoch": 3.848297213622291,
"grad_norm": 4.023077395134154,
"learning_rate": 7.926706700910749e-05,
"loss": 2.4369,
"step": 623
},
{
"epoch": 3.8544891640866874,
"grad_norm": 4.038169577952473,
"learning_rate": 7.91990653393348e-05,
"loss": 2.5278,
"step": 624
},
{
"epoch": 3.8606811145510838,
"grad_norm": 3.8358270442024462,
"learning_rate": 7.913098160995742e-05,
"loss": 2.4348,
"step": 625
},
{
"epoch": 3.86687306501548,
"grad_norm": 4.151270695289064,
"learning_rate": 7.906281601231448e-05,
"loss": 2.4865,
"step": 626
},
{
"epoch": 3.873065015479876,
"grad_norm": 3.532376841609081,
"learning_rate": 7.89945687379752e-05,
"loss": 2.4158,
"step": 627
},
{
"epoch": 3.8792569659442724,
"grad_norm": 3.386589106491462,
"learning_rate": 7.892623997873832e-05,
"loss": 2.4313,
"step": 628
},
{
"epoch": 3.8854489164086687,
"grad_norm": 3.2592172663637657,
"learning_rate": 7.885782992663162e-05,
"loss": 2.3851,
"step": 629
},
{
"epoch": 3.891640866873065,
"grad_norm": 3.83030056141294,
"learning_rate": 7.87893387739113e-05,
"loss": 2.3898,
"step": 630
},
{
"epoch": 3.8978328173374615,
"grad_norm": 3.5379543160228244,
"learning_rate": 7.87207667130615e-05,
"loss": 2.3416,
"step": 631
},
{
"epoch": 3.9040247678018574,
"grad_norm": 3.9003164430953197,
"learning_rate": 7.865211393679373e-05,
"loss": 2.3958,
"step": 632
},
{
"epoch": 3.9102167182662537,
"grad_norm": 4.045214436387836,
"learning_rate": 7.858338063804638e-05,
"loss": 2.3821,
"step": 633
},
{
"epoch": 3.91640866873065,
"grad_norm": 3.772302312812738,
"learning_rate": 7.851456700998405e-05,
"loss": 2.4092,
"step": 634
},
{
"epoch": 3.9226006191950464,
"grad_norm": 4.002774187817698,
"learning_rate": 7.844567324599719e-05,
"loss": 2.3969,
"step": 635
},
{
"epoch": 3.928792569659443,
"grad_norm": 4.175141584948075,
"learning_rate": 7.83766995397014e-05,
"loss": 2.4838,
"step": 636
},
{
"epoch": 3.934984520123839,
"grad_norm": 3.9596695224737126,
"learning_rate": 7.830764608493697e-05,
"loss": 2.4405,
"step": 637
},
{
"epoch": 3.9411764705882355,
"grad_norm": 3.3284358322965133,
"learning_rate": 7.823851307576828e-05,
"loss": 2.3657,
"step": 638
},
{
"epoch": 3.9473684210526314,
"grad_norm": 3.2833880729162614,
"learning_rate": 7.816930070648334e-05,
"loss": 2.4343,
"step": 639
},
{
"epoch": 3.9535603715170278,
"grad_norm": 3.5684648107333707,
"learning_rate": 7.810000917159315e-05,
"loss": 2.4331,
"step": 640
},
{
"epoch": 3.959752321981424,
"grad_norm": 3.447912535908917,
"learning_rate": 7.803063866583118e-05,
"loss": 2.3603,
"step": 641
},
{
"epoch": 3.9659442724458205,
"grad_norm": 3.7491120649608445,
"learning_rate": 7.796118938415289e-05,
"loss": 2.5134,
"step": 642
},
{
"epoch": 3.972136222910217,
"grad_norm": 3.821071728011824,
"learning_rate": 7.789166152173509e-05,
"loss": 2.2568,
"step": 643
},
{
"epoch": 3.9783281733746128,
"grad_norm": 3.4413072268850855,
"learning_rate": 7.782205527397541e-05,
"loss": 2.3745,
"step": 644
},
{
"epoch": 3.984520123839009,
"grad_norm": 3.667189769208347,
"learning_rate": 7.775237083649181e-05,
"loss": 2.5052,
"step": 645
},
{
"epoch": 3.9907120743034055,
"grad_norm": 4.100509084367406,
"learning_rate": 7.7682608405122e-05,
"loss": 2.4049,
"step": 646
},
{
"epoch": 3.996904024767802,
"grad_norm": 3.6578604042215113,
"learning_rate": 7.761276817592282e-05,
"loss": 2.3555,
"step": 647
},
{
"epoch": 4.0,
"grad_norm": 3.5238815757576116,
"learning_rate": 7.75428503451698e-05,
"loss": 1.1803,
"step": 648
},
{
"epoch": 4.006191950464396,
"grad_norm": 3.1639437658042695,
"learning_rate": 7.747285510935654e-05,
"loss": 2.1137,
"step": 649
},
{
"epoch": 4.012383900928793,
"grad_norm": 3.457768945384432,
"learning_rate": 7.74027826651942e-05,
"loss": 2.1498,
"step": 650
},
{
"epoch": 4.018575851393189,
"grad_norm": 3.70029048006138,
"learning_rate": 7.733263320961085e-05,
"loss": 2.1948,
"step": 651
},
{
"epoch": 4.024767801857585,
"grad_norm": 4.126029827412395,
"learning_rate": 7.726240693975111e-05,
"loss": 2.15,
"step": 652
},
{
"epoch": 4.030959752321982,
"grad_norm": 5.558797321581609,
"learning_rate": 7.719210405297537e-05,
"loss": 2.1257,
"step": 653
},
{
"epoch": 4.037151702786378,
"grad_norm": 4.85006309138657,
"learning_rate": 7.712172474685935e-05,
"loss": 2.1376,
"step": 654
},
{
"epoch": 4.043343653250774,
"grad_norm": 6.4395438813022805,
"learning_rate": 7.705126921919359e-05,
"loss": 2.1444,
"step": 655
},
{
"epoch": 4.04953560371517,
"grad_norm": 6.041865846725888,
"learning_rate": 7.698073766798281e-05,
"loss": 2.0954,
"step": 656
},
{
"epoch": 4.055727554179566,
"grad_norm": 4.438802002995133,
"learning_rate": 7.691013029144536e-05,
"loss": 2.112,
"step": 657
},
{
"epoch": 4.061919504643963,
"grad_norm": 4.404903290145634,
"learning_rate": 7.683944728801273e-05,
"loss": 2.0828,
"step": 658
},
{
"epoch": 4.068111455108359,
"grad_norm": 4.778267563170224,
"learning_rate": 7.676868885632893e-05,
"loss": 2.181,
"step": 659
},
{
"epoch": 4.074303405572755,
"grad_norm": 4.459434975412655,
"learning_rate": 7.669785519524993e-05,
"loss": 2.0955,
"step": 660
},
{
"epoch": 4.080495356037152,
"grad_norm": 4.076696574359944,
"learning_rate": 7.662694650384315e-05,
"loss": 2.1908,
"step": 661
},
{
"epoch": 4.086687306501548,
"grad_norm": 4.145823898422436,
"learning_rate": 7.655596298138683e-05,
"loss": 2.1311,
"step": 662
},
{
"epoch": 4.0928792569659445,
"grad_norm": 4.557226017811323,
"learning_rate": 7.648490482736959e-05,
"loss": 2.1075,
"step": 663
},
{
"epoch": 4.099071207430341,
"grad_norm": 5.560413558024817,
"learning_rate": 7.641377224148971e-05,
"loss": 2.1539,
"step": 664
},
{
"epoch": 4.105263157894737,
"grad_norm": 4.87783371849139,
"learning_rate": 7.634256542365468e-05,
"loss": 2.1402,
"step": 665
},
{
"epoch": 4.1114551083591335,
"grad_norm": 5.246452612368868,
"learning_rate": 7.62712845739806e-05,
"loss": 2.1642,
"step": 666
},
{
"epoch": 4.117647058823529,
"grad_norm": 4.779665362725869,
"learning_rate": 7.619992989279167e-05,
"loss": 2.0213,
"step": 667
},
{
"epoch": 4.123839009287925,
"grad_norm": 4.4538999563034425,
"learning_rate": 7.61285015806195e-05,
"loss": 2.086,
"step": 668
},
{
"epoch": 4.130030959752322,
"grad_norm": 4.834496011319403,
"learning_rate": 7.605699983820269e-05,
"loss": 2.1693,
"step": 669
},
{
"epoch": 4.136222910216718,
"grad_norm": 3.9702391292351953,
"learning_rate": 7.598542486648623e-05,
"loss": 2.033,
"step": 670
},
{
"epoch": 4.142414860681114,
"grad_norm": 4.5764661716020205,
"learning_rate": 7.591377686662081e-05,
"loss": 2.0898,
"step": 671
},
{
"epoch": 4.148606811145511,
"grad_norm": 4.346261805320313,
"learning_rate": 7.584205603996246e-05,
"loss": 2.1802,
"step": 672
},
{
"epoch": 4.154798761609907,
"grad_norm": 5.188414338106125,
"learning_rate": 7.577026258807181e-05,
"loss": 2.1883,
"step": 673
},
{
"epoch": 4.1609907120743035,
"grad_norm": 4.5978526365501695,
"learning_rate": 7.569839671271359e-05,
"loss": 1.9735,
"step": 674
},
{
"epoch": 4.1671826625387,
"grad_norm": 4.487198124539332,
"learning_rate": 7.562645861585616e-05,
"loss": 1.9817,
"step": 675
},
{
"epoch": 4.173374613003096,
"grad_norm": 4.7334745708324855,
"learning_rate": 7.555444849967073e-05,
"loss": 2.0571,
"step": 676
},
{
"epoch": 4.179566563467493,
"grad_norm": 5.372008408502007,
"learning_rate": 7.548236656653095e-05,
"loss": 2.1141,
"step": 677
},
{
"epoch": 4.185758513931889,
"grad_norm": 5.144798634366983,
"learning_rate": 7.541021301901234e-05,
"loss": 2.2067,
"step": 678
},
{
"epoch": 4.191950464396285,
"grad_norm": 4.924453474716802,
"learning_rate": 7.533798805989164e-05,
"loss": 2.0421,
"step": 679
},
{
"epoch": 4.198142414860681,
"grad_norm": 4.991735540662648,
"learning_rate": 7.526569189214627e-05,
"loss": 1.9691,
"step": 680
},
{
"epoch": 4.204334365325077,
"grad_norm": 5.092517221614254,
"learning_rate": 7.519332471895384e-05,
"loss": 2.0528,
"step": 681
},
{
"epoch": 4.2105263157894735,
"grad_norm": 5.268399238043823,
"learning_rate": 7.512088674369143e-05,
"loss": 2.112,
"step": 682
},
{
"epoch": 4.21671826625387,
"grad_norm": 4.514936671947266,
"learning_rate": 7.504837816993514e-05,
"loss": 2.1414,
"step": 683
},
{
"epoch": 4.222910216718266,
"grad_norm": 4.674392659937627,
"learning_rate": 7.497579920145945e-05,
"loss": 2.2437,
"step": 684
},
{
"epoch": 4.2291021671826625,
"grad_norm": 4.143299960033866,
"learning_rate": 7.490315004223672e-05,
"loss": 2.1199,
"step": 685
},
{
"epoch": 4.235294117647059,
"grad_norm": 4.38729868716052,
"learning_rate": 7.483043089643653e-05,
"loss": 2.167,
"step": 686
},
{
"epoch": 4.241486068111455,
"grad_norm": 4.39687107676024,
"learning_rate": 7.475764196842516e-05,
"loss": 2.1273,
"step": 687
},
{
"epoch": 4.247678018575852,
"grad_norm": 5.0233257992601335,
"learning_rate": 7.468478346276499e-05,
"loss": 2.0826,
"step": 688
},
{
"epoch": 4.253869969040248,
"grad_norm": 12.022154855461501,
"learning_rate": 7.4611855584214e-05,
"loss": 2.1096,
"step": 689
},
{
"epoch": 4.260061919504644,
"grad_norm": 4.873247412293039,
"learning_rate": 7.453885853772503e-05,
"loss": 2.2862,
"step": 690
},
{
"epoch": 4.266253869969041,
"grad_norm": 4.969070770475527,
"learning_rate": 7.446579252844535e-05,
"loss": 2.0693,
"step": 691
},
{
"epoch": 4.272445820433436,
"grad_norm": 5.014674488114116,
"learning_rate": 7.439265776171611e-05,
"loss": 2.0325,
"step": 692
},
{
"epoch": 4.2786377708978325,
"grad_norm": 4.693562745355228,
"learning_rate": 7.431945444307157e-05,
"loss": 2.0993,
"step": 693
},
{
"epoch": 4.284829721362229,
"grad_norm": 5.091836740664018,
"learning_rate": 7.424618277823873e-05,
"loss": 2.1701,
"step": 694
},
{
"epoch": 4.291021671826625,
"grad_norm": 4.825237510682655,
"learning_rate": 7.417284297313664e-05,
"loss": 2.2776,
"step": 695
},
{
"epoch": 4.2972136222910216,
"grad_norm": 4.748682210483545,
"learning_rate": 7.409943523387586e-05,
"loss": 2.1297,
"step": 696
},
{
"epoch": 4.303405572755418,
"grad_norm": 4.0103272586261065,
"learning_rate": 7.402595976675785e-05,
"loss": 2.1662,
"step": 697
},
{
"epoch": 4.309597523219814,
"grad_norm": 4.275615365901843,
"learning_rate": 7.395241677827438e-05,
"loss": 2.2649,
"step": 698
},
{
"epoch": 4.315789473684211,
"grad_norm": 4.569190967353905,
"learning_rate": 7.387880647510709e-05,
"loss": 1.9498,
"step": 699
},
{
"epoch": 4.321981424148607,
"grad_norm": 3.999971515413252,
"learning_rate": 7.380512906412667e-05,
"loss": 2.0736,
"step": 700
},
{
"epoch": 4.328173374613003,
"grad_norm": 4.799921274732086,
"learning_rate": 7.373138475239249e-05,
"loss": 2.1249,
"step": 701
},
{
"epoch": 4.3343653250774,
"grad_norm": 5.252410492234698,
"learning_rate": 7.365757374715187e-05,
"loss": 2.1682,
"step": 702
},
{
"epoch": 4.340557275541796,
"grad_norm": 4.680441025779536,
"learning_rate": 7.358369625583965e-05,
"loss": 2.1072,
"step": 703
},
{
"epoch": 4.346749226006192,
"grad_norm": 5.346243143190861,
"learning_rate": 7.350975248607743e-05,
"loss": 2.1245,
"step": 704
},
{
"epoch": 4.352941176470588,
"grad_norm": 5.235167621935557,
"learning_rate": 7.34357426456731e-05,
"loss": 1.9862,
"step": 705
},
{
"epoch": 4.359133126934984,
"grad_norm": 4.569103725568221,
"learning_rate": 7.336166694262028e-05,
"loss": 2.0838,
"step": 706
},
{
"epoch": 4.365325077399381,
"grad_norm": 5.117945643586851,
"learning_rate": 7.328752558509761e-05,
"loss": 2.2451,
"step": 707
},
{
"epoch": 4.371517027863777,
"grad_norm": 5.125143550125818,
"learning_rate": 7.321331878146834e-05,
"loss": 2.1555,
"step": 708
},
{
"epoch": 4.377708978328173,
"grad_norm": 4.483992154106004,
"learning_rate": 7.313904674027954e-05,
"loss": 2.0191,
"step": 709
},
{
"epoch": 4.38390092879257,
"grad_norm": 4.574311355191686,
"learning_rate": 7.306470967026169e-05,
"loss": 2.2059,
"step": 710
},
{
"epoch": 4.390092879256966,
"grad_norm": 4.214723810324921,
"learning_rate": 7.299030778032799e-05,
"loss": 2.1897,
"step": 711
},
{
"epoch": 4.396284829721362,
"grad_norm": 4.122839871149707,
"learning_rate": 7.291584127957384e-05,
"loss": 2.2548,
"step": 712
},
{
"epoch": 4.402476780185759,
"grad_norm": 3.6452057194340504,
"learning_rate": 7.284131037727618e-05,
"loss": 2.1921,
"step": 713
},
{
"epoch": 4.408668730650155,
"grad_norm": 4.168442986791581,
"learning_rate": 7.276671528289299e-05,
"loss": 2.2092,
"step": 714
},
{
"epoch": 4.414860681114551,
"grad_norm": 4.773183982640468,
"learning_rate": 7.269205620606259e-05,
"loss": 2.1974,
"step": 715
},
{
"epoch": 4.421052631578947,
"grad_norm": 4.617784023066425,
"learning_rate": 7.261733335660317e-05,
"loss": 2.1342,
"step": 716
},
{
"epoch": 4.427244582043343,
"grad_norm": 5.1080835688057125,
"learning_rate": 7.25425469445121e-05,
"loss": 2.0862,
"step": 717
},
{
"epoch": 4.43343653250774,
"grad_norm": 4.832245298522228,
"learning_rate": 7.246769717996539e-05,
"loss": 2.141,
"step": 718
},
{
"epoch": 4.439628482972136,
"grad_norm": 4.680422668639977,
"learning_rate": 7.239278427331717e-05,
"loss": 2.1117,
"step": 719
},
{
"epoch": 4.445820433436532,
"grad_norm": 4.5217290454953165,
"learning_rate": 7.231780843509889e-05,
"loss": 2.1004,
"step": 720
},
{
"epoch": 4.452012383900929,
"grad_norm": 4.450258321106715,
"learning_rate": 7.224276987601895e-05,
"loss": 2.2117,
"step": 721
},
{
"epoch": 4.458204334365325,
"grad_norm": 4.203679275635295,
"learning_rate": 7.216766880696199e-05,
"loss": 2.1465,
"step": 722
},
{
"epoch": 4.464396284829721,
"grad_norm": 4.355702552457598,
"learning_rate": 7.209250543898834e-05,
"loss": 2.077,
"step": 723
},
{
"epoch": 4.470588235294118,
"grad_norm": 4.1943319549819655,
"learning_rate": 7.201727998333336e-05,
"loss": 2.1524,
"step": 724
},
{
"epoch": 4.476780185758514,
"grad_norm": 4.613968723058964,
"learning_rate": 7.1941992651407e-05,
"loss": 2.1386,
"step": 725
},
{
"epoch": 4.4829721362229105,
"grad_norm": 4.567908592761703,
"learning_rate": 7.1866643654793e-05,
"loss": 2.1904,
"step": 726
},
{
"epoch": 4.489164086687307,
"grad_norm": 5.156072091242819,
"learning_rate": 7.179123320524848e-05,
"loss": 2.0619,
"step": 727
},
{
"epoch": 4.495356037151703,
"grad_norm": 4.8433809937701655,
"learning_rate": 7.171576151470318e-05,
"loss": 2.0997,
"step": 728
},
{
"epoch": 4.5015479876160995,
"grad_norm": 5.27251370417393,
"learning_rate": 7.164022879525902e-05,
"loss": 2.0835,
"step": 729
},
{
"epoch": 4.507739938080495,
"grad_norm": 4.535212409355026,
"learning_rate": 7.156463525918942e-05,
"loss": 2.2414,
"step": 730
},
{
"epoch": 4.513931888544891,
"grad_norm": 4.961853408192757,
"learning_rate": 7.148898111893867e-05,
"loss": 2.2253,
"step": 731
},
{
"epoch": 4.520123839009288,
"grad_norm": 4.139010929875864,
"learning_rate": 7.141326658712143e-05,
"loss": 2.1412,
"step": 732
},
{
"epoch": 4.526315789473684,
"grad_norm": 4.878561442903061,
"learning_rate": 7.133749187652208e-05,
"loss": 2.2329,
"step": 733
},
{
"epoch": 4.53250773993808,
"grad_norm": 4.446855558161751,
"learning_rate": 7.126165720009406e-05,
"loss": 2.2213,
"step": 734
},
{
"epoch": 4.538699690402477,
"grad_norm": 4.824724533257683,
"learning_rate": 7.118576277095944e-05,
"loss": 2.1342,
"step": 735
},
{
"epoch": 4.544891640866873,
"grad_norm": 4.113461661901802,
"learning_rate": 7.110980880240814e-05,
"loss": 2.0658,
"step": 736
},
{
"epoch": 4.5510835913312695,
"grad_norm": 4.581936636361705,
"learning_rate": 7.10337955078974e-05,
"loss": 2.197,
"step": 737
},
{
"epoch": 4.557275541795666,
"grad_norm": 4.943239352273858,
"learning_rate": 7.095772310105124e-05,
"loss": 2.2231,
"step": 738
},
{
"epoch": 4.563467492260062,
"grad_norm": 4.870911601918529,
"learning_rate": 7.088159179565977e-05,
"loss": 2.2161,
"step": 739
},
{
"epoch": 4.569659442724459,
"grad_norm": 4.6578229886289195,
"learning_rate": 7.080540180567862e-05,
"loss": 2.1472,
"step": 740
},
{
"epoch": 4.575851393188854,
"grad_norm": 4.55833159189419,
"learning_rate": 7.07291533452284e-05,
"loss": 2.1873,
"step": 741
},
{
"epoch": 4.58204334365325,
"grad_norm": 4.7889800503626265,
"learning_rate": 7.065284662859395e-05,
"loss": 2.1148,
"step": 742
},
{
"epoch": 4.588235294117647,
"grad_norm": 4.511730121280619,
"learning_rate": 7.05764818702239e-05,
"loss": 2.2623,
"step": 743
},
{
"epoch": 4.594427244582043,
"grad_norm": 4.343408072498585,
"learning_rate": 7.050005928473e-05,
"loss": 2.1682,
"step": 744
},
{
"epoch": 4.6006191950464395,
"grad_norm": 4.44509304970608,
"learning_rate": 7.042357908688646e-05,
"loss": 2.134,
"step": 745
},
{
"epoch": 4.606811145510836,
"grad_norm": 4.581179901763817,
"learning_rate": 7.034704149162944e-05,
"loss": 2.1382,
"step": 746
},
{
"epoch": 4.613003095975232,
"grad_norm": 3.9424465575413827,
"learning_rate": 7.027044671405643e-05,
"loss": 2.2045,
"step": 747
},
{
"epoch": 4.6191950464396285,
"grad_norm": 4.348375839574913,
"learning_rate": 7.019379496942556e-05,
"loss": 2.1715,
"step": 748
},
{
"epoch": 4.625386996904025,
"grad_norm": 4.1548331420612215,
"learning_rate": 7.011708647315509e-05,
"loss": 2.2015,
"step": 749
},
{
"epoch": 4.631578947368421,
"grad_norm": 4.6915592822878045,
"learning_rate": 7.004032144082281e-05,
"loss": 2.2613,
"step": 750
},
{
"epoch": 4.637770897832818,
"grad_norm": 4.630737026765936,
"learning_rate": 6.996350008816532e-05,
"loss": 2.291,
"step": 751
},
{
"epoch": 4.643962848297214,
"grad_norm": 4.378714794510802,
"learning_rate": 6.988662263107754e-05,
"loss": 2.0821,
"step": 752
},
{
"epoch": 4.65015479876161,
"grad_norm": 5.107819820070701,
"learning_rate": 6.980968928561209e-05,
"loss": 2.174,
"step": 753
},
{
"epoch": 4.656346749226007,
"grad_norm": 4.45323535034182,
"learning_rate": 6.97327002679786e-05,
"loss": 2.1216,
"step": 754
},
{
"epoch": 4.662538699690402,
"grad_norm": 4.618964933752295,
"learning_rate": 6.965565579454322e-05,
"loss": 2.1177,
"step": 755
},
{
"epoch": 4.6687306501547985,
"grad_norm": 4.00692278763202,
"learning_rate": 6.957855608182787e-05,
"loss": 2.0692,
"step": 756
},
{
"epoch": 4.674922600619195,
"grad_norm": 4.568606275897397,
"learning_rate": 6.950140134650979e-05,
"loss": 2.1255,
"step": 757
},
{
"epoch": 4.681114551083591,
"grad_norm": 4.7575605136273005,
"learning_rate": 6.942419180542081e-05,
"loss": 2.1642,
"step": 758
},
{
"epoch": 4.687306501547988,
"grad_norm": 4.31453860474588,
"learning_rate": 6.93469276755468e-05,
"loss": 2.2458,
"step": 759
},
{
"epoch": 4.693498452012384,
"grad_norm": 4.929465346006104,
"learning_rate": 6.926960917402701e-05,
"loss": 2.2451,
"step": 760
},
{
"epoch": 4.69969040247678,
"grad_norm": 4.542828192154121,
"learning_rate": 6.919223651815356e-05,
"loss": 2.1707,
"step": 761
},
{
"epoch": 4.705882352941177,
"grad_norm": 4.087682222344042,
"learning_rate": 6.911480992537071e-05,
"loss": 2.2478,
"step": 762
},
{
"epoch": 4.712074303405573,
"grad_norm": 4.253186811157624,
"learning_rate": 6.903732961327432e-05,
"loss": 2.2018,
"step": 763
},
{
"epoch": 4.718266253869969,
"grad_norm": 4.2362550214015595,
"learning_rate": 6.895979579961119e-05,
"loss": 2.2328,
"step": 764
},
{
"epoch": 4.724458204334366,
"grad_norm": 5.196159466568116,
"learning_rate": 6.888220870227853e-05,
"loss": 2.1508,
"step": 765
},
{
"epoch": 4.730650154798761,
"grad_norm": 3.8878861676649175,
"learning_rate": 6.880456853932326e-05,
"loss": 2.1151,
"step": 766
},
{
"epoch": 4.7368421052631575,
"grad_norm": 4.329023340754469,
"learning_rate": 6.872687552894145e-05,
"loss": 2.2116,
"step": 767
},
{
"epoch": 4.743034055727554,
"grad_norm": 4.335216771787078,
"learning_rate": 6.864912988947767e-05,
"loss": 2.1063,
"step": 768
},
{
"epoch": 4.74922600619195,
"grad_norm": 4.280907354310136,
"learning_rate": 6.857133183942442e-05,
"loss": 2.1185,
"step": 769
},
{
"epoch": 4.755417956656347,
"grad_norm": 4.621438296367508,
"learning_rate": 6.849348159742146e-05,
"loss": 2.0787,
"step": 770
},
{
"epoch": 4.761609907120743,
"grad_norm": 4.836976318228352,
"learning_rate": 6.841557938225527e-05,
"loss": 1.9922,
"step": 771
},
{
"epoch": 4.767801857585139,
"grad_norm": 4.89566581269698,
"learning_rate": 6.833762541285836e-05,
"loss": 2.1255,
"step": 772
},
{
"epoch": 4.773993808049536,
"grad_norm": 4.587551174655151,
"learning_rate": 6.82596199083087e-05,
"loss": 2.0862,
"step": 773
},
{
"epoch": 4.780185758513932,
"grad_norm": 4.098948304566244,
"learning_rate": 6.818156308782911e-05,
"loss": 2.1175,
"step": 774
},
{
"epoch": 4.786377708978328,
"grad_norm": 4.627078810465172,
"learning_rate": 6.810345517078657e-05,
"loss": 1.9263,
"step": 775
},
{
"epoch": 4.792569659442725,
"grad_norm": 4.446918518406074,
"learning_rate": 6.80252963766917e-05,
"loss": 2.154,
"step": 776
},
{
"epoch": 4.798761609907121,
"grad_norm": 4.250857173022367,
"learning_rate": 6.794708692519815e-05,
"loss": 2.1464,
"step": 777
},
{
"epoch": 4.804953560371517,
"grad_norm": 4.0136218324719595,
"learning_rate": 6.786882703610182e-05,
"loss": 2.0561,
"step": 778
},
{
"epoch": 4.811145510835914,
"grad_norm": 4.202736014858293,
"learning_rate": 6.779051692934042e-05,
"loss": 2.1862,
"step": 779
},
{
"epoch": 4.817337461300309,
"grad_norm": 4.282532848878059,
"learning_rate": 6.771215682499284e-05,
"loss": 2.1776,
"step": 780
},
{
"epoch": 4.823529411764706,
"grad_norm": 4.515107712023646,
"learning_rate": 6.76337469432784e-05,
"loss": 2.1185,
"step": 781
},
{
"epoch": 4.829721362229102,
"grad_norm": 3.9819214384755806,
"learning_rate": 6.755528750455634e-05,
"loss": 2.2076,
"step": 782
},
{
"epoch": 4.835913312693498,
"grad_norm": 4.6248577553741566,
"learning_rate": 6.747677872932518e-05,
"loss": 2.1507,
"step": 783
},
{
"epoch": 4.842105263157895,
"grad_norm": 4.768420241307226,
"learning_rate": 6.739822083822208e-05,
"loss": 2.2218,
"step": 784
},
{
"epoch": 4.848297213622291,
"grad_norm": 4.486527686954089,
"learning_rate": 6.731961405202224e-05,
"loss": 2.1709,
"step": 785
},
{
"epoch": 4.854489164086687,
"grad_norm": 4.429697591756185,
"learning_rate": 6.724095859163829e-05,
"loss": 2.1713,
"step": 786
},
{
"epoch": 4.860681114551084,
"grad_norm": 4.389569136042767,
"learning_rate": 6.716225467811961e-05,
"loss": 2.0795,
"step": 787
},
{
"epoch": 4.86687306501548,
"grad_norm": 4.715430340153762,
"learning_rate": 6.70835025326518e-05,
"loss": 2.1989,
"step": 788
},
{
"epoch": 4.8730650154798765,
"grad_norm": 4.7231462421730095,
"learning_rate": 6.700470237655596e-05,
"loss": 2.1325,
"step": 789
},
{
"epoch": 4.879256965944273,
"grad_norm": 4.545669489130173,
"learning_rate": 6.692585443128813e-05,
"loss": 2.0339,
"step": 790
},
{
"epoch": 4.885448916408668,
"grad_norm": 4.756837089506344,
"learning_rate": 6.68469589184387e-05,
"loss": 1.984,
"step": 791
},
{
"epoch": 4.891640866873065,
"grad_norm": 4.850076718618227,
"learning_rate": 6.676801605973169e-05,
"loss": 2.0877,
"step": 792
},
{
"epoch": 4.897832817337461,
"grad_norm": 4.587077738976434,
"learning_rate": 6.668902607702419e-05,
"loss": 2.1371,
"step": 793
},
{
"epoch": 4.904024767801857,
"grad_norm": 4.588091204300823,
"learning_rate": 6.660998919230572e-05,
"loss": 2.1427,
"step": 794
},
{
"epoch": 4.910216718266254,
"grad_norm": 4.912079545017205,
"learning_rate": 6.653090562769763e-05,
"loss": 2.1679,
"step": 795
},
{
"epoch": 4.91640866873065,
"grad_norm": 3.9804465702676044,
"learning_rate": 6.645177560545245e-05,
"loss": 2.2245,
"step": 796
},
{
"epoch": 4.922600619195046,
"grad_norm": 4.144902689757148,
"learning_rate": 6.637259934795327e-05,
"loss": 2.0949,
"step": 797
},
{
"epoch": 4.928792569659443,
"grad_norm": 4.456739380615653,
"learning_rate": 6.62933770777131e-05,
"loss": 2.1302,
"step": 798
},
{
"epoch": 4.934984520123839,
"grad_norm": 4.415767596457962,
"learning_rate": 6.62141090173743e-05,
"loss": 2.0592,
"step": 799
},
{
"epoch": 4.9411764705882355,
"grad_norm": 4.359927642233302,
"learning_rate": 6.613479538970789e-05,
"loss": 2.0891,
"step": 800
},
{
"epoch": 4.947368421052632,
"grad_norm": 4.204704189524659,
"learning_rate": 6.605543641761292e-05,
"loss": 2.1229,
"step": 801
},
{
"epoch": 4.953560371517028,
"grad_norm": 4.30895071405075,
"learning_rate": 6.597603232411597e-05,
"loss": 2.2005,
"step": 802
},
{
"epoch": 4.959752321981425,
"grad_norm": 3.9577540550312396,
"learning_rate": 6.589658333237032e-05,
"loss": 2.2028,
"step": 803
},
{
"epoch": 4.965944272445821,
"grad_norm": 4.49768446217552,
"learning_rate": 6.581708966565546e-05,
"loss": 2.1994,
"step": 804
},
{
"epoch": 4.972136222910216,
"grad_norm": 3.9205013178737276,
"learning_rate": 6.573755154737651e-05,
"loss": 2.1986,
"step": 805
},
{
"epoch": 4.978328173374613,
"grad_norm": 3.965357202346724,
"learning_rate": 6.56579692010634e-05,
"loss": 2.0587,
"step": 806
},
{
"epoch": 4.984520123839009,
"grad_norm": 4.5179420881623695,
"learning_rate": 6.557834285037041e-05,
"loss": 2.1724,
"step": 807
},
{
"epoch": 4.9907120743034055,
"grad_norm": 4.482924767634662,
"learning_rate": 6.549867271907553e-05,
"loss": 2.2078,
"step": 808
},
{
"epoch": 4.996904024767802,
"grad_norm": 4.6896685019691,
"learning_rate": 6.541895903107969e-05,
"loss": 1.9702,
"step": 809
},
{
"epoch": 5.0,
"grad_norm": 4.6896685019691,
"learning_rate": 6.533920201040632e-05,
"loss": 1.0959,
"step": 810
},
{
"epoch": 5.006191950464396,
"grad_norm": 5.097135707519353,
"learning_rate": 6.525940188120059e-05,
"loss": 1.8363,
"step": 811
},
{
"epoch": 5.012383900928793,
"grad_norm": 4.297658356502665,
"learning_rate": 6.51795588677288e-05,
"loss": 1.7269,
"step": 812
},
{
"epoch": 5.018575851393189,
"grad_norm": 4.170938719999973,
"learning_rate": 6.509967319437781e-05,
"loss": 1.6751,
"step": 813
},
{
"epoch": 5.024767801857585,
"grad_norm": 5.0993641270081715,
"learning_rate": 6.501974508565437e-05,
"loss": 1.7688,
"step": 814
},
{
"epoch": 5.030959752321982,
"grad_norm": 6.094023609301853,
"learning_rate": 6.493977476618445e-05,
"loss": 1.7442,
"step": 815
},
{
"epoch": 5.037151702786378,
"grad_norm": 7.029385168513187,
"learning_rate": 6.485976246071269e-05,
"loss": 1.7269,
"step": 816
},
{
"epoch": 5.043343653250774,
"grad_norm": 6.469279878438362,
"learning_rate": 6.477970839410166e-05,
"loss": 1.7327,
"step": 817
},
{
"epoch": 5.04953560371517,
"grad_norm": 7.090204265091153,
"learning_rate": 6.469961279133138e-05,
"loss": 1.6868,
"step": 818
},
{
"epoch": 5.055727554179566,
"grad_norm": 6.448327559102867,
"learning_rate": 6.461947587749855e-05,
"loss": 1.8091,
"step": 819
},
{
"epoch": 5.061919504643963,
"grad_norm": 6.285497602063599,
"learning_rate": 6.453929787781595e-05,
"loss": 1.6628,
"step": 820
},
{
"epoch": 5.068111455108359,
"grad_norm": 5.816454250548965,
"learning_rate": 6.445907901761189e-05,
"loss": 1.6682,
"step": 821
},
{
"epoch": 5.074303405572755,
"grad_norm": 6.3789917153433855,
"learning_rate": 6.437881952232947e-05,
"loss": 1.6749,
"step": 822
},
{
"epoch": 5.080495356037152,
"grad_norm": 6.259435870785406,
"learning_rate": 6.429851961752597e-05,
"loss": 1.7692,
"step": 823
},
{
"epoch": 5.086687306501548,
"grad_norm": 7.073683972141054,
"learning_rate": 6.421817952887228e-05,
"loss": 1.5728,
"step": 824
},
{
"epoch": 5.0928792569659445,
"grad_norm": 6.3192650941579975,
"learning_rate": 6.413779948215218e-05,
"loss": 1.6299,
"step": 825
},
{
"epoch": 5.099071207430341,
"grad_norm": 6.069050786842301,
"learning_rate": 6.405737970326179e-05,
"loss": 1.6522,
"step": 826
},
{
"epoch": 5.105263157894737,
"grad_norm": 5.602073764709582,
"learning_rate": 6.397692041820885e-05,
"loss": 1.6706,
"step": 827
},
{
"epoch": 5.1114551083591335,
"grad_norm": 5.638325928752947,
"learning_rate": 6.389642185311215e-05,
"loss": 1.7242,
"step": 828
},
{
"epoch": 5.117647058823529,
"grad_norm": 6.088790913801255,
"learning_rate": 6.381588423420085e-05,
"loss": 1.7472,
"step": 829
},
{
"epoch": 5.123839009287925,
"grad_norm": 5.959373545378721,
"learning_rate": 6.373530778781391e-05,
"loss": 1.7313,
"step": 830
},
{
"epoch": 5.130030959752322,
"grad_norm": 6.005814498691885,
"learning_rate": 6.365469274039936e-05,
"loss": 1.7649,
"step": 831
},
{
"epoch": 5.136222910216718,
"grad_norm": 6.080930391730417,
"learning_rate": 6.357403931851371e-05,
"loss": 1.6265,
"step": 832
},
{
"epoch": 5.142414860681114,
"grad_norm": 6.5286056711387355,
"learning_rate": 6.349334774882137e-05,
"loss": 1.7628,
"step": 833
},
{
"epoch": 5.148606811145511,
"grad_norm": 6.257759792582404,
"learning_rate": 6.341261825809389e-05,
"loss": 1.8872,
"step": 834
},
{
"epoch": 5.154798761609907,
"grad_norm": 6.206382485711229,
"learning_rate": 6.333185107320945e-05,
"loss": 1.7887,
"step": 835
},
{
"epoch": 5.1609907120743035,
"grad_norm": 6.443322713367943,
"learning_rate": 6.325104642115214e-05,
"loss": 1.8509,
"step": 836
},
{
"epoch": 5.1671826625387,
"grad_norm": 6.384300347287064,
"learning_rate": 6.317020452901133e-05,
"loss": 1.8043,
"step": 837
},
{
"epoch": 5.173374613003096,
"grad_norm": 5.890101319337755,
"learning_rate": 6.308932562398109e-05,
"loss": 1.8201,
"step": 838
},
{
"epoch": 5.179566563467493,
"grad_norm": 5.676545837020276,
"learning_rate": 6.300840993335945e-05,
"loss": 1.6744,
"step": 839
},
{
"epoch": 5.185758513931889,
"grad_norm": 5.442639908625545,
"learning_rate": 6.292745768454787e-05,
"loss": 1.6609,
"step": 840
},
{
"epoch": 5.191950464396285,
"grad_norm": 5.735651569209558,
"learning_rate": 6.284646910505054e-05,
"loss": 1.6535,
"step": 841
},
{
"epoch": 5.198142414860681,
"grad_norm": 5.301848332792911,
"learning_rate": 6.276544442247373e-05,
"loss": 1.7173,
"step": 842
},
{
"epoch": 5.204334365325077,
"grad_norm": 5.838168523934679,
"learning_rate": 6.26843838645252e-05,
"loss": 1.7795,
"step": 843
},
{
"epoch": 5.2105263157894735,
"grad_norm": 6.486729476474496,
"learning_rate": 6.260328765901352e-05,
"loss": 1.6847,
"step": 844
},
{
"epoch": 5.21671826625387,
"grad_norm": 6.080973168348242,
"learning_rate": 6.252215603384743e-05,
"loss": 1.8322,
"step": 845
},
{
"epoch": 5.222910216718266,
"grad_norm": 5.553751162760881,
"learning_rate": 6.244098921703524e-05,
"loss": 1.6801,
"step": 846
},
{
"epoch": 5.2291021671826625,
"grad_norm": 7.7187082516070475,
"learning_rate": 6.235978743668415e-05,
"loss": 1.7557,
"step": 847
},
{
"epoch": 5.235294117647059,
"grad_norm": 5.494934240433607,
"learning_rate": 6.227855092099959e-05,
"loss": 1.7204,
"step": 848
},
{
"epoch": 5.241486068111455,
"grad_norm": 5.400078857427008,
"learning_rate": 6.219727989828466e-05,
"loss": 1.8781,
"step": 849
},
{
"epoch": 5.247678018575852,
"grad_norm": 6.282165666257248,
"learning_rate": 6.211597459693939e-05,
"loss": 1.8185,
"step": 850
},
{
"epoch": 5.253869969040248,
"grad_norm": 6.206071203959746,
"learning_rate": 6.203463524546017e-05,
"loss": 1.8299,
"step": 851
},
{
"epoch": 5.260061919504644,
"grad_norm": 6.929875164782504,
"learning_rate": 6.19532620724391e-05,
"loss": 1.8488,
"step": 852
},
{
"epoch": 5.266253869969041,
"grad_norm": 6.94778058551181,
"learning_rate": 6.187185530656328e-05,
"loss": 1.6549,
"step": 853
},
{
"epoch": 5.272445820433436,
"grad_norm": 6.168749749728858,
"learning_rate": 6.179041517661424e-05,
"loss": 1.8255,
"step": 854
},
{
"epoch": 5.2786377708978325,
"grad_norm": 6.388791515764353,
"learning_rate": 6.170894191146733e-05,
"loss": 1.698,
"step": 855
},
{
"epoch": 5.284829721362229,
"grad_norm": 5.749811058169002,
"learning_rate": 6.162743574009094e-05,
"loss": 1.7379,
"step": 856
},
{
"epoch": 5.291021671826625,
"grad_norm": 5.833626724768377,
"learning_rate": 6.154589689154594e-05,
"loss": 1.7003,
"step": 857
},
{
"epoch": 5.2972136222910216,
"grad_norm": 5.5678551392379845,
"learning_rate": 6.146432559498513e-05,
"loss": 1.8987,
"step": 858
},
{
"epoch": 5.303405572755418,
"grad_norm": 5.858138841566609,
"learning_rate": 6.138272207965238e-05,
"loss": 1.8512,
"step": 859
},
{
"epoch": 5.309597523219814,
"grad_norm": 5.673537543847247,
"learning_rate": 6.130108657488219e-05,
"loss": 1.7378,
"step": 860
},
{
"epoch": 5.315789473684211,
"grad_norm": 6.096191086369572,
"learning_rate": 6.121941931009894e-05,
"loss": 1.7862,
"step": 861
},
{
"epoch": 5.321981424148607,
"grad_norm": 6.86038785401145,
"learning_rate": 6.113772051481622e-05,
"loss": 1.6807,
"step": 862
},
{
"epoch": 5.328173374613003,
"grad_norm": 5.864326184212158,
"learning_rate": 6.105599041863631e-05,
"loss": 1.7716,
"step": 863
},
{
"epoch": 5.3343653250774,
"grad_norm": 6.269969066027309,
"learning_rate": 6.09742292512494e-05,
"loss": 1.8457,
"step": 864
},
{
"epoch": 5.340557275541796,
"grad_norm": 6.933266717469394,
"learning_rate": 6.0892437242433035e-05,
"loss": 1.6483,
"step": 865
},
{
"epoch": 5.346749226006192,
"grad_norm": 6.0658603077371405,
"learning_rate": 6.0810614622051396e-05,
"loss": 1.6833,
"step": 866
},
{
"epoch": 5.352941176470588,
"grad_norm": 6.2928527950892805,
"learning_rate": 6.072876162005474e-05,
"loss": 1.7531,
"step": 867
},
{
"epoch": 5.359133126934984,
"grad_norm": 5.7918387029923215,
"learning_rate": 6.064687846647864e-05,
"loss": 1.7937,
"step": 868
},
{
"epoch": 5.365325077399381,
"grad_norm": 6.581852936909009,
"learning_rate": 6.056496539144351e-05,
"loss": 1.6215,
"step": 869
},
{
"epoch": 5.371517027863777,
"grad_norm": 5.973762507331246,
"learning_rate": 6.0483022625153755e-05,
"loss": 1.856,
"step": 870
},
{
"epoch": 5.377708978328173,
"grad_norm": 5.568545853686291,
"learning_rate": 6.040105039789726e-05,
"loss": 1.8014,
"step": 871
},
{
"epoch": 5.38390092879257,
"grad_norm": 5.262578381419057,
"learning_rate": 6.031904894004471e-05,
"loss": 1.729,
"step": 872
},
{
"epoch": 5.390092879256966,
"grad_norm": 6.009994797740327,
"learning_rate": 6.023701848204893e-05,
"loss": 1.7577,
"step": 873
},
{
"epoch": 5.396284829721362,
"grad_norm": 6.362254949364708,
"learning_rate": 6.015495925444429e-05,
"loss": 1.7988,
"step": 874
},
{
"epoch": 5.402476780185759,
"grad_norm": 5.5039823944568775,
"learning_rate": 6.007287148784591e-05,
"loss": 1.8077,
"step": 875
},
{
"epoch": 5.408668730650155,
"grad_norm": 5.968607801035917,
"learning_rate": 5.999075541294921e-05,
"loss": 1.8124,
"step": 876
},
{
"epoch": 5.414860681114551,
"grad_norm": 6.214379080156402,
"learning_rate": 5.9908611260529135e-05,
"loss": 1.8006,
"step": 877
},
{
"epoch": 5.421052631578947,
"grad_norm": 6.204207450749872,
"learning_rate": 5.982643926143954e-05,
"loss": 1.7918,
"step": 878
},
{
"epoch": 5.427244582043343,
"grad_norm": 5.6408213185164895,
"learning_rate": 5.974423964661249e-05,
"loss": 1.8952,
"step": 879
},
{
"epoch": 5.43343653250774,
"grad_norm": 6.347907311283864,
"learning_rate": 5.966201264705778e-05,
"loss": 1.749,
"step": 880
},
{
"epoch": 5.439628482972136,
"grad_norm": 5.809574613917347,
"learning_rate": 5.957975849386202e-05,
"loss": 1.8773,
"step": 881
},
{
"epoch": 5.445820433436532,
"grad_norm": 5.632385846674114,
"learning_rate": 5.949747741818824e-05,
"loss": 1.7677,
"step": 882
},
{
"epoch": 5.452012383900929,
"grad_norm": 5.440696939412205,
"learning_rate": 5.941516965127509e-05,
"loss": 1.8766,
"step": 883
},
{
"epoch": 5.458204334365325,
"grad_norm": 6.022036249548154,
"learning_rate": 5.933283542443622e-05,
"loss": 1.7556,
"step": 884
},
{
"epoch": 5.464396284829721,
"grad_norm": 5.189714160763887,
"learning_rate": 5.925047496905968e-05,
"loss": 1.7301,
"step": 885
},
{
"epoch": 5.470588235294118,
"grad_norm": 6.485624186917087,
"learning_rate": 5.916808851660718e-05,
"loss": 1.9348,
"step": 886
},
{
"epoch": 5.476780185758514,
"grad_norm": 6.217627206801023,
"learning_rate": 5.9085676298613534e-05,
"loss": 1.7815,
"step": 887
},
{
"epoch": 5.4829721362229105,
"grad_norm": 6.97023254823353,
"learning_rate": 5.900323854668597e-05,
"loss": 1.8831,
"step": 888
},
{
"epoch": 5.489164086687307,
"grad_norm": 5.216736331651397,
"learning_rate": 5.892077549250341e-05,
"loss": 1.7848,
"step": 889
},
{
"epoch": 5.495356037151703,
"grad_norm": 5.512287508329415,
"learning_rate": 5.8838287367815966e-05,
"loss": 1.8791,
"step": 890
},
{
"epoch": 5.5015479876160995,
"grad_norm": 5.711267110515132,
"learning_rate": 5.875577440444418e-05,
"loss": 1.7908,
"step": 891
},
{
"epoch": 5.507739938080495,
"grad_norm": 5.433396790662708,
"learning_rate": 5.867323683427836e-05,
"loss": 1.8233,
"step": 892
},
{
"epoch": 5.513931888544891,
"grad_norm": 6.448127794945795,
"learning_rate": 5.8590674889277994e-05,
"loss": 1.7747,
"step": 893
},
{
"epoch": 5.520123839009288,
"grad_norm": 5.7086178958780565,
"learning_rate": 5.85080888014711e-05,
"loss": 1.8987,
"step": 894
},
{
"epoch": 5.526315789473684,
"grad_norm": 5.583443142146346,
"learning_rate": 5.842547880295353e-05,
"loss": 1.8491,
"step": 895
},
{
"epoch": 5.53250773993808,
"grad_norm": 6.079781416829725,
"learning_rate": 5.834284512588831e-05,
"loss": 1.8515,
"step": 896
},
{
"epoch": 5.538699690402477,
"grad_norm": 5.6372864199748385,
"learning_rate": 5.826018800250503e-05,
"loss": 1.8367,
"step": 897
},
{
"epoch": 5.544891640866873,
"grad_norm": 5.515529122384791,
"learning_rate": 5.817750766509915e-05,
"loss": 1.6847,
"step": 898
},
{
"epoch": 5.5510835913312695,
"grad_norm": 5.334874432827247,
"learning_rate": 5.809480434603143e-05,
"loss": 1.7797,
"step": 899
},
{
"epoch": 5.557275541795666,
"grad_norm": 5.80140570708637,
"learning_rate": 5.801207827772714e-05,
"loss": 1.8654,
"step": 900
},
{
"epoch": 5.563467492260062,
"grad_norm": 6.321831941627306,
"learning_rate": 5.7929329692675525e-05,
"loss": 1.7084,
"step": 901
},
{
"epoch": 5.569659442724459,
"grad_norm": 5.412047294377396,
"learning_rate": 5.784655882342912e-05,
"loss": 1.8567,
"step": 902
},
{
"epoch": 5.575851393188854,
"grad_norm": 5.566852785419783,
"learning_rate": 5.776376590260306e-05,
"loss": 1.8087,
"step": 903
},
{
"epoch": 5.58204334365325,
"grad_norm": 5.652096604313271,
"learning_rate": 5.768095116287444e-05,
"loss": 1.7999,
"step": 904
},
{
"epoch": 5.588235294117647,
"grad_norm": 5.975655299586698,
"learning_rate": 5.7598114836981734e-05,
"loss": 1.8325,
"step": 905
},
{
"epoch": 5.594427244582043,
"grad_norm": 5.940380162039415,
"learning_rate": 5.751525715772401e-05,
"loss": 1.8353,
"step": 906
},
{
"epoch": 5.6006191950464395,
"grad_norm": 5.426752630449644,
"learning_rate": 5.7432378357960415e-05,
"loss": 1.7968,
"step": 907
},
{
"epoch": 5.606811145510836,
"grad_norm": 5.3791925890949255,
"learning_rate": 5.734947867060938e-05,
"loss": 1.7644,
"step": 908
},
{
"epoch": 5.613003095975232,
"grad_norm": 5.3770146741879685,
"learning_rate": 5.726655832864809e-05,
"loss": 1.821,
"step": 909
},
{
"epoch": 5.6191950464396285,
"grad_norm": 5.7576425510460405,
"learning_rate": 5.718361756511177e-05,
"loss": 1.8349,
"step": 910
},
{
"epoch": 5.625386996904025,
"grad_norm": 5.545319099253952,
"learning_rate": 5.7100656613093005e-05,
"loss": 1.8334,
"step": 911
},
{
"epoch": 5.631578947368421,
"grad_norm": 5.621097957860418,
"learning_rate": 5.7017675705741156e-05,
"loss": 1.7505,
"step": 912
},
{
"epoch": 5.637770897832818,
"grad_norm": 5.496504675395947,
"learning_rate": 5.693467507626164e-05,
"loss": 1.7661,
"step": 913
},
{
"epoch": 5.643962848297214,
"grad_norm": 5.440416868123972,
"learning_rate": 5.685165495791534e-05,
"loss": 1.761,
"step": 914
},
{
"epoch": 5.65015479876161,
"grad_norm": 5.587396623055035,
"learning_rate": 5.6768615584017804e-05,
"loss": 1.813,
"step": 915
},
{
"epoch": 5.656346749226007,
"grad_norm": 5.6813322643015125,
"learning_rate": 5.6685557187938844e-05,
"loss": 1.9138,
"step": 916
},
{
"epoch": 5.662538699690402,
"grad_norm": 5.724673293483267,
"learning_rate": 5.660248000310162e-05,
"loss": 1.8429,
"step": 917
},
{
"epoch": 5.6687306501547985,
"grad_norm": 5.7928122967608715,
"learning_rate": 5.6519384262982144e-05,
"loss": 1.8223,
"step": 918
},
{
"epoch": 5.674922600619195,
"grad_norm": 5.425499489197008,
"learning_rate": 5.643627020110855e-05,
"loss": 1.8533,
"step": 919
},
{
"epoch": 5.681114551083591,
"grad_norm": 5.36215523508532,
"learning_rate": 5.635313805106047e-05,
"loss": 1.6969,
"step": 920
},
{
"epoch": 5.687306501547988,
"grad_norm": 4.999203099014986,
"learning_rate": 5.626998804646841e-05,
"loss": 1.8429,
"step": 921
},
{
"epoch": 5.693498452012384,
"grad_norm": 5.928680338381235,
"learning_rate": 5.618682042101297e-05,
"loss": 1.8639,
"step": 922
},
{
"epoch": 5.69969040247678,
"grad_norm": 5.697996175032324,
"learning_rate": 5.610363540842435e-05,
"loss": 1.7222,
"step": 923
},
{
"epoch": 5.705882352941177,
"grad_norm": 5.580876924156143,
"learning_rate": 5.602043324248157e-05,
"loss": 1.8367,
"step": 924
},
{
"epoch": 5.712074303405573,
"grad_norm": 5.563961054575789,
"learning_rate": 5.5937214157011884e-05,
"loss": 1.7032,
"step": 925
},
{
"epoch": 5.718266253869969,
"grad_norm": 5.878100256797036,
"learning_rate": 5.585397838589005e-05,
"loss": 1.805,
"step": 926
},
{
"epoch": 5.724458204334366,
"grad_norm": 5.7076843044116075,
"learning_rate": 5.577072616303779e-05,
"loss": 1.8569,
"step": 927
},
{
"epoch": 5.730650154798761,
"grad_norm": 6.05372952918896,
"learning_rate": 5.5687457722423e-05,
"loss": 1.896,
"step": 928
},
{
"epoch": 5.7368421052631575,
"grad_norm": 5.16364163182538,
"learning_rate": 5.5604173298059156e-05,
"loss": 1.6782,
"step": 929
},
{
"epoch": 5.743034055727554,
"grad_norm": 5.6002530820827525,
"learning_rate": 5.55208731240047e-05,
"loss": 1.7387,
"step": 930
},
{
"epoch": 5.74922600619195,
"grad_norm": 5.471173638395719,
"learning_rate": 5.5437557434362305e-05,
"loss": 1.6704,
"step": 931
},
{
"epoch": 5.755417956656347,
"grad_norm": 5.614838368063709,
"learning_rate": 5.535422646327826e-05,
"loss": 1.9572,
"step": 932
},
{
"epoch": 5.761609907120743,
"grad_norm": 6.03455846560755,
"learning_rate": 5.5270880444941764e-05,
"loss": 1.9031,
"step": 933
},
{
"epoch": 5.767801857585139,
"grad_norm": 6.661401789268076,
"learning_rate": 5.518751961358436e-05,
"loss": 1.8565,
"step": 934
},
{
"epoch": 5.773993808049536,
"grad_norm": 5.647161883767869,
"learning_rate": 5.510414420347918e-05,
"loss": 1.8405,
"step": 935
},
{
"epoch": 5.780185758513932,
"grad_norm": 5.334550187571599,
"learning_rate": 5.502075444894035e-05,
"loss": 1.898,
"step": 936
},
{
"epoch": 5.786377708978328,
"grad_norm": 5.0499248707961675,
"learning_rate": 5.493735058432227e-05,
"loss": 1.872,
"step": 937
},
{
"epoch": 5.792569659442725,
"grad_norm": 5.50650425050073,
"learning_rate": 5.485393284401905e-05,
"loss": 1.7519,
"step": 938
},
{
"epoch": 5.798761609907121,
"grad_norm": 5.256311428845345,
"learning_rate": 5.477050146246378e-05,
"loss": 1.703,
"step": 939
},
{
"epoch": 5.804953560371517,
"grad_norm": 5.408181017988561,
"learning_rate": 5.468705667412785e-05,
"loss": 1.8261,
"step": 940
},
{
"epoch": 5.811145510835914,
"grad_norm": 5.484560920918981,
"learning_rate": 5.4603598713520356e-05,
"loss": 1.8205,
"step": 941
},
{
"epoch": 5.817337461300309,
"grad_norm": 5.728999670930413,
"learning_rate": 5.452012781518743e-05,
"loss": 1.6577,
"step": 942
},
{
"epoch": 5.823529411764706,
"grad_norm": 5.502548441271816,
"learning_rate": 5.443664421371153e-05,
"loss": 1.6221,
"step": 943
},
{
"epoch": 5.829721362229102,
"grad_norm": 6.35610081466655,
"learning_rate": 5.435314814371082e-05,
"loss": 1.7112,
"step": 944
},
{
"epoch": 5.835913312693498,
"grad_norm": 5.6172244449375075,
"learning_rate": 5.426963983983853e-05,
"loss": 1.7829,
"step": 945
},
{
"epoch": 5.842105263157895,
"grad_norm": 5.492954166804681,
"learning_rate": 5.4186119536782246e-05,
"loss": 1.7652,
"step": 946
},
{
"epoch": 5.848297213622291,
"grad_norm": 5.7804318465473425,
"learning_rate": 5.410258746926328e-05,
"loss": 1.7701,
"step": 947
},
{
"epoch": 5.854489164086687,
"grad_norm": 6.372706653206452,
"learning_rate": 5.4019043872036015e-05,
"loss": 1.6976,
"step": 948
},
{
"epoch": 5.860681114551084,
"grad_norm": 5.19023867382763,
"learning_rate": 5.393548897988724e-05,
"loss": 1.7425,
"step": 949
},
{
"epoch": 5.86687306501548,
"grad_norm": 5.313027581311771,
"learning_rate": 5.3851923027635475e-05,
"loss": 1.7379,
"step": 950
},
{
"epoch": 5.8730650154798765,
"grad_norm": 5.346043529771426,
"learning_rate": 5.376834625013031e-05,
"loss": 1.7895,
"step": 951
},
{
"epoch": 5.879256965944273,
"grad_norm": 5.6516277074371235,
"learning_rate": 5.3684758882251794e-05,
"loss": 1.6915,
"step": 952
},
{
"epoch": 5.885448916408668,
"grad_norm": 5.60282925527119,
"learning_rate": 5.360116115890972e-05,
"loss": 1.5807,
"step": 953
},
{
"epoch": 5.891640866873065,
"grad_norm": 5.756675233255651,
"learning_rate": 5.351755331504297e-05,
"loss": 1.8783,
"step": 954
},
{
"epoch": 5.897832817337461,
"grad_norm": 6.398985132652635,
"learning_rate": 5.343393558561888e-05,
"loss": 1.8418,
"step": 955
},
{
"epoch": 5.904024767801857,
"grad_norm": 5.714793081519792,
"learning_rate": 5.3350308205632574e-05,
"loss": 1.6848,
"step": 956
},
{
"epoch": 5.910216718266254,
"grad_norm": 5.732024375784504,
"learning_rate": 5.3266671410106306e-05,
"loss": 1.7741,
"step": 957
},
{
"epoch": 5.91640866873065,
"grad_norm": 5.777349638869013,
"learning_rate": 5.318302543408875e-05,
"loss": 1.7644,
"step": 958
},
{
"epoch": 5.922600619195046,
"grad_norm": 5.976598354398558,
"learning_rate": 5.3099370512654426e-05,
"loss": 1.7296,
"step": 959
},
{
"epoch": 5.928792569659443,
"grad_norm": 5.404269095636371,
"learning_rate": 5.3015706880902974e-05,
"loss": 1.6866,
"step": 960
},
{
"epoch": 5.934984520123839,
"grad_norm": 5.527542596369583,
"learning_rate": 5.293203477395851e-05,
"loss": 1.7146,
"step": 961
},
{
"epoch": 5.9411764705882355,
"grad_norm": 5.078043839033948,
"learning_rate": 5.284835442696895e-05,
"loss": 1.8524,
"step": 962
},
{
"epoch": 5.947368421052632,
"grad_norm": 5.772706720501868,
"learning_rate": 5.276466607510544e-05,
"loss": 1.7661,
"step": 963
},
{
"epoch": 5.953560371517028,
"grad_norm": 5.520796373876517,
"learning_rate": 5.2680969953561545e-05,
"loss": 1.882,
"step": 964
},
{
"epoch": 5.959752321981425,
"grad_norm": 5.827251604758201,
"learning_rate": 5.259726629755267e-05,
"loss": 1.8778,
"step": 965
},
{
"epoch": 5.965944272445821,
"grad_norm": 6.207373379919172,
"learning_rate": 5.251355534231546e-05,
"loss": 1.8627,
"step": 966
},
{
"epoch": 5.972136222910216,
"grad_norm": 5.123600713644226,
"learning_rate": 5.2429837323107e-05,
"loss": 1.7472,
"step": 967
},
{
"epoch": 5.978328173374613,
"grad_norm": 5.101450048523518,
"learning_rate": 5.234611247520428e-05,
"loss": 1.7643,
"step": 968
},
{
"epoch": 5.984520123839009,
"grad_norm": 5.408204446605995,
"learning_rate": 5.2262381033903426e-05,
"loss": 1.7875,
"step": 969
},
{
"epoch": 5.9907120743034055,
"grad_norm": 5.379723976429878,
"learning_rate": 5.2178643234519164e-05,
"loss": 1.7445,
"step": 970
},
{
"epoch": 5.996904024767802,
"grad_norm": 5.838975887259199,
"learning_rate": 5.209489931238405e-05,
"loss": 1.7112,
"step": 971
},
{
"epoch": 6.0,
"grad_norm": 5.840227146356309,
"learning_rate": 5.201114950284782e-05,
"loss": 0.8978,
"step": 972
},
{
"epoch": 6.006191950464396,
"grad_norm": 4.275597834318695,
"learning_rate": 5.192739404127679e-05,
"loss": 1.3108,
"step": 973
},
{
"epoch": 6.012383900928793,
"grad_norm": 5.030728682297231,
"learning_rate": 5.1843633163053175e-05,
"loss": 1.2031,
"step": 974
},
{
"epoch": 6.018575851393189,
"grad_norm": 5.8141889393279484,
"learning_rate": 5.175986710357439e-05,
"loss": 1.169,
"step": 975
},
{
"epoch": 6.024767801857585,
"grad_norm": 7.558224479046815,
"learning_rate": 5.167609609825238e-05,
"loss": 1.219,
"step": 976
},
{
"epoch": 6.030959752321982,
"grad_norm": 8.484873862857635,
"learning_rate": 5.159232038251305e-05,
"loss": 1.2784,
"step": 977
},
{
"epoch": 6.037151702786378,
"grad_norm": 9.329078391082556,
"learning_rate": 5.1508540191795506e-05,
"loss": 1.2931,
"step": 978
},
{
"epoch": 6.043343653250774,
"grad_norm": 8.234782351921575,
"learning_rate": 5.142475576155146e-05,
"loss": 1.1175,
"step": 979
},
{
"epoch": 6.04953560371517,
"grad_norm": 8.285760286310255,
"learning_rate": 5.1340967327244496e-05,
"loss": 1.2494,
"step": 980
},
{
"epoch": 6.055727554179566,
"grad_norm": 7.602885755223493,
"learning_rate": 5.1257175124349464e-05,
"loss": 1.2171,
"step": 981
},
{
"epoch": 6.061919504643963,
"grad_norm": 7.073271538574044,
"learning_rate": 5.117337938835186e-05,
"loss": 1.1765,
"step": 982
},
{
"epoch": 6.068111455108359,
"grad_norm": 7.315973298938372,
"learning_rate": 5.1089580354747026e-05,
"loss": 1.118,
"step": 983
},
{
"epoch": 6.074303405572755,
"grad_norm": 6.601903535655793,
"learning_rate": 5.100577825903958e-05,
"loss": 1.2171,
"step": 984
},
{
"epoch": 6.080495356037152,
"grad_norm": 6.757876749392141,
"learning_rate": 5.092197333674286e-05,
"loss": 1.2503,
"step": 985
},
{
"epoch": 6.086687306501548,
"grad_norm": 6.9466240275860835,
"learning_rate": 5.0838165823377995e-05,
"loss": 1.2236,
"step": 986
},
{
"epoch": 6.0928792569659445,
"grad_norm": 7.381969086666438,
"learning_rate": 5.0754355954473466e-05,
"loss": 1.2083,
"step": 987
},
{
"epoch": 6.099071207430341,
"grad_norm": 7.145098862853887,
"learning_rate": 5.0670543965564386e-05,
"loss": 1.3174,
"step": 988
},
{
"epoch": 6.105263157894737,
"grad_norm": 7.137360767402197,
"learning_rate": 5.0586730092191835e-05,
"loss": 1.1661,
"step": 989
},
{
"epoch": 6.1114551083591335,
"grad_norm": 7.417766045611863,
"learning_rate": 5.0502914569902116e-05,
"loss": 1.1905,
"step": 990
},
{
"epoch": 6.117647058823529,
"grad_norm": 7.26464568460221,
"learning_rate": 5.041909763424625e-05,
"loss": 1.2316,
"step": 991
},
{
"epoch": 6.123839009287925,
"grad_norm": 7.5052794586230345,
"learning_rate": 5.033527952077917e-05,
"loss": 1.1215,
"step": 992
},
{
"epoch": 6.130030959752322,
"grad_norm": 7.1970182467527355,
"learning_rate": 5.025146046505917e-05,
"loss": 1.2079,
"step": 993
},
{
"epoch": 6.136222910216718,
"grad_norm": 7.749893049143311,
"learning_rate": 5.0167640702647155e-05,
"loss": 1.3064,
"step": 994
},
{
"epoch": 6.142414860681114,
"grad_norm": 7.551889488671015,
"learning_rate": 5.0083820469106014e-05,
"loss": 1.2398,
"step": 995
},
{
"epoch": 6.148606811145511,
"grad_norm": 7.588653228547087,
"learning_rate": 5e-05,
"loss": 1.2056,
"step": 996
},
{
"epoch": 6.154798761609907,
"grad_norm": 7.447548089817992,
"learning_rate": 4.991617953089399e-05,
"loss": 1.2482,
"step": 997
},
{
"epoch": 6.1609907120743035,
"grad_norm": 7.159080033768566,
"learning_rate": 4.9832359297352856e-05,
"loss": 1.2093,
"step": 998
},
{
"epoch": 6.1671826625387,
"grad_norm": 7.456941150360937,
"learning_rate": 4.9748539534940825e-05,
"loss": 1.2034,
"step": 999
},
{
"epoch": 6.173374613003096,
"grad_norm": 8.179625978947795,
"learning_rate": 4.966472047922083e-05,
"loss": 1.2213,
"step": 1000
},
{
"epoch": 6.179566563467493,
"grad_norm": 7.691969033733335,
"learning_rate": 4.958090236575377e-05,
"loss": 1.2253,
"step": 1001
},
{
"epoch": 6.185758513931889,
"grad_norm": 7.2575621885339645,
"learning_rate": 4.9497085430097896e-05,
"loss": 1.3645,
"step": 1002
},
{
"epoch": 6.191950464396285,
"grad_norm": 6.507233877034855,
"learning_rate": 4.941326990780819e-05,
"loss": 1.1571,
"step": 1003
},
{
"epoch": 6.198142414860681,
"grad_norm": 6.651433405699857,
"learning_rate": 4.932945603443563e-05,
"loss": 1.1652,
"step": 1004
},
{
"epoch": 6.204334365325077,
"grad_norm": 6.650448093340969,
"learning_rate": 4.9245644045526546e-05,
"loss": 1.2783,
"step": 1005
},
{
"epoch": 6.2105263157894735,
"grad_norm": 7.781500509708706,
"learning_rate": 4.916183417662202e-05,
"loss": 1.3173,
"step": 1006
},
{
"epoch": 6.21671826625387,
"grad_norm": 7.517957799121537,
"learning_rate": 4.907802666325716e-05,
"loss": 1.2165,
"step": 1007
},
{
"epoch": 6.222910216718266,
"grad_norm": 7.157971116077995,
"learning_rate": 4.8994221740960424e-05,
"loss": 1.2808,
"step": 1008
},
{
"epoch": 6.2291021671826625,
"grad_norm": 6.816498614027582,
"learning_rate": 4.8910419645253e-05,
"loss": 1.2127,
"step": 1009
},
{
"epoch": 6.235294117647059,
"grad_norm": 6.787835631470309,
"learning_rate": 4.882662061164814e-05,
"loss": 1.2102,
"step": 1010
},
{
"epoch": 6.241486068111455,
"grad_norm": 7.371585623447389,
"learning_rate": 4.874282487565053e-05,
"loss": 1.1692,
"step": 1011
},
{
"epoch": 6.247678018575852,
"grad_norm": 7.24692426387527,
"learning_rate": 4.8659032672755516e-05,
"loss": 1.1388,
"step": 1012
},
{
"epoch": 6.253869969040248,
"grad_norm": 7.6875630481514206,
"learning_rate": 4.8575244238448546e-05,
"loss": 1.2984,
"step": 1013
},
{
"epoch": 6.260061919504644,
"grad_norm": 8.185112676453237,
"learning_rate": 4.8491459808204506e-05,
"loss": 1.1884,
"step": 1014
},
{
"epoch": 6.266253869969041,
"grad_norm": 7.504289154797226,
"learning_rate": 4.8407679617486974e-05,
"loss": 1.1316,
"step": 1015
},
{
"epoch": 6.272445820433436,
"grad_norm": 7.442593673019155,
"learning_rate": 4.832390390174763e-05,
"loss": 1.2232,
"step": 1016
},
{
"epoch": 6.2786377708978325,
"grad_norm": 7.614862457360967,
"learning_rate": 4.824013289642563e-05,
"loss": 1.31,
"step": 1017
},
{
"epoch": 6.284829721362229,
"grad_norm": 7.024726600452601,
"learning_rate": 4.815636683694683e-05,
"loss": 1.2616,
"step": 1018
},
{
"epoch": 6.291021671826625,
"grad_norm": 7.377696892059301,
"learning_rate": 4.807260595872322e-05,
"loss": 1.24,
"step": 1019
},
{
"epoch": 6.2972136222910216,
"grad_norm": 8.404120290677987,
"learning_rate": 4.79888504971522e-05,
"loss": 1.2933,
"step": 1020
},
{
"epoch": 6.303405572755418,
"grad_norm": 6.962131331816545,
"learning_rate": 4.7905100687615956e-05,
"loss": 1.2178,
"step": 1021
},
{
"epoch": 6.309597523219814,
"grad_norm": 7.4231879389884,
"learning_rate": 4.7821356765480834e-05,
"loss": 1.365,
"step": 1022
},
{
"epoch": 6.315789473684211,
"grad_norm": 7.668830962212028,
"learning_rate": 4.773761896609658e-05,
"loss": 1.1617,
"step": 1023
},
{
"epoch": 6.321981424148607,
"grad_norm": 8.124924759999253,
"learning_rate": 4.7653887524795735e-05,
"loss": 1.288,
"step": 1024
},
{
"epoch": 6.328173374613003,
"grad_norm": 6.8212071234113365,
"learning_rate": 4.7570162676893014e-05,
"loss": 1.2787,
"step": 1025
},
{
"epoch": 6.3343653250774,
"grad_norm": 7.628456619758793,
"learning_rate": 4.748644465768457e-05,
"loss": 1.2784,
"step": 1026
},
{
"epoch": 6.340557275541796,
"grad_norm": 6.808373085825395,
"learning_rate": 4.740273370244734e-05,
"loss": 1.165,
"step": 1027
},
{
"epoch": 6.346749226006192,
"grad_norm": 7.305123080991359,
"learning_rate": 4.7319030046438474e-05,
"loss": 1.3038,
"step": 1028
},
{
"epoch": 6.352941176470588,
"grad_norm": 7.031017640077646,
"learning_rate": 4.723533392489457e-05,
"loss": 1.2891,
"step": 1029
},
{
"epoch": 6.359133126934984,
"grad_norm": 7.946312792286679,
"learning_rate": 4.7151645573031064e-05,
"loss": 1.2275,
"step": 1030
},
{
"epoch": 6.365325077399381,
"grad_norm": 7.232948196291571,
"learning_rate": 4.706796522604152e-05,
"loss": 1.2756,
"step": 1031
},
{
"epoch": 6.371517027863777,
"grad_norm": 8.169258262407036,
"learning_rate": 4.698429311909705e-05,
"loss": 1.2046,
"step": 1032
},
{
"epoch": 6.377708978328173,
"grad_norm": 7.139112933535375,
"learning_rate": 4.690062948734558e-05,
"loss": 1.2697,
"step": 1033
},
{
"epoch": 6.38390092879257,
"grad_norm": 7.031067573669555,
"learning_rate": 4.681697456591126e-05,
"loss": 1.3274,
"step": 1034
},
{
"epoch": 6.390092879256966,
"grad_norm": 7.388094833181507,
"learning_rate": 4.673332858989371e-05,
"loss": 1.3151,
"step": 1035
},
{
"epoch": 6.396284829721362,
"grad_norm": 7.525891020121259,
"learning_rate": 4.664969179436744e-05,
"loss": 1.3229,
"step": 1036
},
{
"epoch": 6.402476780185759,
"grad_norm": 7.036939378953139,
"learning_rate": 4.656606441438113e-05,
"loss": 1.3305,
"step": 1037
},
{
"epoch": 6.408668730650155,
"grad_norm": 6.516201091133355,
"learning_rate": 4.648244668495704e-05,
"loss": 1.2263,
"step": 1038
},
{
"epoch": 6.414860681114551,
"grad_norm": 7.819583293034257,
"learning_rate": 4.6398838841090284e-05,
"loss": 1.3778,
"step": 1039
},
{
"epoch": 6.421052631578947,
"grad_norm": 7.405554618748319,
"learning_rate": 4.631524111774822e-05,
"loss": 1.2554,
"step": 1040
},
{
"epoch": 6.427244582043343,
"grad_norm": 7.485395073347659,
"learning_rate": 4.623165374986971e-05,
"loss": 1.3306,
"step": 1041
},
{
"epoch": 6.43343653250774,
"grad_norm": 7.3281945090055824,
"learning_rate": 4.6148076972364544e-05,
"loss": 1.2903,
"step": 1042
},
{
"epoch": 6.439628482972136,
"grad_norm": 6.89408951561583,
"learning_rate": 4.606451102011278e-05,
"loss": 1.218,
"step": 1043
},
{
"epoch": 6.445820433436532,
"grad_norm": 7.208808297092486,
"learning_rate": 4.598095612796398e-05,
"loss": 1.2513,
"step": 1044
},
{
"epoch": 6.452012383900929,
"grad_norm": 8.362888875092093,
"learning_rate": 4.5897412530736735e-05,
"loss": 1.28,
"step": 1045
},
{
"epoch": 6.458204334365325,
"grad_norm": 8.221896268659757,
"learning_rate": 4.5813880463217766e-05,
"loss": 1.3607,
"step": 1046
},
{
"epoch": 6.464396284829721,
"grad_norm": 6.598237194007563,
"learning_rate": 4.573036016016149e-05,
"loss": 1.2459,
"step": 1047
},
{
"epoch": 6.470588235294118,
"grad_norm": 7.17506795330074,
"learning_rate": 4.564685185628919e-05,
"loss": 1.2524,
"step": 1048
},
{
"epoch": 6.476780185758514,
"grad_norm": 7.37054652654002,
"learning_rate": 4.556335578628849e-05,
"loss": 1.2609,
"step": 1049
},
{
"epoch": 6.4829721362229105,
"grad_norm": 7.444318533212484,
"learning_rate": 4.5479872184812575e-05,
"loss": 1.1776,
"step": 1050
},
{
"epoch": 6.489164086687307,
"grad_norm": 6.7548786489662,
"learning_rate": 4.539640128647965e-05,
"loss": 1.1816,
"step": 1051
},
{
"epoch": 6.495356037151703,
"grad_norm": 7.502330023138323,
"learning_rate": 4.531294332587216e-05,
"loss": 1.2786,
"step": 1052
},
{
"epoch": 6.5015479876160995,
"grad_norm": 7.6873595870822475,
"learning_rate": 4.5229498537536235e-05,
"loss": 1.1847,
"step": 1053
},
{
"epoch": 6.507739938080495,
"grad_norm": 7.08085653318228,
"learning_rate": 4.514606715598096e-05,
"loss": 1.2494,
"step": 1054
},
{
"epoch": 6.513931888544891,
"grad_norm": 7.774057962263595,
"learning_rate": 4.506264941567774e-05,
"loss": 1.1855,
"step": 1055
},
{
"epoch": 6.520123839009288,
"grad_norm": 7.957261147038559,
"learning_rate": 4.497924555105966e-05,
"loss": 1.2751,
"step": 1056
},
{
"epoch": 6.526315789473684,
"grad_norm": 7.307938939131949,
"learning_rate": 4.489585579652083e-05,
"loss": 1.3571,
"step": 1057
},
{
"epoch": 6.53250773993808,
"grad_norm": 7.9128153067396605,
"learning_rate": 4.4812480386415655e-05,
"loss": 1.3486,
"step": 1058
},
{
"epoch": 6.538699690402477,
"grad_norm": 7.585860667174547,
"learning_rate": 4.472911955505824e-05,
"loss": 1.3405,
"step": 1059
},
{
"epoch": 6.544891640866873,
"grad_norm": 7.053910641821553,
"learning_rate": 4.464577353672175e-05,
"loss": 1.2813,
"step": 1060
},
{
"epoch": 6.5510835913312695,
"grad_norm": 7.0175847912120535,
"learning_rate": 4.456244256563769e-05,
"loss": 1.2993,
"step": 1061
},
{
"epoch": 6.557275541795666,
"grad_norm": 8.261893892570697,
"learning_rate": 4.4479126875995304e-05,
"loss": 1.3186,
"step": 1062
},
{
"epoch": 6.563467492260062,
"grad_norm": 7.377863177529412,
"learning_rate": 4.439582670194085e-05,
"loss": 1.2835,
"step": 1063
},
{
"epoch": 6.569659442724459,
"grad_norm": 7.988132361082391,
"learning_rate": 4.431254227757703e-05,
"loss": 1.2427,
"step": 1064
},
{
"epoch": 6.575851393188854,
"grad_norm": 7.294581625440724,
"learning_rate": 4.422927383696224e-05,
"loss": 1.4212,
"step": 1065
},
{
"epoch": 6.58204334365325,
"grad_norm": 7.293717549302451,
"learning_rate": 4.414602161410996e-05,
"loss": 1.3549,
"step": 1066
},
{
"epoch": 6.588235294117647,
"grad_norm": 7.058630306102499,
"learning_rate": 4.406278584298813e-05,
"loss": 1.2859,
"step": 1067
},
{
"epoch": 6.594427244582043,
"grad_norm": 6.921408254017408,
"learning_rate": 4.397956675751844e-05,
"loss": 1.2809,
"step": 1068
},
{
"epoch": 6.6006191950464395,
"grad_norm": 7.558955213024363,
"learning_rate": 4.389636459157567e-05,
"loss": 1.3715,
"step": 1069
},
{
"epoch": 6.606811145510836,
"grad_norm": 6.88911221915387,
"learning_rate": 4.381317957898704e-05,
"loss": 1.3168,
"step": 1070
},
{
"epoch": 6.613003095975232,
"grad_norm": 7.249344304210554,
"learning_rate": 4.373001195353159e-05,
"loss": 1.3486,
"step": 1071
},
{
"epoch": 6.6191950464396285,
"grad_norm": 7.779301524807008,
"learning_rate": 4.364686194893952e-05,
"loss": 1.1833,
"step": 1072
},
{
"epoch": 6.625386996904025,
"grad_norm": 6.863098452906575,
"learning_rate": 4.356372979889146e-05,
"loss": 1.3434,
"step": 1073
},
{
"epoch": 6.631578947368421,
"grad_norm": 7.629387293816119,
"learning_rate": 4.348061573701786e-05,
"loss": 1.3763,
"step": 1074
},
{
"epoch": 6.637770897832818,
"grad_norm": 6.759090261353713,
"learning_rate": 4.339751999689839e-05,
"loss": 1.2835,
"step": 1075
},
{
"epoch": 6.643962848297214,
"grad_norm": 6.950291332023515,
"learning_rate": 4.3314442812061174e-05,
"loss": 1.3158,
"step": 1076
},
{
"epoch": 6.65015479876161,
"grad_norm": 7.394433820704982,
"learning_rate": 4.323138441598219e-05,
"loss": 1.3166,
"step": 1077
},
{
"epoch": 6.656346749226007,
"grad_norm": 7.520288383722824,
"learning_rate": 4.3148345042084674e-05,
"loss": 1.3971,
"step": 1078
},
{
"epoch": 6.662538699690402,
"grad_norm": 7.226902674751723,
"learning_rate": 4.306532492373836e-05,
"loss": 1.3676,
"step": 1079
},
{
"epoch": 6.6687306501547985,
"grad_norm": 7.914934791249491,
"learning_rate": 4.2982324294258855e-05,
"loss": 1.2324,
"step": 1080
},
{
"epoch": 6.674922600619195,
"grad_norm": 7.132623440013545,
"learning_rate": 4.289934338690701e-05,
"loss": 1.3126,
"step": 1081
},
{
"epoch": 6.681114551083591,
"grad_norm": 6.94850316262997,
"learning_rate": 4.281638243488823e-05,
"loss": 1.3197,
"step": 1082
},
{
"epoch": 6.687306501547988,
"grad_norm": 7.615556368279129,
"learning_rate": 4.273344167135191e-05,
"loss": 1.2855,
"step": 1083
},
{
"epoch": 6.693498452012384,
"grad_norm": 8.482830314291196,
"learning_rate": 4.265052132939063e-05,
"loss": 1.2906,
"step": 1084
},
{
"epoch": 6.69969040247678,
"grad_norm": 6.951661079605685,
"learning_rate": 4.2567621642039596e-05,
"loss": 1.3183,
"step": 1085
},
{
"epoch": 6.705882352941177,
"grad_norm": 7.328744620320157,
"learning_rate": 4.2484742842276e-05,
"loss": 1.2574,
"step": 1086
},
{
"epoch": 6.712074303405573,
"grad_norm": 7.433086061619386,
"learning_rate": 4.240188516301829e-05,
"loss": 1.2361,
"step": 1087
},
{
"epoch": 6.718266253869969,
"grad_norm": 7.451348957570564,
"learning_rate": 4.2319048837125566e-05,
"loss": 1.4189,
"step": 1088
},
{
"epoch": 6.724458204334366,
"grad_norm": 7.36655253165013,
"learning_rate": 4.223623409739695e-05,
"loss": 1.2414,
"step": 1089
},
{
"epoch": 6.730650154798761,
"grad_norm": 7.50762675670494,
"learning_rate": 4.215344117657088e-05,
"loss": 1.2911,
"step": 1090
},
{
"epoch": 6.7368421052631575,
"grad_norm": 7.2198068102674515,
"learning_rate": 4.207067030732449e-05,
"loss": 1.2591,
"step": 1091
},
{
"epoch": 6.743034055727554,
"grad_norm": 7.423469615444281,
"learning_rate": 4.198792172227287e-05,
"loss": 1.2815,
"step": 1092
},
{
"epoch": 6.74922600619195,
"grad_norm": 7.8244524756087905,
"learning_rate": 4.1905195653968585e-05,
"loss": 1.2598,
"step": 1093
},
{
"epoch": 6.755417956656347,
"grad_norm": 7.455185432784073,
"learning_rate": 4.182249233490084e-05,
"loss": 1.3286,
"step": 1094
},
{
"epoch": 6.761609907120743,
"grad_norm": 7.35007999134081,
"learning_rate": 4.173981199749498e-05,
"loss": 1.3021,
"step": 1095
},
{
"epoch": 6.767801857585139,
"grad_norm": 6.911478606382307,
"learning_rate": 4.1657154874111695e-05,
"loss": 1.3063,
"step": 1096
},
{
"epoch": 6.773993808049536,
"grad_norm": 7.641719895442311,
"learning_rate": 4.157452119704648e-05,
"loss": 1.2602,
"step": 1097
},
{
"epoch": 6.780185758513932,
"grad_norm": 7.401036899860657,
"learning_rate": 4.149191119852891e-05,
"loss": 1.329,
"step": 1098
},
{
"epoch": 6.786377708978328,
"grad_norm": 7.09677983300196,
"learning_rate": 4.140932511072201e-05,
"loss": 1.321,
"step": 1099
},
{
"epoch": 6.792569659442725,
"grad_norm": 7.277552375894589,
"learning_rate": 4.1326763165721655e-05,
"loss": 1.1842,
"step": 1100
},
{
"epoch": 6.798761609907121,
"grad_norm": 8.032866620643269,
"learning_rate": 4.124422559555584e-05,
"loss": 1.383,
"step": 1101
},
{
"epoch": 6.804953560371517,
"grad_norm": 6.913712713930272,
"learning_rate": 4.1161712632184045e-05,
"loss": 1.2827,
"step": 1102
},
{
"epoch": 6.811145510835914,
"grad_norm": 6.835509230572801,
"learning_rate": 4.10792245074966e-05,
"loss": 1.3363,
"step": 1103
},
{
"epoch": 6.817337461300309,
"grad_norm": 7.879506828917882,
"learning_rate": 4.0996761453314056e-05,
"loss": 1.2357,
"step": 1104
},
{
"epoch": 6.823529411764706,
"grad_norm": 7.850103476202386,
"learning_rate": 4.0914323701386464e-05,
"loss": 1.2596,
"step": 1105
},
{
"epoch": 6.829721362229102,
"grad_norm": 7.300238242699234,
"learning_rate": 4.083191148339283e-05,
"loss": 1.3611,
"step": 1106
},
{
"epoch": 6.835913312693498,
"grad_norm": 7.2927400255517165,
"learning_rate": 4.074952503094033e-05,
"loss": 1.2284,
"step": 1107
},
{
"epoch": 6.842105263157895,
"grad_norm": 6.723280082966505,
"learning_rate": 4.0667164575563784e-05,
"loss": 1.1778,
"step": 1108
},
{
"epoch": 6.848297213622291,
"grad_norm": 6.577268826811694,
"learning_rate": 4.0584830348724935e-05,
"loss": 1.318,
"step": 1109
},
{
"epoch": 6.854489164086687,
"grad_norm": 7.463974601182562,
"learning_rate": 4.050252258181177e-05,
"loss": 1.3231,
"step": 1110
},
{
"epoch": 6.860681114551084,
"grad_norm": 7.4880776364105275,
"learning_rate": 4.042024150613799e-05,
"loss": 1.3418,
"step": 1111
},
{
"epoch": 6.86687306501548,
"grad_norm": 8.416888779323035,
"learning_rate": 4.033798735294224e-05,
"loss": 1.366,
"step": 1112
},
{
"epoch": 6.8730650154798765,
"grad_norm": 7.515989995582284,
"learning_rate": 4.025576035338752e-05,
"loss": 1.3318,
"step": 1113
},
{
"epoch": 6.879256965944273,
"grad_norm": 6.861478583396406,
"learning_rate": 4.017356073856049e-05,
"loss": 1.2534,
"step": 1114
},
{
"epoch": 6.885448916408668,
"grad_norm": 6.68693509128183,
"learning_rate": 4.0091388739470884e-05,
"loss": 1.2903,
"step": 1115
},
{
"epoch": 6.891640866873065,
"grad_norm": 8.488756826558117,
"learning_rate": 4.000924458705079e-05,
"loss": 1.4168,
"step": 1116
},
{
"epoch": 6.897832817337461,
"grad_norm": 7.461099214702575,
"learning_rate": 3.9927128512154103e-05,
"loss": 1.3652,
"step": 1117
},
{
"epoch": 6.904024767801857,
"grad_norm": 7.025634053844339,
"learning_rate": 3.984504074555573e-05,
"loss": 1.2759,
"step": 1118
},
{
"epoch": 6.910216718266254,
"grad_norm": 7.076864815181845,
"learning_rate": 3.976298151795107e-05,
"loss": 1.2637,
"step": 1119
},
{
"epoch": 6.91640866873065,
"grad_norm": 7.262703330555911,
"learning_rate": 3.968095105995531e-05,
"loss": 1.1953,
"step": 1120
},
{
"epoch": 6.922600619195046,
"grad_norm": 7.38167980935885,
"learning_rate": 3.959894960210275e-05,
"loss": 1.3502,
"step": 1121
},
{
"epoch": 6.928792569659443,
"grad_norm": 6.487417226403372,
"learning_rate": 3.951697737484625e-05,
"loss": 1.2807,
"step": 1122
},
{
"epoch": 6.934984520123839,
"grad_norm": 6.621086820516981,
"learning_rate": 3.9435034608556504e-05,
"loss": 1.1824,
"step": 1123
},
{
"epoch": 6.9411764705882355,
"grad_norm": 6.198511111175569,
"learning_rate": 3.935312153352137e-05,
"loss": 1.2067,
"step": 1124
},
{
"epoch": 6.947368421052632,
"grad_norm": 7.52708660819256,
"learning_rate": 3.9271238379945283e-05,
"loss": 1.3398,
"step": 1125
},
{
"epoch": 6.953560371517028,
"grad_norm": 7.372110340772461,
"learning_rate": 3.918938537794862e-05,
"loss": 1.1414,
"step": 1126
},
{
"epoch": 6.959752321981425,
"grad_norm": 7.780518558451204,
"learning_rate": 3.910756275756697e-05,
"loss": 1.2668,
"step": 1127
},
{
"epoch": 6.965944272445821,
"grad_norm": 7.463867146831107,
"learning_rate": 3.902577074875061e-05,
"loss": 1.2988,
"step": 1128
},
{
"epoch": 6.972136222910216,
"grad_norm": 7.644638123297254,
"learning_rate": 3.8944009581363696e-05,
"loss": 1.3475,
"step": 1129
},
{
"epoch": 6.978328173374613,
"grad_norm": 7.504426633437561,
"learning_rate": 3.88622794851838e-05,
"loss": 1.2971,
"step": 1130
},
{
"epoch": 6.984520123839009,
"grad_norm": 7.362734099810955,
"learning_rate": 3.878058068990109e-05,
"loss": 1.2548,
"step": 1131
},
{
"epoch": 6.9907120743034055,
"grad_norm": 6.620299195182782,
"learning_rate": 3.869891342511782e-05,
"loss": 1.2814,
"step": 1132
},
{
"epoch": 6.996904024767802,
"grad_norm": 6.829774476356422,
"learning_rate": 3.8617277920347624e-05,
"loss": 1.3188,
"step": 1133
},
{
"epoch": 7.0,
"grad_norm": 6.829774476356422,
"learning_rate": 3.853567440501489e-05,
"loss": 0.5846,
"step": 1134
},
{
"epoch": 7.006191950464396,
"grad_norm": 5.757452200584078,
"learning_rate": 3.845410310845407e-05,
"loss": 0.7691,
"step": 1135
},
{
"epoch": 7.012383900928793,
"grad_norm": 5.3916472112707075,
"learning_rate": 3.8372564259909086e-05,
"loss": 0.7703,
"step": 1136
},
{
"epoch": 7.018575851393189,
"grad_norm": 5.659933567827994,
"learning_rate": 3.829105808853269e-05,
"loss": 0.7488,
"step": 1137
},
{
"epoch": 7.024767801857585,
"grad_norm": 6.129208627573635,
"learning_rate": 3.820958482338575e-05,
"loss": 0.7068,
"step": 1138
},
{
"epoch": 7.030959752321982,
"grad_norm": 6.4799801807011335,
"learning_rate": 3.812814469343674e-05,
"loss": 0.6127,
"step": 1139
},
{
"epoch": 7.037151702786378,
"grad_norm": 7.22503344073261,
"learning_rate": 3.8046737927560916e-05,
"loss": 0.7233,
"step": 1140
},
{
"epoch": 7.043343653250774,
"grad_norm": 8.578266259964357,
"learning_rate": 3.7965364754539845e-05,
"loss": 0.7226,
"step": 1141
},
{
"epoch": 7.04953560371517,
"grad_norm": 8.853781988845313,
"learning_rate": 3.7884025403060635e-05,
"loss": 0.5882,
"step": 1142
},
{
"epoch": 7.055727554179566,
"grad_norm": 7.8476497582946365,
"learning_rate": 3.780272010171535e-05,
"loss": 0.6746,
"step": 1143
},
{
"epoch": 7.061919504643963,
"grad_norm": 8.091888749091458,
"learning_rate": 3.7721449079000413e-05,
"loss": 0.628,
"step": 1144
},
{
"epoch": 7.068111455108359,
"grad_norm": 8.513359984966414,
"learning_rate": 3.7640212563315865e-05,
"loss": 0.6322,
"step": 1145
},
{
"epoch": 7.074303405572755,
"grad_norm": 7.089706535286417,
"learning_rate": 3.7559010782964776e-05,
"loss": 0.584,
"step": 1146
},
{
"epoch": 7.080495356037152,
"grad_norm": 6.652458662446482,
"learning_rate": 3.747784396615258e-05,
"loss": 0.6509,
"step": 1147
},
{
"epoch": 7.086687306501548,
"grad_norm": 7.757592156583572,
"learning_rate": 3.73967123409865e-05,
"loss": 0.6271,
"step": 1148
},
{
"epoch": 7.0928792569659445,
"grad_norm": 8.0459839029957,
"learning_rate": 3.7315616135474805e-05,
"loss": 0.6932,
"step": 1149
},
{
"epoch": 7.099071207430341,
"grad_norm": 7.747894884862985,
"learning_rate": 3.723455557752628e-05,
"loss": 0.6297,
"step": 1150
},
{
"epoch": 7.105263157894737,
"grad_norm": 8.0768419517607,
"learning_rate": 3.715353089494947e-05,
"loss": 0.695,
"step": 1151
},
{
"epoch": 7.1114551083591335,
"grad_norm": 8.347655110729336,
"learning_rate": 3.707254231545214e-05,
"loss": 0.7334,
"step": 1152
},
{
"epoch": 7.117647058823529,
"grad_norm": 8.610878254087831,
"learning_rate": 3.699159006664056e-05,
"loss": 0.6304,
"step": 1153
},
{
"epoch": 7.123839009287925,
"grad_norm": 8.017941711819208,
"learning_rate": 3.691067437601893e-05,
"loss": 0.6681,
"step": 1154
},
{
"epoch": 7.130030959752322,
"grad_norm": 7.751879604109213,
"learning_rate": 3.682979547098867e-05,
"loss": 0.592,
"step": 1155
},
{
"epoch": 7.136222910216718,
"grad_norm": 7.0682048074196,
"learning_rate": 3.674895357884787e-05,
"loss": 0.6628,
"step": 1156
},
{
"epoch": 7.142414860681114,
"grad_norm": 7.524978104619749,
"learning_rate": 3.666814892679056e-05,
"loss": 0.6057,
"step": 1157
},
{
"epoch": 7.148606811145511,
"grad_norm": 7.645817083711223,
"learning_rate": 3.6587381741906126e-05,
"loss": 0.6442,
"step": 1158
},
{
"epoch": 7.154798761609907,
"grad_norm": 7.607342674234635,
"learning_rate": 3.6506652251178665e-05,
"loss": 0.6556,
"step": 1159
},
{
"epoch": 7.1609907120743035,
"grad_norm": 6.670881036041286,
"learning_rate": 3.6425960681486304e-05,
"loss": 0.6234,
"step": 1160
},
{
"epoch": 7.1671826625387,
"grad_norm": 7.376001317028536,
"learning_rate": 3.6345307259600655e-05,
"loss": 0.6421,
"step": 1161
},
{
"epoch": 7.173374613003096,
"grad_norm": 7.991999140107628,
"learning_rate": 3.62646922121861e-05,
"loss": 0.6461,
"step": 1162
},
{
"epoch": 7.179566563467493,
"grad_norm": 7.122464598105445,
"learning_rate": 3.618411576579916e-05,
"loss": 0.5966,
"step": 1163
},
{
"epoch": 7.185758513931889,
"grad_norm": 7.935536657371515,
"learning_rate": 3.6103578146887864e-05,
"loss": 0.5976,
"step": 1164
},
{
"epoch": 7.191950464396285,
"grad_norm": 8.906340192711372,
"learning_rate": 3.6023079581791166e-05,
"loss": 0.6481,
"step": 1165
},
{
"epoch": 7.198142414860681,
"grad_norm": 8.091902630476214,
"learning_rate": 3.594262029673822e-05,
"loss": 0.6261,
"step": 1166
},
{
"epoch": 7.204334365325077,
"grad_norm": 8.613601969483307,
"learning_rate": 3.5862200517847826e-05,
"loss": 0.7235,
"step": 1167
},
{
"epoch": 7.2105263157894735,
"grad_norm": 8.311992100816648,
"learning_rate": 3.578182047112773e-05,
"loss": 0.6724,
"step": 1168
},
{
"epoch": 7.21671826625387,
"grad_norm": 8.307931162702967,
"learning_rate": 3.570148038247404e-05,
"loss": 0.5556,
"step": 1169
},
{
"epoch": 7.222910216718266,
"grad_norm": 6.854708876762124,
"learning_rate": 3.562118047767056e-05,
"loss": 0.5809,
"step": 1170
},
{
"epoch": 7.2291021671826625,
"grad_norm": 7.807928365862507,
"learning_rate": 3.554092098238811e-05,
"loss": 0.6964,
"step": 1171
},
{
"epoch": 7.235294117647059,
"grad_norm": 8.01432971517405,
"learning_rate": 3.5460702122184045e-05,
"loss": 0.6613,
"step": 1172
},
{
"epoch": 7.241486068111455,
"grad_norm": 8.324636492816218,
"learning_rate": 3.5380524122501466e-05,
"loss": 0.7325,
"step": 1173
},
{
"epoch": 7.247678018575852,
"grad_norm": 8.173897222971299,
"learning_rate": 3.5300387208668636e-05,
"loss": 0.6754,
"step": 1174
},
{
"epoch": 7.253869969040248,
"grad_norm": 8.50762012110652,
"learning_rate": 3.5220291605898355e-05,
"loss": 0.662,
"step": 1175
},
{
"epoch": 7.260061919504644,
"grad_norm": 8.12257677652454,
"learning_rate": 3.514023753928734e-05,
"loss": 0.654,
"step": 1176
},
{
"epoch": 7.266253869969041,
"grad_norm": 8.54175725853606,
"learning_rate": 3.506022523381555e-05,
"loss": 0.5887,
"step": 1177
},
{
"epoch": 7.272445820433436,
"grad_norm": 7.299153779760447,
"learning_rate": 3.4980254914345634e-05,
"loss": 0.6315,
"step": 1178
},
{
"epoch": 7.2786377708978325,
"grad_norm": 8.270759623026917,
"learning_rate": 3.4900326805622184e-05,
"loss": 0.668,
"step": 1179
},
{
"epoch": 7.284829721362229,
"grad_norm": 8.817697433292942,
"learning_rate": 3.482044113227121e-05,
"loss": 0.7288,
"step": 1180
},
{
"epoch": 7.291021671826625,
"grad_norm": 7.929767327385549,
"learning_rate": 3.474059811879944e-05,
"loss": 0.7281,
"step": 1181
},
{
"epoch": 7.2972136222910216,
"grad_norm": 8.535652364006877,
"learning_rate": 3.4660797989593685e-05,
"loss": 0.7105,
"step": 1182
},
{
"epoch": 7.303405572755418,
"grad_norm": 8.356481763272463,
"learning_rate": 3.4581040968920307e-05,
"loss": 0.636,
"step": 1183
},
{
"epoch": 7.309597523219814,
"grad_norm": 8.239685126481039,
"learning_rate": 3.450132728092448e-05,
"loss": 0.5836,
"step": 1184
},
{
"epoch": 7.315789473684211,
"grad_norm": 8.160510633814823,
"learning_rate": 3.4421657149629595e-05,
"loss": 0.7017,
"step": 1185
},
{
"epoch": 7.321981424148607,
"grad_norm": 7.9776807141446975,
"learning_rate": 3.434203079893662e-05,
"loss": 0.6823,
"step": 1186
},
{
"epoch": 7.328173374613003,
"grad_norm": 8.82859100339137,
"learning_rate": 3.426244845262351e-05,
"loss": 0.6992,
"step": 1187
},
{
"epoch": 7.3343653250774,
"grad_norm": 7.9540922370357965,
"learning_rate": 3.418291033434454e-05,
"loss": 0.6119,
"step": 1188
},
{
"epoch": 7.340557275541796,
"grad_norm": 8.399930431271818,
"learning_rate": 3.410341666762971e-05,
"loss": 0.6753,
"step": 1189
},
{
"epoch": 7.346749226006192,
"grad_norm": 8.093683857281743,
"learning_rate": 3.4023967675884046e-05,
"loss": 0.6398,
"step": 1190
},
{
"epoch": 7.352941176470588,
"grad_norm": 7.701390167163244,
"learning_rate": 3.3944563582387084e-05,
"loss": 0.6885,
"step": 1191
},
{
"epoch": 7.359133126934984,
"grad_norm": 7.362680274229581,
"learning_rate": 3.386520461029214e-05,
"loss": 0.6913,
"step": 1192
},
{
"epoch": 7.365325077399381,
"grad_norm": 8.383654254826384,
"learning_rate": 3.37858909826257e-05,
"loss": 0.6565,
"step": 1193
},
{
"epoch": 7.371517027863777,
"grad_norm": 8.096659536515782,
"learning_rate": 3.37066229222869e-05,
"loss": 0.6613,
"step": 1194
},
{
"epoch": 7.377708978328173,
"grad_norm": 8.969915988273861,
"learning_rate": 3.362740065204674e-05,
"loss": 0.6977,
"step": 1195
},
{
"epoch": 7.38390092879257,
"grad_norm": 8.00810413870692,
"learning_rate": 3.354822439454756e-05,
"loss": 0.6401,
"step": 1196
},
{
"epoch": 7.390092879256966,
"grad_norm": 7.801977933581489,
"learning_rate": 3.346909437230237e-05,
"loss": 0.624,
"step": 1197
},
{
"epoch": 7.396284829721362,
"grad_norm": 7.876403238913395,
"learning_rate": 3.3390010807694296e-05,
"loss": 0.6615,
"step": 1198
},
{
"epoch": 7.402476780185759,
"grad_norm": 8.311377001905285,
"learning_rate": 3.331097392297582e-05,
"loss": 0.6783,
"step": 1199
},
{
"epoch": 7.408668730650155,
"grad_norm": 8.23551557778447,
"learning_rate": 3.323198394026832e-05,
"loss": 0.6823,
"step": 1200
},
{
"epoch": 7.414860681114551,
"grad_norm": 7.837471162046131,
"learning_rate": 3.3153041081561295e-05,
"loss": 0.7097,
"step": 1201
},
{
"epoch": 7.421052631578947,
"grad_norm": 8.598917363563292,
"learning_rate": 3.307414556871187e-05,
"loss": 0.6873,
"step": 1202
},
{
"epoch": 7.427244582043343,
"grad_norm": 8.158249463290781,
"learning_rate": 3.299529762344406e-05,
"loss": 0.5893,
"step": 1203
},
{
"epoch": 7.43343653250774,
"grad_norm": 8.313299477526693,
"learning_rate": 3.291649746734821e-05,
"loss": 0.7041,
"step": 1204
},
{
"epoch": 7.439628482972136,
"grad_norm": 7.998392093179114,
"learning_rate": 3.283774532188039e-05,
"loss": 0.7022,
"step": 1205
},
{
"epoch": 7.445820433436532,
"grad_norm": 7.083629588891991,
"learning_rate": 3.275904140836172e-05,
"loss": 0.6606,
"step": 1206
},
{
"epoch": 7.452012383900929,
"grad_norm": 8.31730904563776,
"learning_rate": 3.268038594797777e-05,
"loss": 0.6762,
"step": 1207
},
{
"epoch": 7.458204334365325,
"grad_norm": 8.131121047191598,
"learning_rate": 3.260177916177793e-05,
"loss": 0.6383,
"step": 1208
},
{
"epoch": 7.464396284829721,
"grad_norm": 7.963816667204574,
"learning_rate": 3.2523221270674845e-05,
"loss": 0.689,
"step": 1209
},
{
"epoch": 7.470588235294118,
"grad_norm": 7.721059679347194,
"learning_rate": 3.244471249544366e-05,
"loss": 0.6835,
"step": 1210
},
{
"epoch": 7.476780185758514,
"grad_norm": 7.920354396920408,
"learning_rate": 3.236625305672161e-05,
"loss": 0.6929,
"step": 1211
},
{
"epoch": 7.4829721362229105,
"grad_norm": 7.674607331785555,
"learning_rate": 3.228784317500716e-05,
"loss": 0.671,
"step": 1212
},
{
"epoch": 7.489164086687307,
"grad_norm": 8.495915761327536,
"learning_rate": 3.220948307065959e-05,
"loss": 0.7254,
"step": 1213
},
{
"epoch": 7.495356037151703,
"grad_norm": 8.071322310043705,
"learning_rate": 3.213117296389822e-05,
"loss": 0.7481,
"step": 1214
},
{
"epoch": 7.5015479876160995,
"grad_norm": 8.612027411012539,
"learning_rate": 3.2052913074801874e-05,
"loss": 0.7564,
"step": 1215
},
{
"epoch": 7.507739938080495,
"grad_norm": 8.331997297025653,
"learning_rate": 3.197470362330829e-05,
"loss": 0.6429,
"step": 1216
},
{
"epoch": 7.513931888544891,
"grad_norm": 8.400156559319822,
"learning_rate": 3.189654482921344e-05,
"loss": 0.5517,
"step": 1217
},
{
"epoch": 7.520123839009288,
"grad_norm": 7.088840525035982,
"learning_rate": 3.181843691217091e-05,
"loss": 0.6208,
"step": 1218
},
{
"epoch": 7.526315789473684,
"grad_norm": 8.409192028416426,
"learning_rate": 3.17403800916913e-05,
"loss": 0.6873,
"step": 1219
},
{
"epoch": 7.53250773993808,
"grad_norm": 7.838297064321753,
"learning_rate": 3.166237458714165e-05,
"loss": 0.7176,
"step": 1220
},
{
"epoch": 7.538699690402477,
"grad_norm": 8.046052845799926,
"learning_rate": 3.158442061774474e-05,
"loss": 0.693,
"step": 1221
},
{
"epoch": 7.544891640866873,
"grad_norm": 8.79215963366741,
"learning_rate": 3.150651840257855e-05,
"loss": 0.7204,
"step": 1222
},
{
"epoch": 7.5510835913312695,
"grad_norm": 8.744244445066755,
"learning_rate": 3.142866816057559e-05,
"loss": 0.7261,
"step": 1223
},
{
"epoch": 7.557275541795666,
"grad_norm": 7.9685194893419595,
"learning_rate": 3.1350870110522346e-05,
"loss": 0.6169,
"step": 1224
},
{
"epoch": 7.563467492260062,
"grad_norm": 8.619693891486385,
"learning_rate": 3.1273124471058566e-05,
"loss": 0.7385,
"step": 1225
},
{
"epoch": 7.569659442724459,
"grad_norm": 8.54050500452178,
"learning_rate": 3.119543146067675e-05,
"loss": 0.6851,
"step": 1226
},
{
"epoch": 7.575851393188854,
"grad_norm": 8.140674406800304,
"learning_rate": 3.111779129772147e-05,
"loss": 0.7533,
"step": 1227
},
{
"epoch": 7.58204334365325,
"grad_norm": 7.752084913192541,
"learning_rate": 3.104020420038882e-05,
"loss": 0.6617,
"step": 1228
},
{
"epoch": 7.588235294117647,
"grad_norm": 8.202639445478995,
"learning_rate": 3.09626703867257e-05,
"loss": 0.7212,
"step": 1229
},
{
"epoch": 7.594427244582043,
"grad_norm": 8.188391370662625,
"learning_rate": 3.08851900746293e-05,
"loss": 0.6522,
"step": 1230
},
{
"epoch": 7.6006191950464395,
"grad_norm": 7.66888290973458,
"learning_rate": 3.0807763481846454e-05,
"loss": 0.6824,
"step": 1231
},
{
"epoch": 7.606811145510836,
"grad_norm": 8.085980770238914,
"learning_rate": 3.073039082597299e-05,
"loss": 0.6681,
"step": 1232
},
{
"epoch": 7.613003095975232,
"grad_norm": 7.7566651992345435,
"learning_rate": 3.065307232445322e-05,
"loss": 0.6784,
"step": 1233
},
{
"epoch": 7.6191950464396285,
"grad_norm": 7.978641420587795,
"learning_rate": 3.0575808194579204e-05,
"loss": 0.6736,
"step": 1234
},
{
"epoch": 7.625386996904025,
"grad_norm": 8.20959902355321,
"learning_rate": 3.049859865349023e-05,
"loss": 0.7497,
"step": 1235
},
{
"epoch": 7.631578947368421,
"grad_norm": 8.616008443473586,
"learning_rate": 3.0421443918172155e-05,
"loss": 0.7485,
"step": 1236
},
{
"epoch": 7.637770897832818,
"grad_norm": 8.23181397420113,
"learning_rate": 3.0344344205456803e-05,
"loss": 0.6814,
"step": 1237
},
{
"epoch": 7.643962848297214,
"grad_norm": 8.270338931472901,
"learning_rate": 3.0267299732021392e-05,
"loss": 0.6712,
"step": 1238
},
{
"epoch": 7.65015479876161,
"grad_norm": 8.47844326229016,
"learning_rate": 3.0190310714387915e-05,
"loss": 0.6584,
"step": 1239
},
{
"epoch": 7.656346749226007,
"grad_norm": 7.307865533550394,
"learning_rate": 3.011337736892247e-05,
"loss": 0.6361,
"step": 1240
},
{
"epoch": 7.662538699690402,
"grad_norm": 7.992336785797606,
"learning_rate": 3.0036499911834698e-05,
"loss": 0.7118,
"step": 1241
},
{
"epoch": 7.6687306501547985,
"grad_norm": 8.545711521285286,
"learning_rate": 2.9959678559177217e-05,
"loss": 0.6709,
"step": 1242
},
{
"epoch": 7.674922600619195,
"grad_norm": 7.984466724979602,
"learning_rate": 2.988291352684491e-05,
"loss": 0.6251,
"step": 1243
},
{
"epoch": 7.681114551083591,
"grad_norm": 8.118020286806026,
"learning_rate": 2.9806205030574457e-05,
"loss": 0.6173,
"step": 1244
},
{
"epoch": 7.687306501547988,
"grad_norm": 7.808414608011233,
"learning_rate": 2.9729553285943583e-05,
"loss": 0.6466,
"step": 1245
},
{
"epoch": 7.693498452012384,
"grad_norm": 8.06189837506608,
"learning_rate": 2.9652958508370565e-05,
"loss": 0.6322,
"step": 1246
},
{
"epoch": 7.69969040247678,
"grad_norm": 7.26649091945809,
"learning_rate": 2.9576420913113567e-05,
"loss": 0.6161,
"step": 1247
},
{
"epoch": 7.705882352941177,
"grad_norm": 8.528822640413676,
"learning_rate": 2.9499940715270025e-05,
"loss": 0.6576,
"step": 1248
},
{
"epoch": 7.712074303405573,
"grad_norm": 7.519876903624291,
"learning_rate": 2.9423518129776096e-05,
"loss": 0.5952,
"step": 1249
},
{
"epoch": 7.718266253869969,
"grad_norm": 7.771182888891596,
"learning_rate": 2.9347153371406055e-05,
"loss": 0.578,
"step": 1250
},
{
"epoch": 7.724458204334366,
"grad_norm": 7.921734486415399,
"learning_rate": 2.9270846654771624e-05,
"loss": 0.674,
"step": 1251
},
{
"epoch": 7.730650154798761,
"grad_norm": 8.170468598427377,
"learning_rate": 2.9194598194321377e-05,
"loss": 0.6626,
"step": 1252
},
{
"epoch": 7.7368421052631575,
"grad_norm": 8.860381430779642,
"learning_rate": 2.911840820434024e-05,
"loss": 0.6652,
"step": 1253
},
{
"epoch": 7.743034055727554,
"grad_norm": 7.712114629805272,
"learning_rate": 2.904227689894875e-05,
"loss": 0.6307,
"step": 1254
},
{
"epoch": 7.74922600619195,
"grad_norm": 8.303119371595377,
"learning_rate": 2.8966204492102607e-05,
"loss": 0.582,
"step": 1255
},
{
"epoch": 7.755417956656347,
"grad_norm": 8.264302476582474,
"learning_rate": 2.889019119759187e-05,
"loss": 0.6494,
"step": 1256
},
{
"epoch": 7.761609907120743,
"grad_norm": 9.05326486888945,
"learning_rate": 2.8814237229040552e-05,
"loss": 0.6544,
"step": 1257
},
{
"epoch": 7.767801857585139,
"grad_norm": 8.910884920700793,
"learning_rate": 2.873834279990595e-05,
"loss": 0.6165,
"step": 1258
},
{
"epoch": 7.773993808049536,
"grad_norm": 8.13373948747055,
"learning_rate": 2.866250812347795e-05,
"loss": 0.6575,
"step": 1259
},
{
"epoch": 7.780185758513932,
"grad_norm": 8.601515291139803,
"learning_rate": 2.8586733412878586e-05,
"loss": 0.6822,
"step": 1260
},
{
"epoch": 7.786377708978328,
"grad_norm": 8.046468505636659,
"learning_rate": 2.8511018881061345e-05,
"loss": 0.634,
"step": 1261
},
{
"epoch": 7.792569659442725,
"grad_norm": 8.035658308300395,
"learning_rate": 2.8435364740810598e-05,
"loss": 0.6531,
"step": 1262
},
{
"epoch": 7.798761609907121,
"grad_norm": 8.101062025927892,
"learning_rate": 2.8359771204741e-05,
"loss": 0.66,
"step": 1263
},
{
"epoch": 7.804953560371517,
"grad_norm": 8.38621875259201,
"learning_rate": 2.8284238485296838e-05,
"loss": 0.7224,
"step": 1264
},
{
"epoch": 7.811145510835914,
"grad_norm": 9.258934210222563,
"learning_rate": 2.8208766794751516e-05,
"loss": 0.6884,
"step": 1265
},
{
"epoch": 7.817337461300309,
"grad_norm": 8.119723433617475,
"learning_rate": 2.8133356345206997e-05,
"loss": 0.6528,
"step": 1266
},
{
"epoch": 7.823529411764706,
"grad_norm": 7.350294255838845,
"learning_rate": 2.8058007348593e-05,
"loss": 0.598,
"step": 1267
},
{
"epoch": 7.829721362229102,
"grad_norm": 7.419408450013087,
"learning_rate": 2.798272001666663e-05,
"loss": 0.7033,
"step": 1268
},
{
"epoch": 7.835913312693498,
"grad_norm": 7.6967343676625255,
"learning_rate": 2.7907494561011693e-05,
"loss": 0.6119,
"step": 1269
},
{
"epoch": 7.842105263157895,
"grad_norm": 7.437079870880633,
"learning_rate": 2.7832331193038032e-05,
"loss": 0.6005,
"step": 1270
},
{
"epoch": 7.848297213622291,
"grad_norm": 7.558840566026589,
"learning_rate": 2.775723012398107e-05,
"loss": 0.5996,
"step": 1271
},
{
"epoch": 7.854489164086687,
"grad_norm": 8.740518115405091,
"learning_rate": 2.768219156490113e-05,
"loss": 0.6713,
"step": 1272
},
{
"epoch": 7.860681114551084,
"grad_norm": 8.053879690325598,
"learning_rate": 2.760721572668284e-05,
"loss": 0.6038,
"step": 1273
},
{
"epoch": 7.86687306501548,
"grad_norm": 7.7899110456991645,
"learning_rate": 2.7532302820034615e-05,
"loss": 0.6523,
"step": 1274
},
{
"epoch": 7.8730650154798765,
"grad_norm": 9.394868216923037,
"learning_rate": 2.7457453055487926e-05,
"loss": 0.6368,
"step": 1275
},
{
"epoch": 7.879256965944273,
"grad_norm": 8.943817630593156,
"learning_rate": 2.7382666643396826e-05,
"loss": 0.6528,
"step": 1276
},
{
"epoch": 7.885448916408668,
"grad_norm": 8.256568967345226,
"learning_rate": 2.730794379393742e-05,
"loss": 0.6643,
"step": 1277
},
{
"epoch": 7.891640866873065,
"grad_norm": 7.967517426684291,
"learning_rate": 2.7233284717107023e-05,
"loss": 0.6318,
"step": 1278
},
{
"epoch": 7.897832817337461,
"grad_norm": 9.106426627691587,
"learning_rate": 2.7158689622723815e-05,
"loss": 0.62,
"step": 1279
},
{
"epoch": 7.904024767801857,
"grad_norm": 8.626172731573822,
"learning_rate": 2.708415872042618e-05,
"loss": 0.6653,
"step": 1280
},
{
"epoch": 7.910216718266254,
"grad_norm": 7.899309432529724,
"learning_rate": 2.7009692219672023e-05,
"loss": 0.6217,
"step": 1281
},
{
"epoch": 7.91640866873065,
"grad_norm": 7.437612191892225,
"learning_rate": 2.6935290329738328e-05,
"loss": 0.6785,
"step": 1282
},
{
"epoch": 7.922600619195046,
"grad_norm": 7.836176660059616,
"learning_rate": 2.6860953259720472e-05,
"loss": 0.7083,
"step": 1283
},
{
"epoch": 7.928792569659443,
"grad_norm": 7.250084926834517,
"learning_rate": 2.678668121853167e-05,
"loss": 0.6753,
"step": 1284
},
{
"epoch": 7.934984520123839,
"grad_norm": 8.25845490620734,
"learning_rate": 2.67124744149024e-05,
"loss": 0.596,
"step": 1285
},
{
"epoch": 7.9411764705882355,
"grad_norm": 7.65357158563213,
"learning_rate": 2.6638333057379746e-05,
"loss": 0.5703,
"step": 1286
},
{
"epoch": 7.947368421052632,
"grad_norm": 7.780251735268891,
"learning_rate": 2.6564257354326914e-05,
"loss": 0.5831,
"step": 1287
},
{
"epoch": 7.953560371517028,
"grad_norm": 7.638688702033941,
"learning_rate": 2.6490247513922588e-05,
"loss": 0.6565,
"step": 1288
},
{
"epoch": 7.959752321981425,
"grad_norm": 8.272739197943404,
"learning_rate": 2.641630374416036e-05,
"loss": 0.663,
"step": 1289
},
{
"epoch": 7.965944272445821,
"grad_norm": 8.09381850289007,
"learning_rate": 2.634242625284812e-05,
"loss": 0.6729,
"step": 1290
},
{
"epoch": 7.972136222910216,
"grad_norm": 8.327620204390481,
"learning_rate": 2.626861524760753e-05,
"loss": 0.6725,
"step": 1291
},
{
"epoch": 7.978328173374613,
"grad_norm": 8.199164428109075,
"learning_rate": 2.6194870935873337e-05,
"loss": 0.6722,
"step": 1292
},
{
"epoch": 7.984520123839009,
"grad_norm": 8.09738085151205,
"learning_rate": 2.6121193524892918e-05,
"loss": 0.6153,
"step": 1293
},
{
"epoch": 7.9907120743034055,
"grad_norm": 8.716955283383768,
"learning_rate": 2.6047583221725614e-05,
"loss": 0.7093,
"step": 1294
},
{
"epoch": 7.996904024767802,
"grad_norm": 9.226801772932648,
"learning_rate": 2.597404023324217e-05,
"loss": 0.678,
"step": 1295
},
{
"epoch": 8.0,
"grad_norm": 8.387274641654235,
"learning_rate": 2.5900564766124162e-05,
"loss": 0.2941,
"step": 1296
},
{
"epoch": 8.006191950464396,
"grad_norm": 5.211729588648789,
"learning_rate": 2.582715702686337e-05,
"loss": 0.2631,
"step": 1297
},
{
"epoch": 8.012383900928793,
"grad_norm": 5.90311339456378,
"learning_rate": 2.575381722176128e-05,
"loss": 0.272,
"step": 1298
},
{
"epoch": 8.018575851393189,
"grad_norm": 4.971151141966583,
"learning_rate": 2.5680545556928438e-05,
"loss": 0.2773,
"step": 1299
},
{
"epoch": 8.024767801857585,
"grad_norm": 5.6081242166562895,
"learning_rate": 2.5607342238283903e-05,
"loss": 0.2813,
"step": 1300
},
{
"epoch": 8.030959752321982,
"grad_norm": 5.260206176471535,
"learning_rate": 2.5534207471554646e-05,
"loss": 0.2712,
"step": 1301
},
{
"epoch": 8.037151702786378,
"grad_norm": 5.4508788974381535,
"learning_rate": 2.5461141462275002e-05,
"loss": 0.2218,
"step": 1302
},
{
"epoch": 8.043343653250774,
"grad_norm": 6.0916412994862155,
"learning_rate": 2.5388144415786026e-05,
"loss": 0.258,
"step": 1303
},
{
"epoch": 8.04953560371517,
"grad_norm": 5.9748079893438675,
"learning_rate": 2.531521653723501e-05,
"loss": 0.2532,
"step": 1304
},
{
"epoch": 8.055727554179567,
"grad_norm": 6.449372417182895,
"learning_rate": 2.5242358031574853e-05,
"loss": 0.2116,
"step": 1305
},
{
"epoch": 8.061919504643964,
"grad_norm": 5.822373664510092,
"learning_rate": 2.5169569103563485e-05,
"loss": 0.2504,
"step": 1306
},
{
"epoch": 8.06811145510836,
"grad_norm": 6.606672924694336,
"learning_rate": 2.5096849957763292e-05,
"loss": 0.2287,
"step": 1307
},
{
"epoch": 8.074303405572756,
"grad_norm": 6.5527604451433925,
"learning_rate": 2.5024200798540575e-05,
"loss": 0.2291,
"step": 1308
},
{
"epoch": 8.08049535603715,
"grad_norm": 7.954142111358529,
"learning_rate": 2.4951621830064886e-05,
"loss": 0.2777,
"step": 1309
},
{
"epoch": 8.086687306501547,
"grad_norm": 6.50698468501565,
"learning_rate": 2.4879113256308584e-05,
"loss": 0.2005,
"step": 1310
},
{
"epoch": 8.092879256965944,
"grad_norm": 6.98129438122496,
"learning_rate": 2.4806675281046165e-05,
"loss": 0.2766,
"step": 1311
},
{
"epoch": 8.09907120743034,
"grad_norm": 6.951573545459006,
"learning_rate": 2.473430810785372e-05,
"loss": 0.2362,
"step": 1312
},
{
"epoch": 8.105263157894736,
"grad_norm": 6.40999249025508,
"learning_rate": 2.4662011940108382e-05,
"loss": 0.1954,
"step": 1313
},
{
"epoch": 8.111455108359133,
"grad_norm": 6.203154234738614,
"learning_rate": 2.4589786980987672e-05,
"loss": 0.2153,
"step": 1314
},
{
"epoch": 8.117647058823529,
"grad_norm": 5.763352677386173,
"learning_rate": 2.451763343346906e-05,
"loss": 0.2,
"step": 1315
},
{
"epoch": 8.123839009287925,
"grad_norm": 6.877662577956943,
"learning_rate": 2.444555150032929e-05,
"loss": 0.2158,
"step": 1316
},
{
"epoch": 8.130030959752322,
"grad_norm": 7.234338181496438,
"learning_rate": 2.437354138414385e-05,
"loss": 0.1984,
"step": 1317
},
{
"epoch": 8.136222910216718,
"grad_norm": 6.347476827570805,
"learning_rate": 2.4301603287286402e-05,
"loss": 0.231,
"step": 1318
},
{
"epoch": 8.142414860681114,
"grad_norm": 7.80579664610214,
"learning_rate": 2.4229737411928223e-05,
"loss": 0.2432,
"step": 1319
},
{
"epoch": 8.14860681114551,
"grad_norm": 6.238879965547669,
"learning_rate": 2.415794396003756e-05,
"loss": 0.2232,
"step": 1320
},
{
"epoch": 8.154798761609907,
"grad_norm": 6.607933580001818,
"learning_rate": 2.4086223133379198e-05,
"loss": 0.2415,
"step": 1321
},
{
"epoch": 8.160990712074303,
"grad_norm": 6.673368860084464,
"learning_rate": 2.4014575133513782e-05,
"loss": 0.2437,
"step": 1322
},
{
"epoch": 8.1671826625387,
"grad_norm": 6.892100326398936,
"learning_rate": 2.39430001617973e-05,
"loss": 0.2335,
"step": 1323
},
{
"epoch": 8.173374613003096,
"grad_norm": 6.726872628726199,
"learning_rate": 2.387149841938052e-05,
"loss": 0.2153,
"step": 1324
},
{
"epoch": 8.179566563467493,
"grad_norm": 6.812053411668685,
"learning_rate": 2.3800070107208354e-05,
"loss": 0.2419,
"step": 1325
},
{
"epoch": 8.185758513931889,
"grad_norm": 7.3739792106488125,
"learning_rate": 2.3728715426019388e-05,
"loss": 0.2439,
"step": 1326
},
{
"epoch": 8.191950464396285,
"grad_norm": 5.674140988921804,
"learning_rate": 2.3657434576345332e-05,
"loss": 0.2188,
"step": 1327
},
{
"epoch": 8.198142414860682,
"grad_norm": 7.459872374663674,
"learning_rate": 2.3586227758510298e-05,
"loss": 0.2584,
"step": 1328
},
{
"epoch": 8.204334365325078,
"grad_norm": 7.645791139054171,
"learning_rate": 2.3515095172630408e-05,
"loss": 0.2019,
"step": 1329
},
{
"epoch": 8.210526315789474,
"grad_norm": 6.53752820108343,
"learning_rate": 2.3444037018613175e-05,
"loss": 0.2231,
"step": 1330
},
{
"epoch": 8.21671826625387,
"grad_norm": 6.062725182200797,
"learning_rate": 2.3373053496156867e-05,
"loss": 0.2144,
"step": 1331
},
{
"epoch": 8.222910216718267,
"grad_norm": 6.492996483333465,
"learning_rate": 2.3302144804750074e-05,
"loss": 0.2034,
"step": 1332
},
{
"epoch": 8.229102167182663,
"grad_norm": 6.512855794322941,
"learning_rate": 2.3231311143671075e-05,
"loss": 0.2249,
"step": 1333
},
{
"epoch": 8.235294117647058,
"grad_norm": 6.768017013410216,
"learning_rate": 2.3160552711987264e-05,
"loss": 0.2389,
"step": 1334
},
{
"epoch": 8.241486068111454,
"grad_norm": 7.0049409880668705,
"learning_rate": 2.3089869708554657e-05,
"loss": 0.249,
"step": 1335
},
{
"epoch": 8.24767801857585,
"grad_norm": 7.129557311334126,
"learning_rate": 2.301926233201721e-05,
"loss": 0.2301,
"step": 1336
},
{
"epoch": 8.253869969040247,
"grad_norm": 6.815250387575155,
"learning_rate": 2.2948730780806404e-05,
"loss": 0.221,
"step": 1337
},
{
"epoch": 8.260061919504643,
"grad_norm": 7.607829590891377,
"learning_rate": 2.2878275253140664e-05,
"loss": 0.249,
"step": 1338
},
{
"epoch": 8.26625386996904,
"grad_norm": 5.606348163292512,
"learning_rate": 2.2807895947024643e-05,
"loss": 0.1852,
"step": 1339
},
{
"epoch": 8.272445820433436,
"grad_norm": 7.122673389168984,
"learning_rate": 2.2737593060248886e-05,
"loss": 0.2094,
"step": 1340
},
{
"epoch": 8.278637770897832,
"grad_norm": 6.691918362772205,
"learning_rate": 2.266736679038915e-05,
"loss": 0.1943,
"step": 1341
},
{
"epoch": 8.284829721362229,
"grad_norm": 6.213499448759762,
"learning_rate": 2.2597217334805816e-05,
"loss": 0.2415,
"step": 1342
},
{
"epoch": 8.291021671826625,
"grad_norm": 6.523759773656105,
"learning_rate": 2.2527144890643465e-05,
"loss": 0.2133,
"step": 1343
},
{
"epoch": 8.297213622291022,
"grad_norm": 7.243851567114579,
"learning_rate": 2.2457149654830207e-05,
"loss": 0.24,
"step": 1344
},
{
"epoch": 8.303405572755418,
"grad_norm": 6.073498216754716,
"learning_rate": 2.238723182407719e-05,
"loss": 0.1831,
"step": 1345
},
{
"epoch": 8.309597523219814,
"grad_norm": 6.7832682484635,
"learning_rate": 2.2317391594878028e-05,
"loss": 0.2326,
"step": 1346
},
{
"epoch": 8.31578947368421,
"grad_norm": 6.861708354215309,
"learning_rate": 2.2247629163508204e-05,
"loss": 0.2261,
"step": 1347
},
{
"epoch": 8.321981424148607,
"grad_norm": 5.755130687763803,
"learning_rate": 2.2177944726024608e-05,
"loss": 0.2013,
"step": 1348
},
{
"epoch": 8.328173374613003,
"grad_norm": 6.310832057848415,
"learning_rate": 2.2108338478264933e-05,
"loss": 0.2022,
"step": 1349
},
{
"epoch": 8.3343653250774,
"grad_norm": 7.052846963494786,
"learning_rate": 2.2038810615847116e-05,
"loss": 0.2329,
"step": 1350
},
{
"epoch": 8.340557275541796,
"grad_norm": 7.319191156822584,
"learning_rate": 2.196936133416882e-05,
"loss": 0.2554,
"step": 1351
},
{
"epoch": 8.346749226006192,
"grad_norm": 6.439710870985042,
"learning_rate": 2.1899990828406875e-05,
"loss": 0.219,
"step": 1352
},
{
"epoch": 8.352941176470589,
"grad_norm": 8.63536798331795,
"learning_rate": 2.1830699293516677e-05,
"loss": 0.2386,
"step": 1353
},
{
"epoch": 8.359133126934985,
"grad_norm": 6.487908225097029,
"learning_rate": 2.1761486924231727e-05,
"loss": 0.2433,
"step": 1354
},
{
"epoch": 8.365325077399381,
"grad_norm": 7.390858801166949,
"learning_rate": 2.1692353915063046e-05,
"loss": 0.2371,
"step": 1355
},
{
"epoch": 8.371517027863778,
"grad_norm": 6.47598503313347,
"learning_rate": 2.16233004602986e-05,
"loss": 0.2029,
"step": 1356
},
{
"epoch": 8.377708978328174,
"grad_norm": 6.2867756845983145,
"learning_rate": 2.155432675400283e-05,
"loss": 0.2472,
"step": 1357
},
{
"epoch": 8.38390092879257,
"grad_norm": 6.006656853265457,
"learning_rate": 2.1485432990015958e-05,
"loss": 0.248,
"step": 1358
},
{
"epoch": 8.390092879256965,
"grad_norm": 6.6881962837345865,
"learning_rate": 2.1416619361953637e-05,
"loss": 0.2172,
"step": 1359
},
{
"epoch": 8.396284829721361,
"grad_norm": 6.157484880815013,
"learning_rate": 2.1347886063206267e-05,
"loss": 0.2035,
"step": 1360
},
{
"epoch": 8.402476780185758,
"grad_norm": 7.363173374657094,
"learning_rate": 2.1279233286938504e-05,
"loss": 0.1999,
"step": 1361
},
{
"epoch": 8.408668730650154,
"grad_norm": 7.096607612479734,
"learning_rate": 2.1210661226088703e-05,
"loss": 0.2081,
"step": 1362
},
{
"epoch": 8.41486068111455,
"grad_norm": 6.308380615184922,
"learning_rate": 2.1142170073368395e-05,
"loss": 0.1943,
"step": 1363
},
{
"epoch": 8.421052631578947,
"grad_norm": 8.152189818300819,
"learning_rate": 2.1073760021261684e-05,
"loss": 0.244,
"step": 1364
},
{
"epoch": 8.427244582043343,
"grad_norm": 5.708244269346329,
"learning_rate": 2.100543126202481e-05,
"loss": 0.182,
"step": 1365
},
{
"epoch": 8.43343653250774,
"grad_norm": 5.831920283298659,
"learning_rate": 2.0937183987685526e-05,
"loss": 0.2175,
"step": 1366
},
{
"epoch": 8.439628482972136,
"grad_norm": 6.42210420654083,
"learning_rate": 2.0869018390042588e-05,
"loss": 0.2242,
"step": 1367
},
{
"epoch": 8.445820433436532,
"grad_norm": 6.980364372889547,
"learning_rate": 2.0800934660665223e-05,
"loss": 0.2227,
"step": 1368
},
{
"epoch": 8.452012383900929,
"grad_norm": 6.210489409979137,
"learning_rate": 2.073293299089253e-05,
"loss": 0.2499,
"step": 1369
},
{
"epoch": 8.458204334365325,
"grad_norm": 5.795459108999961,
"learning_rate": 2.066501357183305e-05,
"loss": 0.1959,
"step": 1370
},
{
"epoch": 8.464396284829721,
"grad_norm": 7.284135854418629,
"learning_rate": 2.0597176594364148e-05,
"loss": 0.2239,
"step": 1371
},
{
"epoch": 8.470588235294118,
"grad_norm": 8.90348810962731,
"learning_rate": 2.0529422249131496e-05,
"loss": 0.2376,
"step": 1372
},
{
"epoch": 8.476780185758514,
"grad_norm": 7.866093620555389,
"learning_rate": 2.0461750726548556e-05,
"loss": 0.2203,
"step": 1373
},
{
"epoch": 8.48297213622291,
"grad_norm": 6.599369984837879,
"learning_rate": 2.0394162216796024e-05,
"loss": 0.1969,
"step": 1374
},
{
"epoch": 8.489164086687307,
"grad_norm": 6.975347206985036,
"learning_rate": 2.0326656909821263e-05,
"loss": 0.2117,
"step": 1375
},
{
"epoch": 8.495356037151703,
"grad_norm": 6.442740030001861,
"learning_rate": 2.0259234995337863e-05,
"loss": 0.1838,
"step": 1376
},
{
"epoch": 8.5015479876161,
"grad_norm": 5.373775430255688,
"learning_rate": 2.019189666282501e-05,
"loss": 0.1809,
"step": 1377
},
{
"epoch": 8.507739938080496,
"grad_norm": 7.150905800287018,
"learning_rate": 2.0124642101527035e-05,
"loss": 0.2204,
"step": 1378
},
{
"epoch": 8.513931888544892,
"grad_norm": 7.521629447519654,
"learning_rate": 2.0057471500452825e-05,
"loss": 0.2018,
"step": 1379
},
{
"epoch": 8.520123839009289,
"grad_norm": 6.388463961172899,
"learning_rate": 1.9990385048375278e-05,
"loss": 0.2191,
"step": 1380
},
{
"epoch": 8.526315789473685,
"grad_norm": 6.512424992268914,
"learning_rate": 1.9923382933830836e-05,
"loss": 0.2092,
"step": 1381
},
{
"epoch": 8.532507739938081,
"grad_norm": 6.852904088515052,
"learning_rate": 1.9856465345118925e-05,
"loss": 0.2308,
"step": 1382
},
{
"epoch": 8.538699690402478,
"grad_norm": 5.980947731318695,
"learning_rate": 1.9789632470301422e-05,
"loss": 0.1847,
"step": 1383
},
{
"epoch": 8.544891640866872,
"grad_norm": 6.324483999200606,
"learning_rate": 1.9722884497202116e-05,
"loss": 0.1987,
"step": 1384
},
{
"epoch": 8.551083591331269,
"grad_norm": 8.030618441440813,
"learning_rate": 1.9656221613406218e-05,
"loss": 0.2644,
"step": 1385
},
{
"epoch": 8.557275541795665,
"grad_norm": 6.5345265520961435,
"learning_rate": 1.9589644006259756e-05,
"loss": 0.2416,
"step": 1386
},
{
"epoch": 8.563467492260061,
"grad_norm": 6.397613852570123,
"learning_rate": 1.952315186286915e-05,
"loss": 0.222,
"step": 1387
},
{
"epoch": 8.569659442724458,
"grad_norm": 7.04089489831636,
"learning_rate": 1.945674537010062e-05,
"loss": 0.2027,
"step": 1388
},
{
"epoch": 8.575851393188854,
"grad_norm": 6.97787087592902,
"learning_rate": 1.939042471457968e-05,
"loss": 0.2201,
"step": 1389
},
{
"epoch": 8.58204334365325,
"grad_norm": 6.793799389798548,
"learning_rate": 1.9324190082690636e-05,
"loss": 0.2109,
"step": 1390
},
{
"epoch": 8.588235294117647,
"grad_norm": 5.469745216879858,
"learning_rate": 1.925804166057596e-05,
"loss": 0.1942,
"step": 1391
},
{
"epoch": 8.594427244582043,
"grad_norm": 6.219880888435084,
"learning_rate": 1.9191979634135927e-05,
"loss": 0.2339,
"step": 1392
},
{
"epoch": 8.60061919504644,
"grad_norm": 5.747634765720123,
"learning_rate": 1.9126004189027973e-05,
"loss": 0.1637,
"step": 1393
},
{
"epoch": 8.606811145510836,
"grad_norm": 4.803150560563886,
"learning_rate": 1.906011551066622e-05,
"loss": 0.1709,
"step": 1394
},
{
"epoch": 8.613003095975232,
"grad_norm": 6.09335819477407,
"learning_rate": 1.8994313784220942e-05,
"loss": 0.2319,
"step": 1395
},
{
"epoch": 8.619195046439629,
"grad_norm": 6.380835466587691,
"learning_rate": 1.8928599194618052e-05,
"loss": 0.2321,
"step": 1396
},
{
"epoch": 8.625386996904025,
"grad_norm": 6.69520005892915,
"learning_rate": 1.8862971926538553e-05,
"loss": 0.2249,
"step": 1397
},
{
"epoch": 8.631578947368421,
"grad_norm": 6.22365941945241,
"learning_rate": 1.8797432164418072e-05,
"loss": 0.1636,
"step": 1398
},
{
"epoch": 8.637770897832818,
"grad_norm": 7.112655886617888,
"learning_rate": 1.8731980092446306e-05,
"loss": 0.2064,
"step": 1399
},
{
"epoch": 8.643962848297214,
"grad_norm": 7.141282558450403,
"learning_rate": 1.8666615894566515e-05,
"loss": 0.2132,
"step": 1400
},
{
"epoch": 8.65015479876161,
"grad_norm": 5.79129211335436,
"learning_rate": 1.8601339754475005e-05,
"loss": 0.1898,
"step": 1401
},
{
"epoch": 8.656346749226007,
"grad_norm": 7.147248850656452,
"learning_rate": 1.853615185562058e-05,
"loss": 0.2155,
"step": 1402
},
{
"epoch": 8.662538699690403,
"grad_norm": 6.414338629491582,
"learning_rate": 1.847105238120409e-05,
"loss": 0.1884,
"step": 1403
},
{
"epoch": 8.6687306501548,
"grad_norm": 7.722688748172245,
"learning_rate": 1.8406041514177868e-05,
"loss": 0.2542,
"step": 1404
},
{
"epoch": 8.674922600619196,
"grad_norm": 5.937274557351227,
"learning_rate": 1.834111943724523e-05,
"loss": 0.1901,
"step": 1405
},
{
"epoch": 8.681114551083592,
"grad_norm": 6.007815234092331,
"learning_rate": 1.827628633285996e-05,
"loss": 0.1936,
"step": 1406
},
{
"epoch": 8.687306501547988,
"grad_norm": 7.1847410434373975,
"learning_rate": 1.821154238322581e-05,
"loss": 0.2595,
"step": 1407
},
{
"epoch": 8.693498452012385,
"grad_norm": 7.077045758363068,
"learning_rate": 1.8146887770295938e-05,
"loss": 0.1876,
"step": 1408
},
{
"epoch": 8.69969040247678,
"grad_norm": 7.12292958517752,
"learning_rate": 1.808232267577248e-05,
"loss": 0.2041,
"step": 1409
},
{
"epoch": 8.705882352941176,
"grad_norm": 6.1739975800363505,
"learning_rate": 1.801784728110596e-05,
"loss": 0.2048,
"step": 1410
},
{
"epoch": 8.712074303405572,
"grad_norm": 7.536190701947081,
"learning_rate": 1.795346176749484e-05,
"loss": 0.2403,
"step": 1411
},
{
"epoch": 8.718266253869968,
"grad_norm": 6.304015603478306,
"learning_rate": 1.7889166315884975e-05,
"loss": 0.1865,
"step": 1412
},
{
"epoch": 8.724458204334365,
"grad_norm": 7.579468960849294,
"learning_rate": 1.7824961106969124e-05,
"loss": 0.2214,
"step": 1413
},
{
"epoch": 8.730650154798761,
"grad_norm": 7.438004355915212,
"learning_rate": 1.7760846321186386e-05,
"loss": 0.2221,
"step": 1414
},
{
"epoch": 8.736842105263158,
"grad_norm": 7.413131864673324,
"learning_rate": 1.7696822138721798e-05,
"loss": 0.1988,
"step": 1415
},
{
"epoch": 8.743034055727554,
"grad_norm": 6.689688611213779,
"learning_rate": 1.7632888739505742e-05,
"loss": 0.1904,
"step": 1416
},
{
"epoch": 8.74922600619195,
"grad_norm": 6.985013462114578,
"learning_rate": 1.756904630321347e-05,
"loss": 0.217,
"step": 1417
},
{
"epoch": 8.755417956656347,
"grad_norm": 7.116843076668186,
"learning_rate": 1.7505295009264616e-05,
"loss": 0.229,
"step": 1418
},
{
"epoch": 8.761609907120743,
"grad_norm": 6.850052751434371,
"learning_rate": 1.7441635036822624e-05,
"loss": 0.2162,
"step": 1419
},
{
"epoch": 8.76780185758514,
"grad_norm": 5.339204147705435,
"learning_rate": 1.737806656479433e-05,
"loss": 0.1688,
"step": 1420
},
{
"epoch": 8.773993808049536,
"grad_norm": 5.908151183271524,
"learning_rate": 1.7314589771829427e-05,
"loss": 0.1885,
"step": 1421
},
{
"epoch": 8.780185758513932,
"grad_norm": 6.000434459305432,
"learning_rate": 1.7251204836319935e-05,
"loss": 0.1856,
"step": 1422
},
{
"epoch": 8.786377708978328,
"grad_norm": 6.598456106670304,
"learning_rate": 1.718791193639973e-05,
"loss": 0.1943,
"step": 1423
},
{
"epoch": 8.792569659442725,
"grad_norm": 6.881360078803345,
"learning_rate": 1.712471124994407e-05,
"loss": 0.2141,
"step": 1424
},
{
"epoch": 8.798761609907121,
"grad_norm": 6.513661403234104,
"learning_rate": 1.706160295456898e-05,
"loss": 0.1744,
"step": 1425
},
{
"epoch": 8.804953560371517,
"grad_norm": 7.090429945824626,
"learning_rate": 1.6998587227630904e-05,
"loss": 0.2304,
"step": 1426
},
{
"epoch": 8.811145510835914,
"grad_norm": 5.873530151320455,
"learning_rate": 1.6935664246226118e-05,
"loss": 0.2002,
"step": 1427
},
{
"epoch": 8.81733746130031,
"grad_norm": 6.590128634706223,
"learning_rate": 1.6872834187190245e-05,
"loss": 0.2246,
"step": 1428
},
{
"epoch": 8.823529411764707,
"grad_norm": 5.970723590237332,
"learning_rate": 1.6810097227097782e-05,
"loss": 0.1959,
"step": 1429
},
{
"epoch": 8.829721362229103,
"grad_norm": 6.0325861218360615,
"learning_rate": 1.6747453542261538e-05,
"loss": 0.1852,
"step": 1430
},
{
"epoch": 8.8359133126935,
"grad_norm": 6.203393082510939,
"learning_rate": 1.6684903308732232e-05,
"loss": 0.208,
"step": 1431
},
{
"epoch": 8.842105263157894,
"grad_norm": 6.037468114687763,
"learning_rate": 1.6622446702297935e-05,
"loss": 0.208,
"step": 1432
},
{
"epoch": 8.848297213622292,
"grad_norm": 5.4456839372955494,
"learning_rate": 1.6560083898483596e-05,
"loss": 0.1771,
"step": 1433
},
{
"epoch": 8.854489164086687,
"grad_norm": 6.83245124590101,
"learning_rate": 1.6497815072550555e-05,
"loss": 0.2057,
"step": 1434
},
{
"epoch": 8.860681114551083,
"grad_norm": 5.735714387101287,
"learning_rate": 1.6435640399496034e-05,
"loss": 0.1699,
"step": 1435
},
{
"epoch": 8.86687306501548,
"grad_norm": 7.546965433664666,
"learning_rate": 1.6373560054052627e-05,
"loss": 0.1851,
"step": 1436
},
{
"epoch": 8.873065015479876,
"grad_norm": 5.8595469659092565,
"learning_rate": 1.6311574210687862e-05,
"loss": 0.1618,
"step": 1437
},
{
"epoch": 8.879256965944272,
"grad_norm": 6.0645864067771615,
"learning_rate": 1.6249683043603696e-05,
"loss": 0.1613,
"step": 1438
},
{
"epoch": 8.885448916408668,
"grad_norm": 6.816007484225182,
"learning_rate": 1.618788672673598e-05,
"loss": 0.2141,
"step": 1439
},
{
"epoch": 8.891640866873065,
"grad_norm": 6.490848833243287,
"learning_rate": 1.6126185433754033e-05,
"loss": 0.1743,
"step": 1440
},
{
"epoch": 8.897832817337461,
"grad_norm": 6.304776051555149,
"learning_rate": 1.6064579338060087e-05,
"loss": 0.2084,
"step": 1441
},
{
"epoch": 8.904024767801857,
"grad_norm": 6.992876586920148,
"learning_rate": 1.6003068612788864e-05,
"loss": 0.1885,
"step": 1442
},
{
"epoch": 8.910216718266254,
"grad_norm": 6.879469286384281,
"learning_rate": 1.594165343080705e-05,
"loss": 0.2046,
"step": 1443
},
{
"epoch": 8.91640866873065,
"grad_norm": 7.295873180238467,
"learning_rate": 1.588033396471284e-05,
"loss": 0.2236,
"step": 1444
},
{
"epoch": 8.922600619195046,
"grad_norm": 5.737170436178231,
"learning_rate": 1.5819110386835413e-05,
"loss": 0.1985,
"step": 1445
},
{
"epoch": 8.928792569659443,
"grad_norm": 5.2389320863854865,
"learning_rate": 1.575798286923449e-05,
"loss": 0.1792,
"step": 1446
},
{
"epoch": 8.93498452012384,
"grad_norm": 7.020547241770886,
"learning_rate": 1.5696951583699777e-05,
"loss": 0.1968,
"step": 1447
},
{
"epoch": 8.941176470588236,
"grad_norm": 6.067220274895548,
"learning_rate": 1.56360167017506e-05,
"loss": 0.2032,
"step": 1448
},
{
"epoch": 8.947368421052632,
"grad_norm": 6.526140400640363,
"learning_rate": 1.5575178394635316e-05,
"loss": 0.1789,
"step": 1449
},
{
"epoch": 8.953560371517028,
"grad_norm": 6.8961109487146,
"learning_rate": 1.5514436833330892e-05,
"loss": 0.2225,
"step": 1450
},
{
"epoch": 8.959752321981425,
"grad_norm": 5.92248307884942,
"learning_rate": 1.545379218854241e-05,
"loss": 0.1735,
"step": 1451
},
{
"epoch": 8.965944272445821,
"grad_norm": 6.033455110902586,
"learning_rate": 1.5393244630702542e-05,
"loss": 0.1866,
"step": 1452
},
{
"epoch": 8.972136222910217,
"grad_norm": 5.815481467106758,
"learning_rate": 1.5332794329971155e-05,
"loss": 0.1862,
"step": 1453
},
{
"epoch": 8.978328173374614,
"grad_norm": 6.187467776033898,
"learning_rate": 1.5272441456234783e-05,
"loss": 0.1646,
"step": 1454
},
{
"epoch": 8.98452012383901,
"grad_norm": 6.001636770763615,
"learning_rate": 1.5212186179106141e-05,
"loss": 0.1628,
"step": 1455
},
{
"epoch": 8.990712074303406,
"grad_norm": 6.427541654539285,
"learning_rate": 1.5152028667923668e-05,
"loss": 0.1953,
"step": 1456
},
{
"epoch": 8.996904024767801,
"grad_norm": 6.26902193898391,
"learning_rate": 1.5091969091751074e-05,
"loss": 0.1708,
"step": 1457
},
{
"epoch": 9.0,
"grad_norm": 6.26902193898391,
"learning_rate": 1.503200761937677e-05,
"loss": 0.0859,
"step": 1458
},
{
"epoch": 9.006191950464396,
"grad_norm": 4.71081559874955,
"learning_rate": 1.4972144419313527e-05,
"loss": 0.0635,
"step": 1459
},
{
"epoch": 9.012383900928793,
"grad_norm": 2.650543963787472,
"learning_rate": 1.4912379659797909e-05,
"loss": 0.0683,
"step": 1460
},
{
"epoch": 9.018575851393189,
"grad_norm": 3.872097740667003,
"learning_rate": 1.4852713508789833e-05,
"loss": 0.0521,
"step": 1461
},
{
"epoch": 9.024767801857585,
"grad_norm": 3.016760985315231,
"learning_rate": 1.4793146133972108e-05,
"loss": 0.0805,
"step": 1462
},
{
"epoch": 9.030959752321982,
"grad_norm": 2.6210030999981724,
"learning_rate": 1.4733677702749893e-05,
"loss": 0.0445,
"step": 1463
},
{
"epoch": 9.037151702786378,
"grad_norm": 2.7800323781665552,
"learning_rate": 1.4674308382250334e-05,
"loss": 0.0571,
"step": 1464
},
{
"epoch": 9.043343653250774,
"grad_norm": 3.093182907029037,
"learning_rate": 1.4615038339322024e-05,
"loss": 0.0727,
"step": 1465
},
{
"epoch": 9.04953560371517,
"grad_norm": 5.291876382279551,
"learning_rate": 1.4555867740534552e-05,
"loss": 0.0537,
"step": 1466
},
{
"epoch": 9.055727554179567,
"grad_norm": 3.151990570075643,
"learning_rate": 1.4496796752178033e-05,
"loss": 0.0639,
"step": 1467
},
{
"epoch": 9.061919504643964,
"grad_norm": 3.217456007131016,
"learning_rate": 1.4437825540262661e-05,
"loss": 0.0498,
"step": 1468
},
{
"epoch": 9.06811145510836,
"grad_norm": 3.2934700680884474,
"learning_rate": 1.437895427051817e-05,
"loss": 0.0535,
"step": 1469
},
{
"epoch": 9.074303405572756,
"grad_norm": 3.685437107993493,
"learning_rate": 1.4320183108393465e-05,
"loss": 0.0638,
"step": 1470
},
{
"epoch": 9.08049535603715,
"grad_norm": 4.129874568835661,
"learning_rate": 1.4261512219056117e-05,
"loss": 0.0592,
"step": 1471
},
{
"epoch": 9.086687306501547,
"grad_norm": 3.600504013809219,
"learning_rate": 1.420294176739188e-05,
"loss": 0.0565,
"step": 1472
},
{
"epoch": 9.092879256965944,
"grad_norm": 3.7328453732235576,
"learning_rate": 1.4144471918004254e-05,
"loss": 0.0596,
"step": 1473
},
{
"epoch": 9.09907120743034,
"grad_norm": 4.263048430319195,
"learning_rate": 1.4086102835213977e-05,
"loss": 0.0651,
"step": 1474
},
{
"epoch": 9.105263157894736,
"grad_norm": 3.7784042882751634,
"learning_rate": 1.4027834683058639e-05,
"loss": 0.0535,
"step": 1475
},
{
"epoch": 9.111455108359133,
"grad_norm": 3.1593539758220412,
"learning_rate": 1.3969667625292165e-05,
"loss": 0.0478,
"step": 1476
},
{
"epoch": 9.117647058823529,
"grad_norm": 3.719540430139627,
"learning_rate": 1.3911601825384368e-05,
"loss": 0.058,
"step": 1477
},
{
"epoch": 9.123839009287925,
"grad_norm": 3.519368419406696,
"learning_rate": 1.385363744652049e-05,
"loss": 0.0467,
"step": 1478
},
{
"epoch": 9.130030959752322,
"grad_norm": 3.0737082235214945,
"learning_rate": 1.3795774651600757e-05,
"loss": 0.0544,
"step": 1479
},
{
"epoch": 9.136222910216718,
"grad_norm": 3.5400720534438386,
"learning_rate": 1.373801360323988e-05,
"loss": 0.0581,
"step": 1480
},
{
"epoch": 9.142414860681114,
"grad_norm": 2.523122843907356,
"learning_rate": 1.3680354463766642e-05,
"loss": 0.0408,
"step": 1481
},
{
"epoch": 9.14860681114551,
"grad_norm": 3.003305768858099,
"learning_rate": 1.3622797395223442e-05,
"loss": 0.0535,
"step": 1482
},
{
"epoch": 9.154798761609907,
"grad_norm": 4.316472018355042,
"learning_rate": 1.3565342559365807e-05,
"loss": 0.0439,
"step": 1483
},
{
"epoch": 9.160990712074303,
"grad_norm": 2.797056168114164,
"learning_rate": 1.3507990117661968e-05,
"loss": 0.0549,
"step": 1484
},
{
"epoch": 9.1671826625387,
"grad_norm": 3.675743322728183,
"learning_rate": 1.3450740231292352e-05,
"loss": 0.0524,
"step": 1485
},
{
"epoch": 9.173374613003096,
"grad_norm": 4.491453209110375,
"learning_rate": 1.3393593061149224e-05,
"loss": 0.0484,
"step": 1486
},
{
"epoch": 9.179566563467493,
"grad_norm": 3.7689249995584087,
"learning_rate": 1.3336548767836144e-05,
"loss": 0.0617,
"step": 1487
},
{
"epoch": 9.185758513931889,
"grad_norm": 3.454412071410455,
"learning_rate": 1.3279607511667568e-05,
"loss": 0.0571,
"step": 1488
},
{
"epoch": 9.191950464396285,
"grad_norm": 3.991374634067102,
"learning_rate": 1.3222769452668381e-05,
"loss": 0.048,
"step": 1489
},
{
"epoch": 9.198142414860682,
"grad_norm": 3.3810153789316537,
"learning_rate": 1.3166034750573452e-05,
"loss": 0.048,
"step": 1490
},
{
"epoch": 9.204334365325078,
"grad_norm": 3.3604867226265998,
"learning_rate": 1.3109403564827155e-05,
"loss": 0.0707,
"step": 1491
},
{
"epoch": 9.210526315789474,
"grad_norm": 3.6629173401514836,
"learning_rate": 1.3052876054582968e-05,
"loss": 0.0417,
"step": 1492
},
{
"epoch": 9.21671826625387,
"grad_norm": 3.3365860079861114,
"learning_rate": 1.2996452378703012e-05,
"loss": 0.0508,
"step": 1493
},
{
"epoch": 9.222910216718267,
"grad_norm": 4.766038401106651,
"learning_rate": 1.2940132695757573e-05,
"loss": 0.0498,
"step": 1494
},
{
"epoch": 9.229102167182663,
"grad_norm": 3.1082638740258965,
"learning_rate": 1.2883917164024722e-05,
"loss": 0.055,
"step": 1495
},
{
"epoch": 9.235294117647058,
"grad_norm": 3.2643471969192897,
"learning_rate": 1.2827805941489769e-05,
"loss": 0.0558,
"step": 1496
},
{
"epoch": 9.241486068111454,
"grad_norm": 2.757421991613139,
"learning_rate": 1.2771799185844913e-05,
"loss": 0.0473,
"step": 1497
},
{
"epoch": 9.24767801857585,
"grad_norm": 3.585200638599601,
"learning_rate": 1.2715897054488779e-05,
"loss": 0.0404,
"step": 1498
},
{
"epoch": 9.253869969040247,
"grad_norm": 3.4491133898528674,
"learning_rate": 1.266009970452593e-05,
"loss": 0.0411,
"step": 1499
},
{
"epoch": 9.260061919504643,
"grad_norm": 3.232214722741945,
"learning_rate": 1.2604407292766479e-05,
"loss": 0.0434,
"step": 1500
},
{
"epoch": 9.26625386996904,
"grad_norm": 3.792874605987716,
"learning_rate": 1.2548819975725624e-05,
"loss": 0.0382,
"step": 1501
},
{
"epoch": 9.272445820433436,
"grad_norm": 3.6381123169244143,
"learning_rate": 1.2493337909623182e-05,
"loss": 0.0416,
"step": 1502
},
{
"epoch": 9.278637770897832,
"grad_norm": 3.6507679032886444,
"learning_rate": 1.2437961250383207e-05,
"loss": 0.0341,
"step": 1503
},
{
"epoch": 9.284829721362229,
"grad_norm": 3.3853346480918076,
"learning_rate": 1.238269015363352e-05,
"loss": 0.0437,
"step": 1504
},
{
"epoch": 9.291021671826625,
"grad_norm": 2.9883819106280134,
"learning_rate": 1.2327524774705267e-05,
"loss": 0.0437,
"step": 1505
},
{
"epoch": 9.297213622291022,
"grad_norm": 2.698885880068701,
"learning_rate": 1.227246526863251e-05,
"loss": 0.0445,
"step": 1506
},
{
"epoch": 9.303405572755418,
"grad_norm": 3.9540399861290294,
"learning_rate": 1.221751179015172e-05,
"loss": 0.0437,
"step": 1507
},
{
"epoch": 9.309597523219814,
"grad_norm": 3.1667486953886526,
"learning_rate": 1.2162664493701437e-05,
"loss": 0.0316,
"step": 1508
},
{
"epoch": 9.31578947368421,
"grad_norm": 3.8571973311075203,
"learning_rate": 1.2107923533421795e-05,
"loss": 0.0524,
"step": 1509
},
{
"epoch": 9.321981424148607,
"grad_norm": 4.2698398198406675,
"learning_rate": 1.205328906315406e-05,
"loss": 0.0496,
"step": 1510
},
{
"epoch": 9.328173374613003,
"grad_norm": 4.64163479354582,
"learning_rate": 1.1998761236440248e-05,
"loss": 0.0544,
"step": 1511
},
{
"epoch": 9.3343653250774,
"grad_norm": 3.5948294289307836,
"learning_rate": 1.1944340206522665e-05,
"loss": 0.0397,
"step": 1512
},
{
"epoch": 9.340557275541796,
"grad_norm": 4.335126173851826,
"learning_rate": 1.1890026126343445e-05,
"loss": 0.0419,
"step": 1513
},
{
"epoch": 9.346749226006192,
"grad_norm": 2.7728511480148628,
"learning_rate": 1.1835819148544203e-05,
"loss": 0.0444,
"step": 1514
},
{
"epoch": 9.352941176470589,
"grad_norm": 4.46426094098098,
"learning_rate": 1.1781719425465537e-05,
"loss": 0.0398,
"step": 1515
},
{
"epoch": 9.359133126934985,
"grad_norm": 2.9224509475638274,
"learning_rate": 1.172772710914663e-05,
"loss": 0.0349,
"step": 1516
},
{
"epoch": 9.365325077399381,
"grad_norm": 3.2074537383154773,
"learning_rate": 1.1673842351324816e-05,
"loss": 0.0495,
"step": 1517
},
{
"epoch": 9.371517027863778,
"grad_norm": 3.7593720134605926,
"learning_rate": 1.1620065303435123e-05,
"loss": 0.045,
"step": 1518
},
{
"epoch": 9.377708978328174,
"grad_norm": 3.425909943910001,
"learning_rate": 1.1566396116609907e-05,
"loss": 0.0359,
"step": 1519
},
{
"epoch": 9.38390092879257,
"grad_norm": 2.5543285672297107,
"learning_rate": 1.1512834941678392e-05,
"loss": 0.0511,
"step": 1520
},
{
"epoch": 9.390092879256965,
"grad_norm": 3.726490945738635,
"learning_rate": 1.1459381929166251e-05,
"loss": 0.0444,
"step": 1521
},
{
"epoch": 9.396284829721361,
"grad_norm": 3.474670891644341,
"learning_rate": 1.1406037229295168e-05,
"loss": 0.0417,
"step": 1522
},
{
"epoch": 9.402476780185758,
"grad_norm": 4.059480970388856,
"learning_rate": 1.1352800991982466e-05,
"loss": 0.0397,
"step": 1523
},
{
"epoch": 9.408668730650154,
"grad_norm": 3.8019192883784636,
"learning_rate": 1.129967336684059e-05,
"loss": 0.061,
"step": 1524
},
{
"epoch": 9.41486068111455,
"grad_norm": 4.264722681190781,
"learning_rate": 1.1246654503176796e-05,
"loss": 0.0385,
"step": 1525
},
{
"epoch": 9.421052631578947,
"grad_norm": 2.805991694585358,
"learning_rate": 1.1193744549992673e-05,
"loss": 0.0481,
"step": 1526
},
{
"epoch": 9.427244582043343,
"grad_norm": 4.895292229788722,
"learning_rate": 1.1140943655983727e-05,
"loss": 0.0287,
"step": 1527
},
{
"epoch": 9.43343653250774,
"grad_norm": 2.8298220889172376,
"learning_rate": 1.1088251969538971e-05,
"loss": 0.0308,
"step": 1528
},
{
"epoch": 9.439628482972136,
"grad_norm": 3.017727827780755,
"learning_rate": 1.103566963874052e-05,
"loss": 0.0414,
"step": 1529
},
{
"epoch": 9.445820433436532,
"grad_norm": 3.134114519650563,
"learning_rate": 1.0983196811363117e-05,
"loss": 0.0455,
"step": 1530
},
{
"epoch": 9.452012383900929,
"grad_norm": 3.853461797471628,
"learning_rate": 1.093083363487381e-05,
"loss": 0.0471,
"step": 1531
},
{
"epoch": 9.458204334365325,
"grad_norm": 3.784491508037374,
"learning_rate": 1.0878580256431476e-05,
"loss": 0.0333,
"step": 1532
},
{
"epoch": 9.464396284829721,
"grad_norm": 3.1013783605586096,
"learning_rate": 1.082643682288641e-05,
"loss": 0.0445,
"step": 1533
},
{
"epoch": 9.470588235294118,
"grad_norm": 3.024386287468399,
"learning_rate": 1.0774403480779954e-05,
"loss": 0.0398,
"step": 1534
},
{
"epoch": 9.476780185758514,
"grad_norm": 2.6304318358318546,
"learning_rate": 1.0722480376343996e-05,
"loss": 0.0365,
"step": 1535
},
{
"epoch": 9.48297213622291,
"grad_norm": 3.6055280547773823,
"learning_rate": 1.0670667655500665e-05,
"loss": 0.0384,
"step": 1536
},
{
"epoch": 9.489164086687307,
"grad_norm": 4.25296133447424,
"learning_rate": 1.0618965463861868e-05,
"loss": 0.0699,
"step": 1537
},
{
"epoch": 9.495356037151703,
"grad_norm": 3.0854868556894988,
"learning_rate": 1.0567373946728882e-05,
"loss": 0.037,
"step": 1538
},
{
"epoch": 9.5015479876161,
"grad_norm": 3.536570489303969,
"learning_rate": 1.0515893249091935e-05,
"loss": 0.0421,
"step": 1539
},
{
"epoch": 9.507739938080496,
"grad_norm": 3.922968140133842,
"learning_rate": 1.0464523515629849e-05,
"loss": 0.0457,
"step": 1540
},
{
"epoch": 9.513931888544892,
"grad_norm": 3.7980928274918324,
"learning_rate": 1.0413264890709546e-05,
"loss": 0.041,
"step": 1541
},
{
"epoch": 9.520123839009289,
"grad_norm": 2.9689972166106764,
"learning_rate": 1.0362117518385733e-05,
"loss": 0.0345,
"step": 1542
},
{
"epoch": 9.526315789473685,
"grad_norm": 3.7882224991986995,
"learning_rate": 1.0311081542400453e-05,
"loss": 0.0503,
"step": 1543
},
{
"epoch": 9.532507739938081,
"grad_norm": 3.4973928447710425,
"learning_rate": 1.0260157106182672e-05,
"loss": 0.0426,
"step": 1544
},
{
"epoch": 9.538699690402478,
"grad_norm": 4.1842225938523585,
"learning_rate": 1.0209344352847921e-05,
"loss": 0.0336,
"step": 1545
},
{
"epoch": 9.544891640866872,
"grad_norm": 2.9611668772280684,
"learning_rate": 1.0158643425197818e-05,
"loss": 0.0361,
"step": 1546
},
{
"epoch": 9.551083591331269,
"grad_norm": 2.6698421025685297,
"learning_rate": 1.0108054465719736e-05,
"loss": 0.0326,
"step": 1547
},
{
"epoch": 9.557275541795665,
"grad_norm": 2.907941122419673,
"learning_rate": 1.0057577616586384e-05,
"loss": 0.0475,
"step": 1548
},
{
"epoch": 9.563467492260061,
"grad_norm": 3.3638184639805364,
"learning_rate": 1.0007213019655393e-05,
"loss": 0.0376,
"step": 1549
},
{
"epoch": 9.569659442724458,
"grad_norm": 4.1309601156713605,
"learning_rate": 9.95696081646893e-06,
"loss": 0.0369,
"step": 1550
},
{
"epoch": 9.575851393188854,
"grad_norm": 3.003076302057863,
"learning_rate": 9.906821148253304e-06,
"loss": 0.0372,
"step": 1551
},
{
"epoch": 9.58204334365325,
"grad_norm": 2.860865564137365,
"learning_rate": 9.856794155918525e-06,
"loss": 0.0398,
"step": 1552
},
{
"epoch": 9.588235294117647,
"grad_norm": 3.4619849919989623,
"learning_rate": 9.806879980057993e-06,
"loss": 0.035,
"step": 1553
},
{
"epoch": 9.594427244582043,
"grad_norm": 2.6605517772057286,
"learning_rate": 9.757078760948013e-06,
"loss": 0.036,
"step": 1554
},
{
"epoch": 9.60061919504644,
"grad_norm": 2.8748160828272398,
"learning_rate": 9.70739063854748e-06,
"loss": 0.0419,
"step": 1555
},
{
"epoch": 9.606811145510836,
"grad_norm": 3.993546683123117,
"learning_rate": 9.657815752497428e-06,
"loss": 0.0462,
"step": 1556
},
{
"epoch": 9.613003095975232,
"grad_norm": 3.6620329861297436,
"learning_rate": 9.608354242120638e-06,
"loss": 0.0473,
"step": 1557
},
{
"epoch": 9.619195046439629,
"grad_norm": 3.6542556245789153,
"learning_rate": 9.559006246421292e-06,
"loss": 0.0313,
"step": 1558
},
{
"epoch": 9.625386996904025,
"grad_norm": 3.488855274052928,
"learning_rate": 9.509771904084557e-06,
"loss": 0.0452,
"step": 1559
},
{
"epoch": 9.631578947368421,
"grad_norm": 2.707239777436621,
"learning_rate": 9.46065135347618e-06,
"loss": 0.0332,
"step": 1560
},
{
"epoch": 9.637770897832818,
"grad_norm": 3.463158996741532,
"learning_rate": 9.411644732642122e-06,
"loss": 0.0356,
"step": 1561
},
{
"epoch": 9.643962848297214,
"grad_norm": 2.950857156541241,
"learning_rate": 9.362752179308176e-06,
"loss": 0.0332,
"step": 1562
},
{
"epoch": 9.65015479876161,
"grad_norm": 6.4610121276298855,
"learning_rate": 9.313973830879513e-06,
"loss": 0.0584,
"step": 1563
},
{
"epoch": 9.656346749226007,
"grad_norm": 3.783311982771598,
"learning_rate": 9.265309824440404e-06,
"loss": 0.0383,
"step": 1564
},
{
"epoch": 9.662538699690403,
"grad_norm": 2.8594458661572455,
"learning_rate": 9.216760296753756e-06,
"loss": 0.0352,
"step": 1565
},
{
"epoch": 9.6687306501548,
"grad_norm": 2.568500500523965,
"learning_rate": 9.16832538426075e-06,
"loss": 0.0321,
"step": 1566
},
{
"epoch": 9.674922600619196,
"grad_norm": 3.742127090672312,
"learning_rate": 9.120005223080485e-06,
"loss": 0.0467,
"step": 1567
},
{
"epoch": 9.681114551083592,
"grad_norm": 4.219351190704549,
"learning_rate": 9.071799949009508e-06,
"loss": 0.0365,
"step": 1568
},
{
"epoch": 9.687306501547988,
"grad_norm": 3.1755991636644545,
"learning_rate": 9.023709697521543e-06,
"loss": 0.0407,
"step": 1569
},
{
"epoch": 9.693498452012385,
"grad_norm": 3.635230157595945,
"learning_rate": 8.975734603767056e-06,
"loss": 0.0402,
"step": 1570
},
{
"epoch": 9.69969040247678,
"grad_norm": 3.78300603635811,
"learning_rate": 8.92787480257286e-06,
"loss": 0.0414,
"step": 1571
},
{
"epoch": 9.705882352941176,
"grad_norm": 3.365315495479003,
"learning_rate": 8.880130428441774e-06,
"loss": 0.039,
"step": 1572
},
{
"epoch": 9.712074303405572,
"grad_norm": 3.974574854772795,
"learning_rate": 8.832501615552224e-06,
"loss": 0.0472,
"step": 1573
},
{
"epoch": 9.718266253869968,
"grad_norm": 3.452851500943343,
"learning_rate": 8.784988497757846e-06,
"loss": 0.0348,
"step": 1574
},
{
"epoch": 9.724458204334365,
"grad_norm": 3.062904518542008,
"learning_rate": 8.737591208587158e-06,
"loss": 0.0325,
"step": 1575
},
{
"epoch": 9.730650154798761,
"grad_norm": 3.6314811791043593,
"learning_rate": 8.690309881243148e-06,
"loss": 0.0418,
"step": 1576
},
{
"epoch": 9.736842105263158,
"grad_norm": 4.1744000045735214,
"learning_rate": 8.643144648602913e-06,
"loss": 0.0375,
"step": 1577
},
{
"epoch": 9.743034055727554,
"grad_norm": 3.256942957126009,
"learning_rate": 8.5960956432173e-06,
"loss": 0.0492,
"step": 1578
},
{
"epoch": 9.74922600619195,
"grad_norm": 4.165939797178315,
"learning_rate": 8.549162997310467e-06,
"loss": 0.0445,
"step": 1579
},
{
"epoch": 9.755417956656347,
"grad_norm": 3.0519982083557933,
"learning_rate": 8.502346842779618e-06,
"loss": 0.0376,
"step": 1580
},
{
"epoch": 9.761609907120743,
"grad_norm": 3.760682337684378,
"learning_rate": 8.455647311194537e-06,
"loss": 0.032,
"step": 1581
},
{
"epoch": 9.76780185758514,
"grad_norm": 2.756832052460076,
"learning_rate": 8.409064533797284e-06,
"loss": 0.0389,
"step": 1582
},
{
"epoch": 9.773993808049536,
"grad_norm": 4.340348325659682,
"learning_rate": 8.362598641501773e-06,
"loss": 0.0337,
"step": 1583
},
{
"epoch": 9.780185758513932,
"grad_norm": 2.7303960940322756,
"learning_rate": 8.31624976489347e-06,
"loss": 0.0492,
"step": 1584
},
{
"epoch": 9.786377708978328,
"grad_norm": 4.578121431627828,
"learning_rate": 8.270018034228916e-06,
"loss": 0.0389,
"step": 1585
},
{
"epoch": 9.792569659442725,
"grad_norm": 3.1414915838212494,
"learning_rate": 8.223903579435499e-06,
"loss": 0.0302,
"step": 1586
},
{
"epoch": 9.798761609907121,
"grad_norm": 3.773023692273092,
"learning_rate": 8.177906530110996e-06,
"loss": 0.0395,
"step": 1587
},
{
"epoch": 9.804953560371517,
"grad_norm": 2.8811396061099033,
"learning_rate": 8.132027015523219e-06,
"loss": 0.0371,
"step": 1588
},
{
"epoch": 9.811145510835914,
"grad_norm": 2.643361725427002,
"learning_rate": 8.086265164609708e-06,
"loss": 0.0426,
"step": 1589
},
{
"epoch": 9.81733746130031,
"grad_norm": 3.7025529170637914,
"learning_rate": 8.040621105977264e-06,
"loss": 0.0373,
"step": 1590
},
{
"epoch": 9.823529411764707,
"grad_norm": 3.6884225848392185,
"learning_rate": 7.9950949679017e-06,
"loss": 0.0386,
"step": 1591
},
{
"epoch": 9.829721362229103,
"grad_norm": 3.3916461921735164,
"learning_rate": 7.949686878327428e-06,
"loss": 0.0324,
"step": 1592
},
{
"epoch": 9.8359133126935,
"grad_norm": 2.8304230940358,
"learning_rate": 7.90439696486705e-06,
"loss": 0.0422,
"step": 1593
},
{
"epoch": 9.842105263157894,
"grad_norm": 3.0651290512485727,
"learning_rate": 7.859225354801137e-06,
"loss": 0.0368,
"step": 1594
},
{
"epoch": 9.848297213622292,
"grad_norm": 4.475021970118916,
"learning_rate": 7.814172175077738e-06,
"loss": 0.0456,
"step": 1595
},
{
"epoch": 9.854489164086687,
"grad_norm": 3.215831651434377,
"learning_rate": 7.769237552312048e-06,
"loss": 0.0339,
"step": 1596
},
{
"epoch": 9.860681114551083,
"grad_norm": 4.138337590415728,
"learning_rate": 7.724421612786109e-06,
"loss": 0.0308,
"step": 1597
},
{
"epoch": 9.86687306501548,
"grad_norm": 2.4632014500120865,
"learning_rate": 7.679724482448424e-06,
"loss": 0.0305,
"step": 1598
},
{
"epoch": 9.873065015479876,
"grad_norm": 2.6731726515248675,
"learning_rate": 7.635146286913586e-06,
"loss": 0.0539,
"step": 1599
},
{
"epoch": 9.879256965944272,
"grad_norm": 4.600988512774891,
"learning_rate": 7.59068715146195e-06,
"loss": 0.0279,
"step": 1600
},
{
"epoch": 9.885448916408668,
"grad_norm": 3.5572086266550556,
"learning_rate": 7.546347201039255e-06,
"loss": 0.041,
"step": 1601
},
{
"epoch": 9.891640866873065,
"grad_norm": 3.0537265767069863,
"learning_rate": 7.502126560256301e-06,
"loss": 0.0459,
"step": 1602
},
{
"epoch": 9.897832817337461,
"grad_norm": 3.5419809775843833,
"learning_rate": 7.458025353388593e-06,
"loss": 0.0334,
"step": 1603
},
{
"epoch": 9.904024767801857,
"grad_norm": 3.414732524788787,
"learning_rate": 7.414043704375944e-06,
"loss": 0.0489,
"step": 1604
},
{
"epoch": 9.910216718266254,
"grad_norm": 3.3256856595567967,
"learning_rate": 7.370181736822229e-06,
"loss": 0.0363,
"step": 1605
},
{
"epoch": 9.91640866873065,
"grad_norm": 3.895158991978571,
"learning_rate": 7.326439573994953e-06,
"loss": 0.0327,
"step": 1606
},
{
"epoch": 9.922600619195046,
"grad_norm": 2.6763490690043152,
"learning_rate": 7.282817338824893e-06,
"loss": 0.0314,
"step": 1607
},
{
"epoch": 9.928792569659443,
"grad_norm": 2.958990660663684,
"learning_rate": 7.2393151539058235e-06,
"loss": 0.0327,
"step": 1608
},
{
"epoch": 9.93498452012384,
"grad_norm": 3.6738392405654303,
"learning_rate": 7.195933141494133e-06,
"loss": 0.0367,
"step": 1609
},
{
"epoch": 9.941176470588236,
"grad_norm": 2.6964217016336183,
"learning_rate": 7.1526714235084725e-06,
"loss": 0.0328,
"step": 1610
},
{
"epoch": 9.947368421052632,
"grad_norm": 4.1020842351654085,
"learning_rate": 7.109530121529439e-06,
"loss": 0.0357,
"step": 1611
},
{
"epoch": 9.953560371517028,
"grad_norm": 4.013825911292633,
"learning_rate": 7.066509356799189e-06,
"loss": 0.045,
"step": 1612
},
{
"epoch": 9.959752321981425,
"grad_norm": 2.890908035100116,
"learning_rate": 7.023609250221153e-06,
"loss": 0.0383,
"step": 1613
},
{
"epoch": 9.965944272445821,
"grad_norm": 2.866464736264314,
"learning_rate": 6.980829922359666e-06,
"loss": 0.0316,
"step": 1614
},
{
"epoch": 9.972136222910217,
"grad_norm": 3.532104979673028,
"learning_rate": 6.938171493439622e-06,
"loss": 0.0321,
"step": 1615
},
{
"epoch": 9.978328173374614,
"grad_norm": 2.4739025667718315,
"learning_rate": 6.895634083346158e-06,
"loss": 0.0333,
"step": 1616
},
{
"epoch": 9.98452012383901,
"grad_norm": 2.686121620365639,
"learning_rate": 6.853217811624313e-06,
"loss": 0.0325,
"step": 1617
},
{
"epoch": 9.990712074303406,
"grad_norm": 3.1692020283541273,
"learning_rate": 6.810922797478653e-06,
"loss": 0.0441,
"step": 1618
},
{
"epoch": 9.996904024767801,
"grad_norm": 3.084011355069829,
"learning_rate": 6.7687491597729915e-06,
"loss": 0.0358,
"step": 1619
},
{
"epoch": 10.0,
"grad_norm": 3.492275686041303,
"learning_rate": 6.7266970170300315e-06,
"loss": 0.0115,
"step": 1620
},
{
"epoch": 10.006191950464396,
"grad_norm": 0.9690041621709453,
"learning_rate": 6.684766487431027e-06,
"loss": 0.0136,
"step": 1621
},
{
"epoch": 10.012383900928793,
"grad_norm": 0.6941026113485895,
"learning_rate": 6.642957688815476e-06,
"loss": 0.0105,
"step": 1622
},
{
"epoch": 10.018575851393189,
"grad_norm": 1.6892873132285695,
"learning_rate": 6.601270738680721e-06,
"loss": 0.0132,
"step": 1623
},
{
"epoch": 10.024767801857585,
"grad_norm": 0.7795458575261905,
"learning_rate": 6.559705754181711e-06,
"loss": 0.0117,
"step": 1624
},
{
"epoch": 10.030959752321982,
"grad_norm": 1.816508512253383,
"learning_rate": 6.518262852130625e-06,
"loss": 0.0168,
"step": 1625
},
{
"epoch": 10.037151702786378,
"grad_norm": 1.1419893923025035,
"learning_rate": 6.476942148996529e-06,
"loss": 0.0132,
"step": 1626
},
{
"epoch": 10.043343653250774,
"grad_norm": 0.8090659548366029,
"learning_rate": 6.435743760905083e-06,
"loss": 0.0095,
"step": 1627
},
{
"epoch": 10.04953560371517,
"grad_norm": 0.619249601903476,
"learning_rate": 6.394667803638199e-06,
"loss": 0.0093,
"step": 1628
},
{
"epoch": 10.055727554179567,
"grad_norm": 0.7657768966180624,
"learning_rate": 6.353714392633697e-06,
"loss": 0.0104,
"step": 1629
},
{
"epoch": 10.061919504643964,
"grad_norm": 0.8324925619707857,
"learning_rate": 6.312883642985013e-06,
"loss": 0.0117,
"step": 1630
},
{
"epoch": 10.06811145510836,
"grad_norm": 1.3470052156475658,
"learning_rate": 6.272175669440861e-06,
"loss": 0.0119,
"step": 1631
},
{
"epoch": 10.074303405572756,
"grad_norm": 1.1332209729806688,
"learning_rate": 6.2315905864049175e-06,
"loss": 0.011,
"step": 1632
},
{
"epoch": 10.08049535603715,
"grad_norm": 0.8623305294648557,
"learning_rate": 6.1911285079354785e-06,
"loss": 0.01,
"step": 1633
},
{
"epoch": 10.086687306501547,
"grad_norm": 1.9984087842934972,
"learning_rate": 6.150789547745178e-06,
"loss": 0.0137,
"step": 1634
},
{
"epoch": 10.092879256965944,
"grad_norm": 0.797048668269128,
"learning_rate": 6.110573819200605e-06,
"loss": 0.0123,
"step": 1635
},
{
"epoch": 10.09907120743034,
"grad_norm": 1.6465166602549526,
"learning_rate": 6.070481435322062e-06,
"loss": 0.0174,
"step": 1636
},
{
"epoch": 10.105263157894736,
"grad_norm": 0.9915252884551151,
"learning_rate": 6.030512508783187e-06,
"loss": 0.0117,
"step": 1637
},
{
"epoch": 10.111455108359133,
"grad_norm": 0.8757703472508664,
"learning_rate": 5.990667151910678e-06,
"loss": 0.0123,
"step": 1638
},
{
"epoch": 10.117647058823529,
"grad_norm": 0.6976670441674919,
"learning_rate": 5.950945476683956e-06,
"loss": 0.0102,
"step": 1639
},
{
"epoch": 10.123839009287925,
"grad_norm": 1.671517191093078,
"learning_rate": 5.911347594734823e-06,
"loss": 0.0139,
"step": 1640
},
{
"epoch": 10.130030959752322,
"grad_norm": 1.9299239166858515,
"learning_rate": 5.871873617347217e-06,
"loss": 0.0133,
"step": 1641
},
{
"epoch": 10.136222910216718,
"grad_norm": 1.2988206835620129,
"learning_rate": 5.832523655456845e-06,
"loss": 0.0153,
"step": 1642
},
{
"epoch": 10.142414860681114,
"grad_norm": 1.4569247038428583,
"learning_rate": 5.793297819650884e-06,
"loss": 0.0127,
"step": 1643
},
{
"epoch": 10.14860681114551,
"grad_norm": 0.7816795380471435,
"learning_rate": 5.754196220167679e-06,
"loss": 0.0098,
"step": 1644
},
{
"epoch": 10.154798761609907,
"grad_norm": 1.3591968793264988,
"learning_rate": 5.715218966896435e-06,
"loss": 0.0133,
"step": 1645
},
{
"epoch": 10.160990712074303,
"grad_norm": 0.6915571090654116,
"learning_rate": 5.676366169376873e-06,
"loss": 0.0109,
"step": 1646
},
{
"epoch": 10.1671826625387,
"grad_norm": 1.1901382752213034,
"learning_rate": 5.637637936798979e-06,
"loss": 0.0105,
"step": 1647
},
{
"epoch": 10.173374613003096,
"grad_norm": 0.697074801951405,
"learning_rate": 5.599034378002649e-06,
"loss": 0.0105,
"step": 1648
},
{
"epoch": 10.179566563467493,
"grad_norm": 1.5164025066944178,
"learning_rate": 5.560555601477418e-06,
"loss": 0.014,
"step": 1649
},
{
"epoch": 10.185758513931889,
"grad_norm": 0.9434268887247322,
"learning_rate": 5.522201715362135e-06,
"loss": 0.0092,
"step": 1650
},
{
"epoch": 10.191950464396285,
"grad_norm": 0.851637419723935,
"learning_rate": 5.483972827444644e-06,
"loss": 0.0101,
"step": 1651
},
{
"epoch": 10.198142414860682,
"grad_norm": 0.5702932276533784,
"learning_rate": 5.4458690451615225e-06,
"loss": 0.0073,
"step": 1652
},
{
"epoch": 10.204334365325078,
"grad_norm": 1.0316361049272975,
"learning_rate": 5.40789047559776e-06,
"loss": 0.0104,
"step": 1653
},
{
"epoch": 10.210526315789474,
"grad_norm": 1.5701282669378862,
"learning_rate": 5.370037225486413e-06,
"loss": 0.0138,
"step": 1654
},
{
"epoch": 10.21671826625387,
"grad_norm": 1.614278502654183,
"learning_rate": 5.332309401208408e-06,
"loss": 0.0109,
"step": 1655
},
{
"epoch": 10.222910216718267,
"grad_norm": 1.5812343173850434,
"learning_rate": 5.294707108792146e-06,
"loss": 0.0115,
"step": 1656
},
{
"epoch": 10.229102167182663,
"grad_norm": 0.7741625900478416,
"learning_rate": 5.257230453913237e-06,
"loss": 0.0087,
"step": 1657
},
{
"epoch": 10.235294117647058,
"grad_norm": 0.8181768770881074,
"learning_rate": 5.219879541894213e-06,
"loss": 0.0114,
"step": 1658
},
{
"epoch": 10.241486068111454,
"grad_norm": 0.5613606941784166,
"learning_rate": 5.182654477704229e-06,
"loss": 0.0068,
"step": 1659
},
{
"epoch": 10.24767801857585,
"grad_norm": 1.4465625168944913,
"learning_rate": 5.145555365958754e-06,
"loss": 0.0129,
"step": 1660
},
{
"epoch": 10.253869969040247,
"grad_norm": 1.264523084436169,
"learning_rate": 5.108582310919302e-06,
"loss": 0.0115,
"step": 1661
},
{
"epoch": 10.260061919504643,
"grad_norm": 1.5695424386876935,
"learning_rate": 5.071735416493095e-06,
"loss": 0.014,
"step": 1662
},
{
"epoch": 10.26625386996904,
"grad_norm": 1.5234094429714513,
"learning_rate": 5.0350147862328275e-06,
"loss": 0.0118,
"step": 1663
},
{
"epoch": 10.272445820433436,
"grad_norm": 1.1168371246093773,
"learning_rate": 4.998420523336344e-06,
"loss": 0.0081,
"step": 1664
},
{
"epoch": 10.278637770897832,
"grad_norm": 0.7440344852931872,
"learning_rate": 4.961952730646319e-06,
"loss": 0.0086,
"step": 1665
},
{
"epoch": 10.284829721362229,
"grad_norm": 1.8493349338190865,
"learning_rate": 4.9256115106500575e-06,
"loss": 0.0149,
"step": 1666
},
{
"epoch": 10.291021671826625,
"grad_norm": 1.3877650925456666,
"learning_rate": 4.889396965479115e-06,
"loss": 0.0114,
"step": 1667
},
{
"epoch": 10.297213622291022,
"grad_norm": 0.7281950423300454,
"learning_rate": 4.853309196909045e-06,
"loss": 0.0096,
"step": 1668
},
{
"epoch": 10.303405572755418,
"grad_norm": 2.8181618879647354,
"learning_rate": 4.817348306359121e-06,
"loss": 0.0123,
"step": 1669
},
{
"epoch": 10.309597523219814,
"grad_norm": 0.5273928142352734,
"learning_rate": 4.78151439489205e-06,
"loss": 0.0075,
"step": 1670
},
{
"epoch": 10.31578947368421,
"grad_norm": 1.2995849665706596,
"learning_rate": 4.7458075632136776e-06,
"loss": 0.0116,
"step": 1671
},
{
"epoch": 10.321981424148607,
"grad_norm": 1.5898543976941963,
"learning_rate": 4.710227911672721e-06,
"loss": 0.013,
"step": 1672
},
{
"epoch": 10.328173374613003,
"grad_norm": 1.3417056303557735,
"learning_rate": 4.674775540260456e-06,
"loss": 0.0103,
"step": 1673
},
{
"epoch": 10.3343653250774,
"grad_norm": 1.4685272029621523,
"learning_rate": 4.6394505486104685e-06,
"loss": 0.0131,
"step": 1674
},
{
"epoch": 10.340557275541796,
"grad_norm": 0.8277822917699998,
"learning_rate": 4.604253035998379e-06,
"loss": 0.0116,
"step": 1675
},
{
"epoch": 10.346749226006192,
"grad_norm": 1.1121863929552394,
"learning_rate": 4.569183101341501e-06,
"loss": 0.0114,
"step": 1676
},
{
"epoch": 10.352941176470589,
"grad_norm": 0.677948454774533,
"learning_rate": 4.534240843198662e-06,
"loss": 0.0091,
"step": 1677
},
{
"epoch": 10.359133126934985,
"grad_norm": 0.8533445134270156,
"learning_rate": 4.499426359769859e-06,
"loss": 0.0095,
"step": 1678
},
{
"epoch": 10.365325077399381,
"grad_norm": 1.0518039551642777,
"learning_rate": 4.464739748895963e-06,
"loss": 0.0104,
"step": 1679
},
{
"epoch": 10.371517027863778,
"grad_norm": 1.4222729965189793,
"learning_rate": 4.430181108058517e-06,
"loss": 0.0116,
"step": 1680
},
{
"epoch": 10.377708978328174,
"grad_norm": 0.7831784087988597,
"learning_rate": 4.395750534379411e-06,
"loss": 0.0088,
"step": 1681
},
{
"epoch": 10.38390092879257,
"grad_norm": 0.7708963964507719,
"learning_rate": 4.361448124620621e-06,
"loss": 0.0097,
"step": 1682
},
{
"epoch": 10.390092879256965,
"grad_norm": 0.8954947088275224,
"learning_rate": 4.327273975183949e-06,
"loss": 0.0112,
"step": 1683
},
{
"epoch": 10.396284829721361,
"grad_norm": 1.8242329830813873,
"learning_rate": 4.293228182110704e-06,
"loss": 0.0099,
"step": 1684
},
{
"epoch": 10.402476780185758,
"grad_norm": 3.6333512677727655,
"learning_rate": 4.259310841081515e-06,
"loss": 0.0204,
"step": 1685
},
{
"epoch": 10.408668730650154,
"grad_norm": 1.892780654876686,
"learning_rate": 4.225522047415992e-06,
"loss": 0.0189,
"step": 1686
},
{
"epoch": 10.41486068111455,
"grad_norm": 1.7700319894140126,
"learning_rate": 4.191861896072457e-06,
"loss": 0.0118,
"step": 1687
},
{
"epoch": 10.421052631578947,
"grad_norm": 0.9200528055389516,
"learning_rate": 4.1583304816477706e-06,
"loss": 0.0125,
"step": 1688
},
{
"epoch": 10.427244582043343,
"grad_norm": 0.7789740816150398,
"learning_rate": 4.12492789837694e-06,
"loss": 0.0109,
"step": 1689
},
{
"epoch": 10.43343653250774,
"grad_norm": 0.783564421561771,
"learning_rate": 4.091654240132925e-06,
"loss": 0.0116,
"step": 1690
},
{
"epoch": 10.439628482972136,
"grad_norm": 0.6118045323063145,
"learning_rate": 4.058509600426358e-06,
"loss": 0.0077,
"step": 1691
},
{
"epoch": 10.445820433436532,
"grad_norm": 0.7260828388384113,
"learning_rate": 4.0254940724053005e-06,
"loss": 0.0093,
"step": 1692
},
{
"epoch": 10.452012383900929,
"grad_norm": 2.228638075154875,
"learning_rate": 3.992607748854954e-06,
"loss": 0.0181,
"step": 1693
},
{
"epoch": 10.458204334365325,
"grad_norm": 1.0717379632301736,
"learning_rate": 3.959850722197411e-06,
"loss": 0.0098,
"step": 1694
},
{
"epoch": 10.464396284829721,
"grad_norm": 1.3420482541310115,
"learning_rate": 3.927223084491388e-06,
"loss": 0.0117,
"step": 1695
},
{
"epoch": 10.470588235294118,
"grad_norm": 0.6950594630885271,
"learning_rate": 3.8947249274319805e-06,
"loss": 0.0076,
"step": 1696
},
{
"epoch": 10.476780185758514,
"grad_norm": 1.702546172738537,
"learning_rate": 3.86235634235041e-06,
"loss": 0.0126,
"step": 1697
},
{
"epoch": 10.48297213622291,
"grad_norm": 0.9719819269381079,
"learning_rate": 3.830117420213713e-06,
"loss": 0.0089,
"step": 1698
},
{
"epoch": 10.489164086687307,
"grad_norm": 1.4004291750343034,
"learning_rate": 3.798008251624585e-06,
"loss": 0.0132,
"step": 1699
},
{
"epoch": 10.495356037151703,
"grad_norm": 0.7473199645427587,
"learning_rate": 3.7660289268210415e-06,
"loss": 0.008,
"step": 1700
},
{
"epoch": 10.5015479876161,
"grad_norm": 0.9275640876435634,
"learning_rate": 3.7341795356761688e-06,
"loss": 0.0091,
"step": 1701
},
{
"epoch": 10.507739938080496,
"grad_norm": 0.6574915351770422,
"learning_rate": 3.70246016769793e-06,
"loss": 0.0079,
"step": 1702
},
{
"epoch": 10.513931888544892,
"grad_norm": 0.7561543389827127,
"learning_rate": 3.6708709120288565e-06,
"loss": 0.0087,
"step": 1703
},
{
"epoch": 10.520123839009289,
"grad_norm": 0.9256024134932721,
"learning_rate": 3.639411857445829e-06,
"loss": 0.012,
"step": 1704
},
{
"epoch": 10.526315789473685,
"grad_norm": 1.3251515213309393,
"learning_rate": 3.6080830923598263e-06,
"loss": 0.0106,
"step": 1705
},
{
"epoch": 10.532507739938081,
"grad_norm": 0.804076898474573,
"learning_rate": 3.5768847048156305e-06,
"loss": 0.0076,
"step": 1706
},
{
"epoch": 10.538699690402478,
"grad_norm": 0.5449255097242203,
"learning_rate": 3.5458167824916566e-06,
"loss": 0.0076,
"step": 1707
},
{
"epoch": 10.544891640866872,
"grad_norm": 0.8837706768356558,
"learning_rate": 3.5148794126996563e-06,
"loss": 0.0112,
"step": 1708
},
{
"epoch": 10.551083591331269,
"grad_norm": 0.7222977773251738,
"learning_rate": 3.484072682384465e-06,
"loss": 0.0085,
"step": 1709
},
{
"epoch": 10.557275541795665,
"grad_norm": 1.0329452823258398,
"learning_rate": 3.4533966781237992e-06,
"loss": 0.0118,
"step": 1710
},
{
"epoch": 10.563467492260061,
"grad_norm": 0.9527789813475251,
"learning_rate": 3.422851486127987e-06,
"loss": 0.01,
"step": 1711
},
{
"epoch": 10.569659442724458,
"grad_norm": 0.8268111323007847,
"learning_rate": 3.3924371922397003e-06,
"loss": 0.0112,
"step": 1712
},
{
"epoch": 10.575851393188854,
"grad_norm": 0.6727829069778749,
"learning_rate": 3.3621538819337772e-06,
"loss": 0.0102,
"step": 1713
},
{
"epoch": 10.58204334365325,
"grad_norm": 0.6973365017429423,
"learning_rate": 3.332001640316923e-06,
"loss": 0.0085,
"step": 1714
},
{
"epoch": 10.588235294117647,
"grad_norm": 1.2335275863619866,
"learning_rate": 3.3019805521275095e-06,
"loss": 0.0119,
"step": 1715
},
{
"epoch": 10.594427244582043,
"grad_norm": 0.8442287627038453,
"learning_rate": 3.272090701735314e-06,
"loss": 0.0109,
"step": 1716
},
{
"epoch": 10.60061919504644,
"grad_norm": 0.7037863282128368,
"learning_rate": 3.242332173141277e-06,
"loss": 0.0077,
"step": 1717
},
{
"epoch": 10.606811145510836,
"grad_norm": 2.0129171525647354,
"learning_rate": 3.212705049977299e-06,
"loss": 0.0129,
"step": 1718
},
{
"epoch": 10.613003095975232,
"grad_norm": 1.6794016066259494,
"learning_rate": 3.183209415505978e-06,
"loss": 0.0158,
"step": 1719
},
{
"epoch": 10.619195046439629,
"grad_norm": 0.8047045412104622,
"learning_rate": 3.153845352620377e-06,
"loss": 0.0098,
"step": 1720
},
{
"epoch": 10.625386996904025,
"grad_norm": 0.8135051004087362,
"learning_rate": 3.1246129438438076e-06,
"loss": 0.01,
"step": 1721
},
{
"epoch": 10.631578947368421,
"grad_norm": 1.169581392676557,
"learning_rate": 3.095512271329587e-06,
"loss": 0.0099,
"step": 1722
},
{
"epoch": 10.637770897832818,
"grad_norm": 0.7159938157722547,
"learning_rate": 3.0665434168607842e-06,
"loss": 0.0084,
"step": 1723
},
{
"epoch": 10.643962848297214,
"grad_norm": 1.5038994222527753,
"learning_rate": 3.0377064618500426e-06,
"loss": 0.0123,
"step": 1724
},
{
"epoch": 10.65015479876161,
"grad_norm": 0.9341949537858699,
"learning_rate": 3.009001487339308e-06,
"loss": 0.0094,
"step": 1725
},
{
"epoch": 10.656346749226007,
"grad_norm": 1.059506028399342,
"learning_rate": 2.9804285739996053e-06,
"loss": 0.0091,
"step": 1726
},
{
"epoch": 10.662538699690403,
"grad_norm": 0.4376975526720586,
"learning_rate": 2.9519878021308622e-06,
"loss": 0.0059,
"step": 1727
},
{
"epoch": 10.6687306501548,
"grad_norm": 0.5828319926747679,
"learning_rate": 2.9236792516615817e-06,
"loss": 0.0077,
"step": 1728
},
{
"epoch": 10.674922600619196,
"grad_norm": 0.7758125938963785,
"learning_rate": 2.895503002148725e-06,
"loss": 0.0107,
"step": 1729
},
{
"epoch": 10.681114551083592,
"grad_norm": 0.9203794960207291,
"learning_rate": 2.867459132777417e-06,
"loss": 0.0103,
"step": 1730
},
{
"epoch": 10.687306501547988,
"grad_norm": 0.8585940595715308,
"learning_rate": 2.8395477223607693e-06,
"loss": 0.0086,
"step": 1731
},
{
"epoch": 10.693498452012385,
"grad_norm": 0.7370337401926742,
"learning_rate": 2.811768849339619e-06,
"loss": 0.0095,
"step": 1732
},
{
"epoch": 10.69969040247678,
"grad_norm": 0.7979822392217546,
"learning_rate": 2.7841225917823345e-06,
"loss": 0.0094,
"step": 1733
},
{
"epoch": 10.705882352941176,
"grad_norm": 1.0772577268099952,
"learning_rate": 2.7566090273845824e-06,
"loss": 0.0101,
"step": 1734
},
{
"epoch": 10.712074303405572,
"grad_norm": 0.8668097368420958,
"learning_rate": 2.729228233469117e-06,
"loss": 0.0106,
"step": 1735
},
{
"epoch": 10.718266253869968,
"grad_norm": 0.6870847423896851,
"learning_rate": 2.7019802869855783e-06,
"loss": 0.009,
"step": 1736
},
{
"epoch": 10.724458204334365,
"grad_norm": 0.7982906139809085,
"learning_rate": 2.6748652645102177e-06,
"loss": 0.0112,
"step": 1737
},
{
"epoch": 10.730650154798761,
"grad_norm": 0.7354991019101021,
"learning_rate": 2.647883242245769e-06,
"loss": 0.0093,
"step": 1738
},
{
"epoch": 10.736842105263158,
"grad_norm": 0.9678810687189697,
"learning_rate": 2.6210342960211744e-06,
"loss": 0.0091,
"step": 1739
},
{
"epoch": 10.743034055727554,
"grad_norm": 0.6660482019733218,
"learning_rate": 2.5943185012913594e-06,
"loss": 0.0095,
"step": 1740
},
{
"epoch": 10.74922600619195,
"grad_norm": 1.06839550994224,
"learning_rate": 2.567735933137083e-06,
"loss": 0.0088,
"step": 1741
},
{
"epoch": 10.755417956656347,
"grad_norm": 0.9426678760153843,
"learning_rate": 2.5412866662646697e-06,
"loss": 0.008,
"step": 1742
},
{
"epoch": 10.761609907120743,
"grad_norm": 0.7572329345990066,
"learning_rate": 2.5149707750058314e-06,
"loss": 0.0093,
"step": 1743
},
{
"epoch": 10.76780185758514,
"grad_norm": 1.1459689282860193,
"learning_rate": 2.4887883333174435e-06,
"loss": 0.0102,
"step": 1744
},
{
"epoch": 10.773993808049536,
"grad_norm": 0.7313816351516814,
"learning_rate": 2.4627394147813343e-06,
"loss": 0.0073,
"step": 1745
},
{
"epoch": 10.780185758513932,
"grad_norm": 1.207241558988607,
"learning_rate": 2.4368240926041008e-06,
"loss": 0.0087,
"step": 1746
},
{
"epoch": 10.786377708978328,
"grad_norm": 0.7977655013840111,
"learning_rate": 2.411042439616873e-06,
"loss": 0.0078,
"step": 1747
},
{
"epoch": 10.792569659442725,
"grad_norm": 0.890577896345556,
"learning_rate": 2.3853945282751257e-06,
"loss": 0.0113,
"step": 1748
},
{
"epoch": 10.798761609907121,
"grad_norm": 0.8583291783885854,
"learning_rate": 2.3598804306584843e-06,
"loss": 0.0092,
"step": 1749
},
{
"epoch": 10.804953560371517,
"grad_norm": 1.0913490870442912,
"learning_rate": 2.334500218470509e-06,
"loss": 0.0114,
"step": 1750
},
{
"epoch": 10.811145510835914,
"grad_norm": 1.0439131308846614,
"learning_rate": 2.309253963038477e-06,
"loss": 0.0101,
"step": 1751
},
{
"epoch": 10.81733746130031,
"grad_norm": 0.8933322010979203,
"learning_rate": 2.284141735313211e-06,
"loss": 0.0096,
"step": 1752
},
{
"epoch": 10.823529411764707,
"grad_norm": 0.7395638861766168,
"learning_rate": 2.2591636058688804e-06,
"loss": 0.0084,
"step": 1753
},
{
"epoch": 10.829721362229103,
"grad_norm": 0.9780330229569952,
"learning_rate": 2.2343196449027716e-06,
"loss": 0.0126,
"step": 1754
},
{
"epoch": 10.8359133126935,
"grad_norm": 0.952534486696127,
"learning_rate": 2.209609922235134e-06,
"loss": 0.0114,
"step": 1755
},
{
"epoch": 10.842105263157894,
"grad_norm": 1.0367024109289174,
"learning_rate": 2.185034507308925e-06,
"loss": 0.0081,
"step": 1756
},
{
"epoch": 10.848297213622292,
"grad_norm": 1.4665715967848132,
"learning_rate": 2.1605934691896867e-06,
"loss": 0.0123,
"step": 1757
},
{
"epoch": 10.854489164086687,
"grad_norm": 0.4968073457268509,
"learning_rate": 2.136286876565302e-06,
"loss": 0.0066,
"step": 1758
},
{
"epoch": 10.860681114551083,
"grad_norm": 0.7006736750841475,
"learning_rate": 2.1121147977457954e-06,
"loss": 0.0079,
"step": 1759
},
{
"epoch": 10.86687306501548,
"grad_norm": 0.9662755200351817,
"learning_rate": 2.0880773006631935e-06,
"loss": 0.0092,
"step": 1760
},
{
"epoch": 10.873065015479876,
"grad_norm": 0.639567274002592,
"learning_rate": 2.0641744528712925e-06,
"loss": 0.0076,
"step": 1761
},
{
"epoch": 10.879256965944272,
"grad_norm": 0.6402348827422334,
"learning_rate": 2.0404063215454515e-06,
"loss": 0.0063,
"step": 1762
},
{
"epoch": 10.885448916408668,
"grad_norm": 1.4081204793940463,
"learning_rate": 2.016772973482456e-06,
"loss": 0.0123,
"step": 1763
},
{
"epoch": 10.891640866873065,
"grad_norm": 0.6397554058823934,
"learning_rate": 1.993274475100293e-06,
"loss": 0.0079,
"step": 1764
},
{
"epoch": 10.897832817337461,
"grad_norm": 0.8403423770212202,
"learning_rate": 1.9699108924379818e-06,
"loss": 0.0092,
"step": 1765
},
{
"epoch": 10.904024767801857,
"grad_norm": 0.700545288483362,
"learning_rate": 1.9466822911553774e-06,
"loss": 0.0084,
"step": 1766
},
{
"epoch": 10.910216718266254,
"grad_norm": 0.8071855492769477,
"learning_rate": 1.9235887365329776e-06,
"loss": 0.0078,
"step": 1767
},
{
"epoch": 10.91640866873065,
"grad_norm": 0.9467118760989457,
"learning_rate": 1.9006302934717723e-06,
"loss": 0.0073,
"step": 1768
},
{
"epoch": 10.922600619195046,
"grad_norm": 0.9326749378450487,
"learning_rate": 1.8778070264930281e-06,
"loss": 0.0115,
"step": 1769
},
{
"epoch": 10.928792569659443,
"grad_norm": 0.9362248437621925,
"learning_rate": 1.8551189997381147e-06,
"loss": 0.0091,
"step": 1770
},
{
"epoch": 10.93498452012384,
"grad_norm": 0.7116804801712812,
"learning_rate": 1.832566276968345e-06,
"loss": 0.0088,
"step": 1771
},
{
"epoch": 10.941176470588236,
"grad_norm": 1.1736566878259078,
"learning_rate": 1.8101489215647803e-06,
"loss": 0.0123,
"step": 1772
},
{
"epoch": 10.947368421052632,
"grad_norm": 1.3595920924818499,
"learning_rate": 1.7878669965280314e-06,
"loss": 0.0106,
"step": 1773
},
{
"epoch": 10.953560371517028,
"grad_norm": 0.6260934080310849,
"learning_rate": 1.7657205644781128e-06,
"loss": 0.0074,
"step": 1774
},
{
"epoch": 10.959752321981425,
"grad_norm": 1.1283883512215795,
"learning_rate": 1.7437096876542713e-06,
"loss": 0.0094,
"step": 1775
},
{
"epoch": 10.965944272445821,
"grad_norm": 1.0216294157436547,
"learning_rate": 1.7218344279147702e-06,
"loss": 0.0098,
"step": 1776
},
{
"epoch": 10.972136222910217,
"grad_norm": 1.574604863160782,
"learning_rate": 1.7000948467367717e-06,
"loss": 0.011,
"step": 1777
},
{
"epoch": 10.978328173374614,
"grad_norm": 0.6374174964157301,
"learning_rate": 1.678491005216093e-06,
"loss": 0.0077,
"step": 1778
},
{
"epoch": 10.98452012383901,
"grad_norm": 0.6088874624769877,
"learning_rate": 1.657022964067112e-06,
"loss": 0.007,
"step": 1779
},
{
"epoch": 10.990712074303406,
"grad_norm": 1.0079770500653065,
"learning_rate": 1.6356907836225565e-06,
"loss": 0.0117,
"step": 1780
},
{
"epoch": 10.996904024767801,
"grad_norm": 0.5552932658491616,
"learning_rate": 1.6144945238332986e-06,
"loss": 0.0066,
"step": 1781
},
{
"epoch": 11.0,
"grad_norm": 0.5552932658491616,
"learning_rate": 1.5934342442682826e-06,
"loss": 0.0035,
"step": 1782
},
{
"epoch": 11.006191950464396,
"grad_norm": 0.8590230780171667,
"learning_rate": 1.5725100041142692e-06,
"loss": 0.0081,
"step": 1783
},
{
"epoch": 11.012383900928793,
"grad_norm": 1.9743623604731195,
"learning_rate": 1.5517218621756924e-06,
"loss": 0.0084,
"step": 1784
},
{
"epoch": 11.018575851393189,
"grad_norm": 0.35183110478447976,
"learning_rate": 1.5310698768745247e-06,
"loss": 0.0051,
"step": 1785
},
{
"epoch": 11.024767801857585,
"grad_norm": 0.6042641844577623,
"learning_rate": 1.5105541062500839e-06,
"loss": 0.0066,
"step": 1786
},
{
"epoch": 11.030959752321982,
"grad_norm": 0.5993920654526943,
"learning_rate": 1.4901746079588553e-06,
"loss": 0.0094,
"step": 1787
},
{
"epoch": 11.037151702786378,
"grad_norm": 0.54204549668726,
"learning_rate": 1.4699314392743969e-06,
"loss": 0.0068,
"step": 1788
},
{
"epoch": 11.043343653250774,
"grad_norm": 0.39963210647239444,
"learning_rate": 1.4498246570870844e-06,
"loss": 0.0072,
"step": 1789
},
{
"epoch": 11.04953560371517,
"grad_norm": 0.9314148398330484,
"learning_rate": 1.4298543179040337e-06,
"loss": 0.0062,
"step": 1790
},
{
"epoch": 11.055727554179567,
"grad_norm": 0.3696579757438963,
"learning_rate": 1.4100204778488946e-06,
"loss": 0.0074,
"step": 1791
},
{
"epoch": 11.061919504643964,
"grad_norm": 0.7288361787060351,
"learning_rate": 1.390323192661691e-06,
"loss": 0.0087,
"step": 1792
},
{
"epoch": 11.06811145510836,
"grad_norm": 0.581066884072162,
"learning_rate": 1.3707625176987149e-06,
"loss": 0.0069,
"step": 1793
},
{
"epoch": 11.074303405572756,
"grad_norm": 0.5857256793638341,
"learning_rate": 1.3513385079323094e-06,
"loss": 0.0076,
"step": 1794
},
{
"epoch": 11.08049535603715,
"grad_norm": 0.5199433056180636,
"learning_rate": 1.3320512179507526e-06,
"loss": 0.0064,
"step": 1795
},
{
"epoch": 11.086687306501547,
"grad_norm": 0.6046228576130303,
"learning_rate": 1.3129007019580752e-06,
"loss": 0.0087,
"step": 1796
},
{
"epoch": 11.092879256965944,
"grad_norm": 0.581548624238609,
"learning_rate": 1.293887013773959e-06,
"loss": 0.0078,
"step": 1797
},
{
"epoch": 11.09907120743034,
"grad_norm": 0.5301680741343029,
"learning_rate": 1.275010206833499e-06,
"loss": 0.0079,
"step": 1798
},
{
"epoch": 11.105263157894736,
"grad_norm": 0.42022687774794415,
"learning_rate": 1.2562703341871707e-06,
"loss": 0.0061,
"step": 1799
},
{
"epoch": 11.111455108359133,
"grad_norm": 0.665077350032754,
"learning_rate": 1.2376674485005568e-06,
"loss": 0.0072,
"step": 1800
},
{
"epoch": 11.117647058823529,
"grad_norm": 0.5579968541747495,
"learning_rate": 1.2192016020542984e-06,
"loss": 0.0084,
"step": 1801
},
{
"epoch": 11.123839009287925,
"grad_norm": 0.4954904144880409,
"learning_rate": 1.2008728467438946e-06,
"loss": 0.0056,
"step": 1802
},
{
"epoch": 11.130030959752322,
"grad_norm": 0.44858005274599605,
"learning_rate": 1.1826812340795523e-06,
"loss": 0.009,
"step": 1803
},
{
"epoch": 11.136222910216718,
"grad_norm": 0.804259291040737,
"learning_rate": 1.1646268151860984e-06,
"loss": 0.0067,
"step": 1804
},
{
"epoch": 11.142414860681114,
"grad_norm": 0.6782383183205587,
"learning_rate": 1.1467096408027678e-06,
"loss": 0.0064,
"step": 1805
},
{
"epoch": 11.14860681114551,
"grad_norm": 0.42505724970819314,
"learning_rate": 1.1289297612830985e-06,
"loss": 0.0066,
"step": 1806
},
{
"epoch": 11.154798761609907,
"grad_norm": 0.5140160581787362,
"learning_rate": 1.1112872265947815e-06,
"loss": 0.0055,
"step": 1807
},
{
"epoch": 11.160990712074303,
"grad_norm": 0.43576435472775915,
"learning_rate": 1.0937820863195225e-06,
"loss": 0.007,
"step": 1808
},
{
"epoch": 11.1671826625387,
"grad_norm": 0.593651067275537,
"learning_rate": 1.0764143896528966e-06,
"loss": 0.0084,
"step": 1809
},
{
"epoch": 11.173374613003096,
"grad_norm": 0.4183258589178314,
"learning_rate": 1.0591841854042218e-06,
"loss": 0.0062,
"step": 1810
},
{
"epoch": 11.179566563467493,
"grad_norm": 0.6129626912513965,
"learning_rate": 1.0420915219964023e-06,
"loss": 0.0056,
"step": 1811
},
{
"epoch": 11.185758513931889,
"grad_norm": 0.7988572762895868,
"learning_rate": 1.0251364474658187e-06,
"loss": 0.0067,
"step": 1812
},
{
"epoch": 11.191950464396285,
"grad_norm": 0.7057642530613543,
"learning_rate": 1.0083190094621719e-06,
"loss": 0.0073,
"step": 1813
},
{
"epoch": 11.198142414860682,
"grad_norm": 0.42926277169282456,
"learning_rate": 9.91639255248339e-07,
"loss": 0.0055,
"step": 1814
},
{
"epoch": 11.204334365325078,
"grad_norm": 0.41431734580291646,
"learning_rate": 9.750972317002949e-07,
"loss": 0.0065,
"step": 1815
},
{
"epoch": 11.210526315789474,
"grad_norm": 0.5919073415860816,
"learning_rate": 9.586929853069138e-07,
"loss": 0.0091,
"step": 1816
},
{
"epoch": 11.21671826625387,
"grad_norm": 0.8579825741603267,
"learning_rate": 9.424265621698736e-07,
"loss": 0.0077,
"step": 1817
},
{
"epoch": 11.222910216718267,
"grad_norm": 0.3910087337186186,
"learning_rate": 9.262980080035288e-07,
"loss": 0.0086,
"step": 1818
},
{
"epoch": 11.229102167182663,
"grad_norm": 0.5442690307510305,
"learning_rate": 9.103073681347607e-07,
"loss": 0.0059,
"step": 1819
},
{
"epoch": 11.235294117647058,
"grad_norm": 0.4656439315898283,
"learning_rate": 8.944546875028714e-07,
"loss": 0.0074,
"step": 1820
},
{
"epoch": 11.241486068111454,
"grad_norm": 0.5839593061344903,
"learning_rate": 8.787400106594568e-07,
"loss": 0.0073,
"step": 1821
},
{
"epoch": 11.24767801857585,
"grad_norm": 0.5211620061867827,
"learning_rate": 8.631633817682505e-07,
"loss": 0.008,
"step": 1822
},
{
"epoch": 11.253869969040247,
"grad_norm": 0.565491467227642,
"learning_rate": 8.477248446050523e-07,
"loss": 0.0074,
"step": 1823
},
{
"epoch": 11.260061919504643,
"grad_norm": 0.6939557316334681,
"learning_rate": 8.3242444255755e-07,
"loss": 0.0067,
"step": 1824
},
{
"epoch": 11.26625386996904,
"grad_norm": 0.4705178104025061,
"learning_rate": 8.172622186252421e-07,
"loss": 0.005,
"step": 1825
},
{
"epoch": 11.272445820433436,
"grad_norm": 0.34070656855140247,
"learning_rate": 8.02238215419282e-07,
"loss": 0.0062,
"step": 1826
},
{
"epoch": 11.278637770897832,
"grad_norm": 0.4739816540016086,
"learning_rate": 7.873524751624006e-07,
"loss": 0.0071,
"step": 1827
},
{
"epoch": 11.284829721362229,
"grad_norm": 0.5132177212112234,
"learning_rate": 7.72605039688723e-07,
"loss": 0.006,
"step": 1828
},
{
"epoch": 11.291021671826625,
"grad_norm": 0.39156911709743286,
"learning_rate": 7.579959504437184e-07,
"loss": 0.0062,
"step": 1829
},
{
"epoch": 11.297213622291022,
"grad_norm": 0.43103836317419303,
"learning_rate": 7.435252484840505e-07,
"loss": 0.0058,
"step": 1830
},
{
"epoch": 11.303405572755418,
"grad_norm": 0.5129600013904935,
"learning_rate": 7.291929744774495e-07,
"loss": 0.0071,
"step": 1831
},
{
"epoch": 11.309597523219814,
"grad_norm": 0.640816840030639,
"learning_rate": 7.149991687026514e-07,
"loss": 0.0083,
"step": 1832
},
{
"epoch": 11.31578947368421,
"grad_norm": 0.7358989372167142,
"learning_rate": 7.009438710491978e-07,
"loss": 0.0066,
"step": 1833
},
{
"epoch": 11.321981424148607,
"grad_norm": 0.5454191148891866,
"learning_rate": 6.870271210174139e-07,
"loss": 0.0078,
"step": 1834
},
{
"epoch": 11.328173374613003,
"grad_norm": 0.5617217241243505,
"learning_rate": 6.732489577182421e-07,
"loss": 0.0073,
"step": 1835
},
{
"epoch": 11.3343653250774,
"grad_norm": 0.4411947377739936,
"learning_rate": 6.596094198731473e-07,
"loss": 0.0059,
"step": 1836
},
{
"epoch": 11.340557275541796,
"grad_norm": 0.4038338751853977,
"learning_rate": 6.461085458140059e-07,
"loss": 0.0056,
"step": 1837
},
{
"epoch": 11.346749226006192,
"grad_norm": 0.45642980515517534,
"learning_rate": 6.327463734830174e-07,
"loss": 0.007,
"step": 1838
},
{
"epoch": 11.352941176470589,
"grad_norm": 0.712071078719854,
"learning_rate": 6.195229404325542e-07,
"loss": 0.0079,
"step": 1839
},
{
"epoch": 11.359133126934985,
"grad_norm": 0.40511189890054594,
"learning_rate": 6.064382838251059e-07,
"loss": 0.0069,
"step": 1840
},
{
"epoch": 11.365325077399381,
"grad_norm": 1.6722730305147262,
"learning_rate": 5.934924404331355e-07,
"loss": 0.0068,
"step": 1841
},
{
"epoch": 11.371517027863778,
"grad_norm": 0.77392054804963,
"learning_rate": 5.806854466390011e-07,
"loss": 0.0075,
"step": 1842
},
{
"epoch": 11.377708978328174,
"grad_norm": 0.509056072217164,
"learning_rate": 5.680173384348453e-07,
"loss": 0.0058,
"step": 1843
},
{
"epoch": 11.38390092879257,
"grad_norm": 0.37899654795682763,
"learning_rate": 5.554881514224897e-07,
"loss": 0.0049,
"step": 1844
},
{
"epoch": 11.390092879256965,
"grad_norm": 0.46168661436291886,
"learning_rate": 5.430979208133402e-07,
"loss": 0.0081,
"step": 1845
},
{
"epoch": 11.396284829721361,
"grad_norm": 0.8165580802526049,
"learning_rate": 5.30846681428282e-07,
"loss": 0.0066,
"step": 1846
},
{
"epoch": 11.402476780185758,
"grad_norm": 0.4790357107913533,
"learning_rate": 5.187344676976014e-07,
"loss": 0.0068,
"step": 1847
},
{
"epoch": 11.408668730650154,
"grad_norm": 0.46815040147214804,
"learning_rate": 5.067613136608473e-07,
"loss": 0.0093,
"step": 1848
},
{
"epoch": 11.41486068111455,
"grad_norm": 0.5326417302012959,
"learning_rate": 4.949272529667926e-07,
"loss": 0.0065,
"step": 1849
},
{
"epoch": 11.421052631578947,
"grad_norm": 0.5816630635354162,
"learning_rate": 4.832323188732835e-07,
"loss": 0.0075,
"step": 1850
},
{
"epoch": 11.427244582043343,
"grad_norm": 0.48427773798496576,
"learning_rate": 4.7167654424718487e-07,
"loss": 0.0068,
"step": 1851
},
{
"epoch": 11.43343653250774,
"grad_norm": 0.5497025572817426,
"learning_rate": 4.6025996156428e-07,
"loss": 0.0078,
"step": 1852
},
{
"epoch": 11.439628482972136,
"grad_norm": 0.5510500182306229,
"learning_rate": 4.4898260290915927e-07,
"loss": 0.0075,
"step": 1853
},
{
"epoch": 11.445820433436532,
"grad_norm": 0.44627509003732324,
"learning_rate": 4.3784449997515406e-07,
"loss": 0.0077,
"step": 1854
},
{
"epoch": 11.452012383900929,
"grad_norm": 0.4334395348335764,
"learning_rate": 4.2684568406423654e-07,
"loss": 0.0061,
"step": 1855
},
{
"epoch": 11.458204334365325,
"grad_norm": 0.4842234079078647,
"learning_rate": 4.159861860869307e-07,
"loss": 0.0068,
"step": 1856
},
{
"epoch": 11.464396284829721,
"grad_norm": 0.45320319836783474,
"learning_rate": 4.0526603656223516e-07,
"loss": 0.0073,
"step": 1857
},
{
"epoch": 11.470588235294118,
"grad_norm": 0.42575539666611417,
"learning_rate": 3.946852656175226e-07,
"loss": 0.0068,
"step": 1858
},
{
"epoch": 11.476780185758514,
"grad_norm": 0.4652897286828862,
"learning_rate": 3.8424390298846815e-07,
"loss": 0.0071,
"step": 1859
},
{
"epoch": 11.48297213622291,
"grad_norm": 0.5116973817438624,
"learning_rate": 3.739419780189657e-07,
"loss": 0.0085,
"step": 1860
},
{
"epoch": 11.489164086687307,
"grad_norm": 0.6935134709789724,
"learning_rate": 3.6377951966102276e-07,
"loss": 0.007,
"step": 1861
},
{
"epoch": 11.495356037151703,
"grad_norm": 0.41416047318174065,
"learning_rate": 3.537565564747103e-07,
"loss": 0.0074,
"step": 1862
},
{
"epoch": 11.5015479876161,
"grad_norm": 0.5586880199039042,
"learning_rate": 3.4387311662807397e-07,
"loss": 0.0073,
"step": 1863
},
{
"epoch": 11.507739938080496,
"grad_norm": 0.4896846203221236,
"learning_rate": 3.341292278970398e-07,
"loss": 0.0065,
"step": 1864
},
{
"epoch": 11.513931888544892,
"grad_norm": 0.5620994205712971,
"learning_rate": 3.24524917665342e-07,
"loss": 0.0066,
"step": 1865
},
{
"epoch": 11.520123839009289,
"grad_norm": 0.5017437806583144,
"learning_rate": 3.1506021292447287e-07,
"loss": 0.0067,
"step": 1866
},
{
"epoch": 11.526315789473685,
"grad_norm": 0.5215174105998739,
"learning_rate": 3.057351402735553e-07,
"loss": 0.0082,
"step": 1867
},
{
"epoch": 11.532507739938081,
"grad_norm": 0.4829789050249859,
"learning_rate": 2.965497259193151e-07,
"loss": 0.0067,
"step": 1868
},
{
"epoch": 11.538699690402478,
"grad_norm": 0.5822395608103282,
"learning_rate": 2.8750399567599173e-07,
"loss": 0.0067,
"step": 1869
},
{
"epoch": 11.544891640866872,
"grad_norm": 0.45867408894473916,
"learning_rate": 2.7859797496523897e-07,
"loss": 0.0072,
"step": 1870
},
{
"epoch": 11.551083591331269,
"grad_norm": 0.5205589054718633,
"learning_rate": 2.6983168881611895e-07,
"loss": 0.0055,
"step": 1871
},
{
"epoch": 11.557275541795665,
"grad_norm": 0.39738862312410056,
"learning_rate": 2.612051618649525e-07,
"loss": 0.006,
"step": 1872
},
{
"epoch": 11.563467492260061,
"grad_norm": 0.5013445636761662,
"learning_rate": 2.527184183553022e-07,
"loss": 0.0079,
"step": 1873
},
{
"epoch": 11.569659442724458,
"grad_norm": 0.5104992056179086,
"learning_rate": 2.443714821379006e-07,
"loss": 0.0085,
"step": 1874
},
{
"epoch": 11.575851393188854,
"grad_norm": 0.691277343224864,
"learning_rate": 2.3616437667055015e-07,
"loss": 0.008,
"step": 1875
},
{
"epoch": 11.58204334365325,
"grad_norm": 0.4802053768271533,
"learning_rate": 2.2809712501810632e-07,
"loss": 0.0067,
"step": 1876
},
{
"epoch": 11.588235294117647,
"grad_norm": 0.605809537482388,
"learning_rate": 2.2016974985236693e-07,
"loss": 0.0071,
"step": 1877
},
{
"epoch": 11.594427244582043,
"grad_norm": 0.7705825543183333,
"learning_rate": 2.1238227345202754e-07,
"loss": 0.0073,
"step": 1878
},
{
"epoch": 11.60061919504644,
"grad_norm": 0.4715519242573243,
"learning_rate": 2.047347177026371e-07,
"loss": 0.007,
"step": 1879
},
{
"epoch": 11.606811145510836,
"grad_norm": 1.0390341494339923,
"learning_rate": 1.9722710409650347e-07,
"loss": 0.007,
"step": 1880
},
{
"epoch": 11.613003095975232,
"grad_norm": 0.44854359389428083,
"learning_rate": 1.898594537326437e-07,
"loss": 0.005,
"step": 1881
},
{
"epoch": 11.619195046439629,
"grad_norm": 0.38508653976333107,
"learning_rate": 1.8263178731675047e-07,
"loss": 0.0063,
"step": 1882
},
{
"epoch": 11.625386996904025,
"grad_norm": 0.6846671924981257,
"learning_rate": 1.7554412516108677e-07,
"loss": 0.0075,
"step": 1883
},
{
"epoch": 11.631578947368421,
"grad_norm": 0.4785043704303202,
"learning_rate": 1.685964871844692e-07,
"loss": 0.007,
"step": 1884
},
{
"epoch": 11.637770897832818,
"grad_norm": 0.5534942405207746,
"learning_rate": 1.6178889291220133e-07,
"loss": 0.0066,
"step": 1885
},
{
"epoch": 11.643962848297214,
"grad_norm": 0.6334060900627472,
"learning_rate": 1.5512136147600164e-07,
"loss": 0.0067,
"step": 1886
},
{
"epoch": 11.65015479876161,
"grad_norm": 0.5157177078600752,
"learning_rate": 1.4859391161397007e-07,
"loss": 0.0065,
"step": 1887
},
{
"epoch": 11.656346749226007,
"grad_norm": 0.45312969046876567,
"learning_rate": 1.4220656167053258e-07,
"loss": 0.0079,
"step": 1888
},
{
"epoch": 11.662538699690403,
"grad_norm": 0.5352641681979426,
"learning_rate": 1.3595932959638013e-07,
"loss": 0.0069,
"step": 1889
},
{
"epoch": 11.6687306501548,
"grad_norm": 0.42825165496945455,
"learning_rate": 1.2985223294841865e-07,
"loss": 0.0068,
"step": 1890
},
{
"epoch": 11.674922600619196,
"grad_norm": 0.46191991329180665,
"learning_rate": 1.2388528888973017e-07,
"loss": 0.0063,
"step": 1891
},
{
"epoch": 11.681114551083592,
"grad_norm": 0.5332588937649859,
"learning_rate": 1.1805851418952296e-07,
"loss": 0.0056,
"step": 1892
},
{
"epoch": 11.687306501547988,
"grad_norm": 0.5317234947401625,
"learning_rate": 1.1237192522307593e-07,
"loss": 0.006,
"step": 1893
},
{
"epoch": 11.693498452012385,
"grad_norm": 0.49736650975887164,
"learning_rate": 1.068255379716887e-07,
"loss": 0.0081,
"step": 1894
},
{
"epoch": 11.69969040247678,
"grad_norm": 0.47643332055864485,
"learning_rate": 1.0141936802265939e-07,
"loss": 0.007,
"step": 1895
},
{
"epoch": 11.705882352941176,
"grad_norm": 0.41908853217415565,
"learning_rate": 9.615343056922354e-08,
"loss": 0.0059,
"step": 1896
},
{
"epoch": 11.712074303405572,
"grad_norm": 0.5727683236930392,
"learning_rate": 9.102774041049866e-08,
"loss": 0.0063,
"step": 1897
},
{
"epoch": 11.718266253869968,
"grad_norm": 0.43881012020217974,
"learning_rate": 8.604231195147861e-08,
"loss": 0.007,
"step": 1898
},
{
"epoch": 11.724458204334365,
"grad_norm": 0.434012984801701,
"learning_rate": 8.119715920296145e-08,
"loss": 0.0071,
"step": 1899
},
{
"epoch": 11.730650154798761,
"grad_norm": 0.5668726891302649,
"learning_rate": 7.649229578152728e-08,
"loss": 0.0056,
"step": 1900
},
{
"epoch": 11.736842105263158,
"grad_norm": 0.5656514949124021,
"learning_rate": 7.192773490948823e-08,
"loss": 0.0072,
"step": 1901
},
{
"epoch": 11.743034055727554,
"grad_norm": 0.6164067533137532,
"learning_rate": 6.750348941486073e-08,
"loss": 0.0092,
"step": 1902
},
{
"epoch": 11.74922600619195,
"grad_norm": 1.2472133137974002,
"learning_rate": 6.321957173132664e-08,
"loss": 0.0066,
"step": 1903
},
{
"epoch": 11.755417956656347,
"grad_norm": 0.4392934542646206,
"learning_rate": 5.9075993898188854e-08,
"loss": 0.0075,
"step": 1904
},
{
"epoch": 11.761609907120743,
"grad_norm": 0.49472358771014996,
"learning_rate": 5.507276756036017e-08,
"loss": 0.0072,
"step": 1905
},
{
"epoch": 11.76780185758514,
"grad_norm": 1.1317299783753785,
"learning_rate": 5.1209903968302276e-08,
"loss": 0.0067,
"step": 1906
},
{
"epoch": 11.773993808049536,
"grad_norm": 0.5987121278705873,
"learning_rate": 4.7487413978025696e-08,
"loss": 0.0059,
"step": 1907
},
{
"epoch": 11.780185758513932,
"grad_norm": 0.4778726713011685,
"learning_rate": 4.390530805102322e-08,
"loss": 0.0069,
"step": 1908
},
{
"epoch": 11.786377708978328,
"grad_norm": 0.5588314148416086,
"learning_rate": 4.0463596254269875e-08,
"loss": 0.0063,
"step": 1909
},
{
"epoch": 11.792569659442725,
"grad_norm": 0.4740418584164989,
"learning_rate": 3.7162288260178536e-08,
"loss": 0.0094,
"step": 1910
},
{
"epoch": 11.798761609907121,
"grad_norm": 0.6334094812461899,
"learning_rate": 3.4001393346588806e-08,
"loss": 0.0063,
"step": 1911
},
{
"epoch": 11.804953560371517,
"grad_norm": 0.8497258020172912,
"learning_rate": 3.098092039671707e-08,
"loss": 0.0097,
"step": 1912
},
{
"epoch": 11.811145510835914,
"grad_norm": 0.5492510215863654,
"learning_rate": 2.8100877899156498e-08,
"loss": 0.0069,
"step": 1913
},
{
"epoch": 11.81733746130031,
"grad_norm": 0.46923531975887417,
"learning_rate": 2.5361273947827058e-08,
"loss": 0.0087,
"step": 1914
},
{
"epoch": 11.823529411764707,
"grad_norm": 0.5954689826520188,
"learning_rate": 2.2762116241981103e-08,
"loss": 0.0072,
"step": 1915
},
{
"epoch": 11.829721362229103,
"grad_norm": 0.6339550310655667,
"learning_rate": 2.0303412086164485e-08,
"loss": 0.0072,
"step": 1916
},
{
"epoch": 11.8359133126935,
"grad_norm": 0.4144783163072785,
"learning_rate": 1.7985168390194372e-08,
"loss": 0.0056,
"step": 1917
},
{
"epoch": 11.842105263157894,
"grad_norm": 0.40612660529647476,
"learning_rate": 1.580739166914813e-08,
"loss": 0.0055,
"step": 1918
},
{
"epoch": 11.848297213622292,
"grad_norm": 0.4767050444962352,
"learning_rate": 1.3770088043335572e-08,
"loss": 0.0081,
"step": 1919
},
{
"epoch": 11.854489164086687,
"grad_norm": 0.6771418638491573,
"learning_rate": 1.187326323830451e-08,
"loss": 0.0059,
"step": 1920
},
{
"epoch": 11.860681114551083,
"grad_norm": 0.5268183349235755,
"learning_rate": 1.01169225847908e-08,
"loss": 0.0084,
"step": 1921
},
{
"epoch": 11.86687306501548,
"grad_norm": 0.5714529234379931,
"learning_rate": 8.501071018729434e-09,
"loss": 0.0085,
"step": 1922
},
{
"epoch": 11.873065015479876,
"grad_norm": 0.641796851723476,
"learning_rate": 7.025713081232344e-09,
"loss": 0.0061,
"step": 1923
},
{
"epoch": 11.879256965944272,
"grad_norm": 0.5786492854731642,
"learning_rate": 5.690852918566192e-09,
"loss": 0.0074,
"step": 1924
},
{
"epoch": 11.885448916408668,
"grad_norm": 0.4395964372874901,
"learning_rate": 4.496494282157926e-09,
"loss": 0.0077,
"step": 1925
},
{
"epoch": 11.891640866873065,
"grad_norm": 0.6144859848178597,
"learning_rate": 3.4426405285725717e-09,
"loss": 0.0062,
"step": 1926
},
{
"epoch": 11.897832817337461,
"grad_norm": 0.45027881629827304,
"learning_rate": 2.529294619513234e-09,
"loss": 0.0053,
"step": 1927
},
{
"epoch": 11.904024767801857,
"grad_norm": 0.4100958029737214,
"learning_rate": 1.7564591217933412e-09,
"loss": 0.0071,
"step": 1928
},
{
"epoch": 11.910216718266254,
"grad_norm": 0.49678388184393957,
"learning_rate": 1.1241362073588502e-09,
"loss": 0.0084,
"step": 1929
},
{
"epoch": 11.91640866873065,
"grad_norm": 0.5684024725276834,
"learning_rate": 6.323276532604893e-10,
"loss": 0.0067,
"step": 1930
},
{
"epoch": 11.922600619195046,
"grad_norm": 0.5054176005019508,
"learning_rate": 2.8103484164820893e-10,
"loss": 0.0071,
"step": 1931
},
{
"epoch": 11.928792569659443,
"grad_norm": 0.6033011906551876,
"learning_rate": 7.025875977673124e-11,
"loss": 0.0063,
"step": 1932
},
{
"epoch": 11.928792569659443,
"step": 1932,
"total_flos": 6213184927956992.0,
"train_loss": 1.5306146964587475,
"train_runtime": 72417.837,
"train_samples_per_second": 3.424,
"train_steps_per_second": 0.027
}
],
"logging_steps": 1.0,
"max_steps": 1932,
"num_input_tokens_seen": 0,
"num_train_epochs": 12,
"save_steps": 50000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 6213184927956992.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}