Qwen2.5-1.5B-Open-R1-Distill / trainer_state.json
calledice666's picture
Model save
7ce18ae verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9999074502545118,
"eval_steps": 100,
"global_step": 2701,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0018509949097639982,
"grad_norm": 2.749824922426202,
"learning_rate": 3.690036900369004e-07,
"loss": 1.1027,
"step": 5
},
{
"epoch": 0.0037019898195279964,
"grad_norm": 2.7133127995482353,
"learning_rate": 7.380073800738008e-07,
"loss": 1.107,
"step": 10
},
{
"epoch": 0.005552984729291994,
"grad_norm": 2.5060043757547894,
"learning_rate": 1.1070110701107011e-06,
"loss": 1.0411,
"step": 15
},
{
"epoch": 0.007403979639055993,
"grad_norm": 2.580301155440442,
"learning_rate": 1.4760147601476015e-06,
"loss": 1.0394,
"step": 20
},
{
"epoch": 0.00925497454881999,
"grad_norm": 2.101103627946831,
"learning_rate": 1.845018450184502e-06,
"loss": 1.025,
"step": 25
},
{
"epoch": 0.011105969458583989,
"grad_norm": 1.7798274009053514,
"learning_rate": 2.2140221402214023e-06,
"loss": 1.0642,
"step": 30
},
{
"epoch": 0.012956964368347987,
"grad_norm": 1.4788395927397677,
"learning_rate": 2.5830258302583027e-06,
"loss": 0.9636,
"step": 35
},
{
"epoch": 0.014807959278111986,
"grad_norm": 1.4694159866777265,
"learning_rate": 2.952029520295203e-06,
"loss": 1.0495,
"step": 40
},
{
"epoch": 0.016658954187875982,
"grad_norm": 1.529975198759594,
"learning_rate": 3.3210332103321034e-06,
"loss": 0.9308,
"step": 45
},
{
"epoch": 0.01850994909763998,
"grad_norm": 1.4286077278093599,
"learning_rate": 3.690036900369004e-06,
"loss": 0.9909,
"step": 50
},
{
"epoch": 0.02036094400740398,
"grad_norm": 1.336405308330729,
"learning_rate": 4.059040590405905e-06,
"loss": 0.948,
"step": 55
},
{
"epoch": 0.022211938917167977,
"grad_norm": 1.411657382339046,
"learning_rate": 4.428044280442805e-06,
"loss": 0.9434,
"step": 60
},
{
"epoch": 0.024062933826931976,
"grad_norm": 1.3370491817510695,
"learning_rate": 4.797047970479705e-06,
"loss": 0.9112,
"step": 65
},
{
"epoch": 0.025913928736695974,
"grad_norm": 1.238804041054237,
"learning_rate": 5.166051660516605e-06,
"loss": 0.864,
"step": 70
},
{
"epoch": 0.027764923646459973,
"grad_norm": 1.5683504507074115,
"learning_rate": 5.535055350553506e-06,
"loss": 0.927,
"step": 75
},
{
"epoch": 0.02961591855622397,
"grad_norm": 1.3407079562724307,
"learning_rate": 5.904059040590406e-06,
"loss": 0.8787,
"step": 80
},
{
"epoch": 0.03146691346598797,
"grad_norm": 1.1838954340614056,
"learning_rate": 6.273062730627307e-06,
"loss": 0.9208,
"step": 85
},
{
"epoch": 0.033317908375751965,
"grad_norm": 1.5881922549502974,
"learning_rate": 6.642066420664207e-06,
"loss": 0.8529,
"step": 90
},
{
"epoch": 0.035168903285515966,
"grad_norm": 1.4451601927412367,
"learning_rate": 7.011070110701108e-06,
"loss": 0.9031,
"step": 95
},
{
"epoch": 0.03701989819527996,
"grad_norm": 1.2272450676249622,
"learning_rate": 7.380073800738008e-06,
"loss": 0.8384,
"step": 100
},
{
"epoch": 0.03701989819527996,
"eval_loss": 0.8794825673103333,
"eval_runtime": 125.7829,
"eval_samples_per_second": 1.018,
"eval_steps_per_second": 0.509,
"step": 100
},
{
"epoch": 0.03887089310504396,
"grad_norm": 1.1893073354157337,
"learning_rate": 7.749077490774908e-06,
"loss": 0.889,
"step": 105
},
{
"epoch": 0.04072188801480796,
"grad_norm": 1.3052090560256295,
"learning_rate": 8.11808118081181e-06,
"loss": 0.8133,
"step": 110
},
{
"epoch": 0.04257288292457196,
"grad_norm": 1.37698008425289,
"learning_rate": 8.48708487084871e-06,
"loss": 0.865,
"step": 115
},
{
"epoch": 0.044423877834335955,
"grad_norm": 1.303834447656376,
"learning_rate": 8.85608856088561e-06,
"loss": 0.846,
"step": 120
},
{
"epoch": 0.04627487274409996,
"grad_norm": 1.2327474858734326,
"learning_rate": 9.22509225092251e-06,
"loss": 0.7942,
"step": 125
},
{
"epoch": 0.04812586765386395,
"grad_norm": 1.2058003679999099,
"learning_rate": 9.59409594095941e-06,
"loss": 0.8417,
"step": 130
},
{
"epoch": 0.04997686256362795,
"grad_norm": 1.2290254882715699,
"learning_rate": 9.963099630996312e-06,
"loss": 0.8701,
"step": 135
},
{
"epoch": 0.05182785747339195,
"grad_norm": 1.1760267952662564,
"learning_rate": 1.033210332103321e-05,
"loss": 0.8252,
"step": 140
},
{
"epoch": 0.053678852383155944,
"grad_norm": 1.269131376133505,
"learning_rate": 1.0701107011070112e-05,
"loss": 0.8302,
"step": 145
},
{
"epoch": 0.055529847292919945,
"grad_norm": 1.1597954970515953,
"learning_rate": 1.1070110701107012e-05,
"loss": 0.8479,
"step": 150
},
{
"epoch": 0.05738084220268394,
"grad_norm": 1.2448895450141289,
"learning_rate": 1.1439114391143913e-05,
"loss": 0.8521,
"step": 155
},
{
"epoch": 0.05923183711244794,
"grad_norm": 1.1897405950550726,
"learning_rate": 1.1808118081180812e-05,
"loss": 0.7592,
"step": 160
},
{
"epoch": 0.06108283202221194,
"grad_norm": 1.314365342475427,
"learning_rate": 1.2177121771217713e-05,
"loss": 0.8289,
"step": 165
},
{
"epoch": 0.06293382693197594,
"grad_norm": 1.3600021570595577,
"learning_rate": 1.2546125461254614e-05,
"loss": 0.8119,
"step": 170
},
{
"epoch": 0.06478482184173993,
"grad_norm": 1.2381592999546391,
"learning_rate": 1.2915129151291515e-05,
"loss": 0.8081,
"step": 175
},
{
"epoch": 0.06663581675150393,
"grad_norm": 1.2650464435907114,
"learning_rate": 1.3284132841328414e-05,
"loss": 0.8173,
"step": 180
},
{
"epoch": 0.06848681166126794,
"grad_norm": 1.2523135217657222,
"learning_rate": 1.3653136531365315e-05,
"loss": 0.8105,
"step": 185
},
{
"epoch": 0.07033780657103193,
"grad_norm": 1.1760807838355647,
"learning_rate": 1.4022140221402215e-05,
"loss": 0.8644,
"step": 190
},
{
"epoch": 0.07218880148079593,
"grad_norm": 1.2557569983889052,
"learning_rate": 1.4391143911439116e-05,
"loss": 0.8031,
"step": 195
},
{
"epoch": 0.07403979639055992,
"grad_norm": 1.2032960225457723,
"learning_rate": 1.4760147601476015e-05,
"loss": 0.8639,
"step": 200
},
{
"epoch": 0.07403979639055992,
"eval_loss": 0.8400982022285461,
"eval_runtime": 125.9532,
"eval_samples_per_second": 1.016,
"eval_steps_per_second": 0.508,
"step": 200
},
{
"epoch": 0.07589079130032392,
"grad_norm": 1.394672996108549,
"learning_rate": 1.5129151291512916e-05,
"loss": 0.8164,
"step": 205
},
{
"epoch": 0.07774178621008793,
"grad_norm": 1.3016173569055736,
"learning_rate": 1.5498154981549817e-05,
"loss": 0.8478,
"step": 210
},
{
"epoch": 0.07959278111985192,
"grad_norm": 1.406780438265723,
"learning_rate": 1.5867158671586716e-05,
"loss": 0.841,
"step": 215
},
{
"epoch": 0.08144377602961592,
"grad_norm": 1.2377690323264836,
"learning_rate": 1.623616236162362e-05,
"loss": 0.8287,
"step": 220
},
{
"epoch": 0.08329477093937991,
"grad_norm": 1.39607937433599,
"learning_rate": 1.6605166051660518e-05,
"loss": 0.8176,
"step": 225
},
{
"epoch": 0.08514576584914392,
"grad_norm": 1.3980920756979587,
"learning_rate": 1.697416974169742e-05,
"loss": 0.7867,
"step": 230
},
{
"epoch": 0.08699676075890792,
"grad_norm": 1.2563235907029433,
"learning_rate": 1.734317343173432e-05,
"loss": 0.8447,
"step": 235
},
{
"epoch": 0.08884775566867191,
"grad_norm": 1.4806853438242649,
"learning_rate": 1.771217712177122e-05,
"loss": 0.807,
"step": 240
},
{
"epoch": 0.0906987505784359,
"grad_norm": 1.2447141754207756,
"learning_rate": 1.8081180811808117e-05,
"loss": 0.7962,
"step": 245
},
{
"epoch": 0.09254974548819991,
"grad_norm": 1.3080325661809735,
"learning_rate": 1.845018450184502e-05,
"loss": 0.7745,
"step": 250
},
{
"epoch": 0.09440074039796391,
"grad_norm": 1.179323447687895,
"learning_rate": 1.8819188191881922e-05,
"loss": 0.8574,
"step": 255
},
{
"epoch": 0.0962517353077279,
"grad_norm": 1.2896705705328326,
"learning_rate": 1.918819188191882e-05,
"loss": 0.8296,
"step": 260
},
{
"epoch": 0.0981027302174919,
"grad_norm": 1.3644542064844416,
"learning_rate": 1.955719557195572e-05,
"loss": 0.8251,
"step": 265
},
{
"epoch": 0.0999537251272559,
"grad_norm": 1.286282268357473,
"learning_rate": 1.9926199261992623e-05,
"loss": 0.7754,
"step": 270
},
{
"epoch": 0.1018047200370199,
"grad_norm": 1.2329112299042364,
"learning_rate": 1.999986628620426e-05,
"loss": 0.7915,
"step": 275
},
{
"epoch": 0.1036557149467839,
"grad_norm": 1.3173883195813032,
"learning_rate": 1.9999323080037623e-05,
"loss": 0.866,
"step": 280
},
{
"epoch": 0.10550670985654789,
"grad_norm": 1.2611947477770145,
"learning_rate": 1.9998362047068548e-05,
"loss": 0.83,
"step": 285
},
{
"epoch": 0.10735770476631189,
"grad_norm": 1.1796656758558173,
"learning_rate": 1.9996983227454284e-05,
"loss": 0.8357,
"step": 290
},
{
"epoch": 0.1092086996760759,
"grad_norm": 1.1353613990508302,
"learning_rate": 1.9995186678809513e-05,
"loss": 0.7181,
"step": 295
},
{
"epoch": 0.11105969458583989,
"grad_norm": 1.2567062023606574,
"learning_rate": 1.999297247620393e-05,
"loss": 0.8598,
"step": 300
},
{
"epoch": 0.11105969458583989,
"eval_loss": 0.8292834758758545,
"eval_runtime": 125.837,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.509,
"step": 300
},
{
"epoch": 0.11291068949560389,
"grad_norm": 1.3854824655270053,
"learning_rate": 1.999034071215912e-05,
"loss": 0.869,
"step": 305
},
{
"epoch": 0.11476168440536788,
"grad_norm": 1.4180966128082728,
"learning_rate": 1.998729149664468e-05,
"loss": 0.7802,
"step": 310
},
{
"epoch": 0.11661267931513189,
"grad_norm": 1.2687340435719516,
"learning_rate": 1.9983824957073632e-05,
"loss": 0.7751,
"step": 315
},
{
"epoch": 0.11846367422489588,
"grad_norm": 1.2012101992327227,
"learning_rate": 1.99799412382971e-05,
"loss": 0.7831,
"step": 320
},
{
"epoch": 0.12031466913465988,
"grad_norm": 1.2085415925635907,
"learning_rate": 1.9975640502598243e-05,
"loss": 0.788,
"step": 325
},
{
"epoch": 0.12216566404442387,
"grad_norm": 1.2183184035114982,
"learning_rate": 1.9970922929685496e-05,
"loss": 0.7998,
"step": 330
},
{
"epoch": 0.12401665895418787,
"grad_norm": 1.2657715995907186,
"learning_rate": 1.996578871668504e-05,
"loss": 0.7978,
"step": 335
},
{
"epoch": 0.12586765386395188,
"grad_norm": 1.1898542144979154,
"learning_rate": 1.996023807813258e-05,
"loss": 0.7861,
"step": 340
},
{
"epoch": 0.12771864877371586,
"grad_norm": 1.2193829841446657,
"learning_rate": 1.995427124596437e-05,
"loss": 0.8067,
"step": 345
},
{
"epoch": 0.12956964368347987,
"grad_norm": 1.1563840168462576,
"learning_rate": 1.9947888469507527e-05,
"loss": 0.7666,
"step": 350
},
{
"epoch": 0.13142063859324388,
"grad_norm": 1.3445764875998873,
"learning_rate": 1.9941090015469614e-05,
"loss": 0.7855,
"step": 355
},
{
"epoch": 0.13327163350300786,
"grad_norm": 1.2515125913797396,
"learning_rate": 1.9933876167927494e-05,
"loss": 0.8211,
"step": 360
},
{
"epoch": 0.13512262841277187,
"grad_norm": 1.2363244217826992,
"learning_rate": 1.9926247228315455e-05,
"loss": 0.7726,
"step": 365
},
{
"epoch": 0.13697362332253588,
"grad_norm": 1.1398366050989162,
"learning_rate": 1.9918203515412616e-05,
"loss": 0.7809,
"step": 370
},
{
"epoch": 0.13882461823229986,
"grad_norm": 1.1609837477400817,
"learning_rate": 1.9909745365329623e-05,
"loss": 0.7981,
"step": 375
},
{
"epoch": 0.14067561314206387,
"grad_norm": 1.1428050730123682,
"learning_rate": 1.990087313149457e-05,
"loss": 0.8055,
"step": 380
},
{
"epoch": 0.14252660805182785,
"grad_norm": 1.1811407177495057,
"learning_rate": 1.9891587184638274e-05,
"loss": 0.7898,
"step": 385
},
{
"epoch": 0.14437760296159186,
"grad_norm": 1.1089125886531135,
"learning_rate": 1.9881887912778738e-05,
"loss": 0.8435,
"step": 390
},
{
"epoch": 0.14622859787135586,
"grad_norm": 1.1264567214650973,
"learning_rate": 1.9871775721204973e-05,
"loss": 0.8027,
"step": 395
},
{
"epoch": 0.14807959278111985,
"grad_norm": 1.2984452085681775,
"learning_rate": 1.9861251032460053e-05,
"loss": 0.8238,
"step": 400
},
{
"epoch": 0.14807959278111985,
"eval_loss": 0.8187944889068604,
"eval_runtime": 125.8941,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 400
},
{
"epoch": 0.14993058769088385,
"grad_norm": 1.262457467293623,
"learning_rate": 1.985031428632345e-05,
"loss": 0.7656,
"step": 405
},
{
"epoch": 0.15178158260064784,
"grad_norm": 1.222235652953353,
"learning_rate": 1.9838965939792666e-05,
"loss": 0.8033,
"step": 410
},
{
"epoch": 0.15363257751041184,
"grad_norm": 1.0872174045726444,
"learning_rate": 1.9827206467064133e-05,
"loss": 0.8079,
"step": 415
},
{
"epoch": 0.15548357242017585,
"grad_norm": 1.188894944380026,
"learning_rate": 1.9815036359513408e-05,
"loss": 0.7931,
"step": 420
},
{
"epoch": 0.15733456732993983,
"grad_norm": 1.0474799030998936,
"learning_rate": 1.980245612567462e-05,
"loss": 0.7752,
"step": 425
},
{
"epoch": 0.15918556223970384,
"grad_norm": 1.0714711020851349,
"learning_rate": 1.9789466291219246e-05,
"loss": 0.7787,
"step": 430
},
{
"epoch": 0.16103655714946785,
"grad_norm": 1.2916090137573928,
"learning_rate": 1.9776067398934122e-05,
"loss": 0.8464,
"step": 435
},
{
"epoch": 0.16288755205923183,
"grad_norm": 1.0837760464706834,
"learning_rate": 1.9762260008698787e-05,
"loss": 0.742,
"step": 440
},
{
"epoch": 0.16473854696899584,
"grad_norm": 1.0638766335715402,
"learning_rate": 1.974804469746206e-05,
"loss": 0.7909,
"step": 445
},
{
"epoch": 0.16658954187875982,
"grad_norm": 1.1650137243413088,
"learning_rate": 1.9733422059217954e-05,
"loss": 0.8142,
"step": 450
},
{
"epoch": 0.16844053678852383,
"grad_norm": 1.1494193680521951,
"learning_rate": 1.9718392704980852e-05,
"loss": 0.7187,
"step": 455
},
{
"epoch": 0.17029153169828784,
"grad_norm": 1.141831785166932,
"learning_rate": 1.9702957262759964e-05,
"loss": 0.7338,
"step": 460
},
{
"epoch": 0.17214252660805182,
"grad_norm": 1.1597583259361357,
"learning_rate": 1.9687116377533102e-05,
"loss": 0.802,
"step": 465
},
{
"epoch": 0.17399352151781583,
"grad_norm": 1.1829435674368602,
"learning_rate": 1.9670870711219708e-05,
"loss": 0.7982,
"step": 470
},
{
"epoch": 0.1758445164275798,
"grad_norm": 1.2456510178235929,
"learning_rate": 1.9654220942653223e-05,
"loss": 0.796,
"step": 475
},
{
"epoch": 0.17769551133734382,
"grad_norm": 1.1776301109970682,
"learning_rate": 1.9637167767552687e-05,
"loss": 0.7754,
"step": 480
},
{
"epoch": 0.17954650624710783,
"grad_norm": 1.116849386127364,
"learning_rate": 1.9619711898493707e-05,
"loss": 0.7919,
"step": 485
},
{
"epoch": 0.1813975011568718,
"grad_norm": 1.090344008956795,
"learning_rate": 1.9601854064878645e-05,
"loss": 0.84,
"step": 490
},
{
"epoch": 0.18324849606663582,
"grad_norm": 1.0862481361440726,
"learning_rate": 1.9583595012906173e-05,
"loss": 0.7697,
"step": 495
},
{
"epoch": 0.18509949097639983,
"grad_norm": 1.142567519924341,
"learning_rate": 1.956493550554006e-05,
"loss": 0.7635,
"step": 500
},
{
"epoch": 0.18509949097639983,
"eval_loss": 0.8105577230453491,
"eval_runtime": 125.9442,
"eval_samples_per_second": 1.016,
"eval_steps_per_second": 0.508,
"step": 500
},
{
"epoch": 0.1869504858861638,
"grad_norm": 1.023479803471901,
"learning_rate": 1.954587632247732e-05,
"loss": 0.7747,
"step": 505
},
{
"epoch": 0.18880148079592782,
"grad_norm": 1.063385408309198,
"learning_rate": 1.9526418260115615e-05,
"loss": 0.75,
"step": 510
},
{
"epoch": 0.1906524757056918,
"grad_norm": 1.1699474105718337,
"learning_rate": 1.9506562131519978e-05,
"loss": 0.7914,
"step": 515
},
{
"epoch": 0.1925034706154558,
"grad_norm": 1.2205261625534392,
"learning_rate": 1.9486308766388843e-05,
"loss": 0.8228,
"step": 520
},
{
"epoch": 0.19435446552521982,
"grad_norm": 1.2493323461591765,
"learning_rate": 1.9465659011019384e-05,
"loss": 0.7494,
"step": 525
},
{
"epoch": 0.1962054604349838,
"grad_norm": 1.083050859566808,
"learning_rate": 1.9444613728272127e-05,
"loss": 0.7478,
"step": 530
},
{
"epoch": 0.1980564553447478,
"grad_norm": 1.122573251625821,
"learning_rate": 1.9423173797534924e-05,
"loss": 0.8088,
"step": 535
},
{
"epoch": 0.1999074502545118,
"grad_norm": 1.2528116581295516,
"learning_rate": 1.9401340114686187e-05,
"loss": 0.7854,
"step": 540
},
{
"epoch": 0.2017584451642758,
"grad_norm": 1.2245250113750425,
"learning_rate": 1.9379113592057468e-05,
"loss": 0.7745,
"step": 545
},
{
"epoch": 0.2036094400740398,
"grad_norm": 1.3603783721746314,
"learning_rate": 1.9356495158395317e-05,
"loss": 0.8214,
"step": 550
},
{
"epoch": 0.20546043498380379,
"grad_norm": 1.1676680192314415,
"learning_rate": 1.9333485758822495e-05,
"loss": 0.7731,
"step": 555
},
{
"epoch": 0.2073114298935678,
"grad_norm": 1.2142076942764124,
"learning_rate": 1.931008635479847e-05,
"loss": 0.8128,
"step": 560
},
{
"epoch": 0.2091624248033318,
"grad_norm": 1.1799986315552842,
"learning_rate": 1.9286297924079244e-05,
"loss": 0.7536,
"step": 565
},
{
"epoch": 0.21101341971309578,
"grad_norm": 1.0965045706780743,
"learning_rate": 1.92621214606765e-05,
"loss": 0.7744,
"step": 570
},
{
"epoch": 0.2128644146228598,
"grad_norm": 1.2434542202635104,
"learning_rate": 1.9237557974816063e-05,
"loss": 0.7965,
"step": 575
},
{
"epoch": 0.21471540953262377,
"grad_norm": 1.2091983359476026,
"learning_rate": 1.921260849289568e-05,
"loss": 0.7837,
"step": 580
},
{
"epoch": 0.21656640444238778,
"grad_norm": 1.1940311597580997,
"learning_rate": 1.9187274057442153e-05,
"loss": 0.774,
"step": 585
},
{
"epoch": 0.2184173993521518,
"grad_norm": 1.0743701754676238,
"learning_rate": 1.916155572706776e-05,
"loss": 0.8083,
"step": 590
},
{
"epoch": 0.22026839426191577,
"grad_norm": 1.119330031173641,
"learning_rate": 1.913545457642601e-05,
"loss": 0.7692,
"step": 595
},
{
"epoch": 0.22211938917167978,
"grad_norm": 1.1838482419139507,
"learning_rate": 1.9108971696166778e-05,
"loss": 0.7388,
"step": 600
},
{
"epoch": 0.22211938917167978,
"eval_loss": 0.8023950457572937,
"eval_runtime": 125.9122,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 600
},
{
"epoch": 0.22397038408144376,
"grad_norm": 1.1918846048877687,
"learning_rate": 1.9082108192890673e-05,
"loss": 0.806,
"step": 605
},
{
"epoch": 0.22582137899120777,
"grad_norm": 1.1474696064483147,
"learning_rate": 1.905486518910286e-05,
"loss": 0.7756,
"step": 610
},
{
"epoch": 0.22767237390097178,
"grad_norm": 1.1873077682787359,
"learning_rate": 1.9027243823166107e-05,
"loss": 0.7807,
"step": 615
},
{
"epoch": 0.22952336881073576,
"grad_norm": 1.1370410152260748,
"learning_rate": 1.8999245249253247e-05,
"loss": 0.8518,
"step": 620
},
{
"epoch": 0.23137436372049977,
"grad_norm": 1.1592714123770727,
"learning_rate": 1.8970870637298936e-05,
"loss": 0.7947,
"step": 625
},
{
"epoch": 0.23322535863026378,
"grad_norm": 1.167379979369555,
"learning_rate": 1.8942121172950765e-05,
"loss": 0.8175,
"step": 630
},
{
"epoch": 0.23507635354002776,
"grad_norm": 1.0148428174894528,
"learning_rate": 1.8912998057519735e-05,
"loss": 0.7813,
"step": 635
},
{
"epoch": 0.23692734844979177,
"grad_norm": 1.2294496173862697,
"learning_rate": 1.8883502507930044e-05,
"loss": 0.8282,
"step": 640
},
{
"epoch": 0.23877834335955575,
"grad_norm": 1.153834698104633,
"learning_rate": 1.885363575666823e-05,
"loss": 0.819,
"step": 645
},
{
"epoch": 0.24062933826931976,
"grad_norm": 1.0699702689958812,
"learning_rate": 1.8823399051731698e-05,
"loss": 0.7333,
"step": 650
},
{
"epoch": 0.24248033317908377,
"grad_norm": 1.0339682532907033,
"learning_rate": 1.8792793656576544e-05,
"loss": 0.7723,
"step": 655
},
{
"epoch": 0.24433132808884775,
"grad_norm": 1.083476975643199,
"learning_rate": 1.876182085006478e-05,
"loss": 0.7932,
"step": 660
},
{
"epoch": 0.24618232299861176,
"grad_norm": 0.9842008335663365,
"learning_rate": 1.873048192641088e-05,
"loss": 0.7978,
"step": 665
},
{
"epoch": 0.24803331790837574,
"grad_norm": 1.0859880941267326,
"learning_rate": 1.8698778195127715e-05,
"loss": 0.7903,
"step": 670
},
{
"epoch": 0.24988431281813975,
"grad_norm": 1.0920881647401042,
"learning_rate": 1.866671098097183e-05,
"loss": 0.7805,
"step": 675
},
{
"epoch": 0.25173530772790376,
"grad_norm": 1.0447814725747497,
"learning_rate": 1.863428162388808e-05,
"loss": 0.8156,
"step": 680
},
{
"epoch": 0.25358630263766774,
"grad_norm": 1.122277511505867,
"learning_rate": 1.860149147895366e-05,
"loss": 0.7803,
"step": 685
},
{
"epoch": 0.2554372975474317,
"grad_norm": 1.1085629206709198,
"learning_rate": 1.856834191632144e-05,
"loss": 0.8045,
"step": 690
},
{
"epoch": 0.25728829245719576,
"grad_norm": 1.087490803470686,
"learning_rate": 1.8534834321162778e-05,
"loss": 0.822,
"step": 695
},
{
"epoch": 0.25913928736695974,
"grad_norm": 1.0114699174712363,
"learning_rate": 1.850097009360958e-05,
"loss": 0.7234,
"step": 700
},
{
"epoch": 0.25913928736695974,
"eval_loss": 0.7961795926094055,
"eval_runtime": 125.8984,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 700
},
{
"epoch": 0.2609902822767237,
"grad_norm": 1.1666771129689475,
"learning_rate": 1.8466750648695826e-05,
"loss": 0.8284,
"step": 705
},
{
"epoch": 0.26284127718648775,
"grad_norm": 1.0304043488373609,
"learning_rate": 1.843217741629843e-05,
"loss": 0.7025,
"step": 710
},
{
"epoch": 0.26469227209625174,
"grad_norm": 1.0509818064952419,
"learning_rate": 1.83972518410775e-05,
"loss": 0.739,
"step": 715
},
{
"epoch": 0.2665432670060157,
"grad_norm": 0.9979057999152209,
"learning_rate": 1.836197538241596e-05,
"loss": 0.7452,
"step": 720
},
{
"epoch": 0.26839426191577975,
"grad_norm": 1.0805505088606497,
"learning_rate": 1.8326349514358594e-05,
"loss": 0.7787,
"step": 725
},
{
"epoch": 0.27024525682554373,
"grad_norm": 1.091857948129845,
"learning_rate": 1.8290375725550417e-05,
"loss": 0.8327,
"step": 730
},
{
"epoch": 0.2720962517353077,
"grad_norm": 1.0581119370365102,
"learning_rate": 1.8254055519174502e-05,
"loss": 0.7103,
"step": 735
},
{
"epoch": 0.27394724664507175,
"grad_norm": 1.1460921794947827,
"learning_rate": 1.821739041288915e-05,
"loss": 0.7927,
"step": 740
},
{
"epoch": 0.27579824155483573,
"grad_norm": 0.9651613777166085,
"learning_rate": 1.818038193876448e-05,
"loss": 0.7406,
"step": 745
},
{
"epoch": 0.2776492364645997,
"grad_norm": 0.9502996970486759,
"learning_rate": 1.8143031643218413e-05,
"loss": 0.7027,
"step": 750
},
{
"epoch": 0.2795002313743637,
"grad_norm": 1.1463924114726602,
"learning_rate": 1.8105341086952052e-05,
"loss": 0.8102,
"step": 755
},
{
"epoch": 0.28135122628412773,
"grad_norm": 1.1234130601114334,
"learning_rate": 1.806731184488447e-05,
"loss": 0.7802,
"step": 760
},
{
"epoch": 0.2832022211938917,
"grad_norm": 1.0971581054230324,
"learning_rate": 1.8028945506086898e-05,
"loss": 0.7516,
"step": 765
},
{
"epoch": 0.2850532161036557,
"grad_norm": 1.107825346177917,
"learning_rate": 1.799024367371631e-05,
"loss": 0.7435,
"step": 770
},
{
"epoch": 0.28690421101341973,
"grad_norm": 1.1319764102322498,
"learning_rate": 1.795120796494848e-05,
"loss": 0.8346,
"step": 775
},
{
"epoch": 0.2887552059231837,
"grad_norm": 1.054146904567116,
"learning_rate": 1.791184001091035e-05,
"loss": 0.7668,
"step": 780
},
{
"epoch": 0.2906062008329477,
"grad_norm": 0.9546703867947438,
"learning_rate": 1.7872141456611904e-05,
"loss": 0.7729,
"step": 785
},
{
"epoch": 0.29245719574271173,
"grad_norm": 1.0520354517409871,
"learning_rate": 1.7832113960877445e-05,
"loss": 0.7666,
"step": 790
},
{
"epoch": 0.2943081906524757,
"grad_norm": 1.0339039958969705,
"learning_rate": 1.779175919627624e-05,
"loss": 0.7879,
"step": 795
},
{
"epoch": 0.2961591855622397,
"grad_norm": 1.0374069302809197,
"learning_rate": 1.775107884905266e-05,
"loss": 0.7693,
"step": 800
},
{
"epoch": 0.2961591855622397,
"eval_loss": 0.7911367416381836,
"eval_runtime": 125.8736,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 800
},
{
"epoch": 0.2980101804720037,
"grad_norm": 1.023031947891097,
"learning_rate": 1.7710074619055707e-05,
"loss": 0.7701,
"step": 805
},
{
"epoch": 0.2998611753817677,
"grad_norm": 1.02308459176092,
"learning_rate": 1.7668748219668007e-05,
"loss": 0.7079,
"step": 810
},
{
"epoch": 0.3017121702915317,
"grad_norm": 1.0193146042109342,
"learning_rate": 1.7627101377734176e-05,
"loss": 0.7814,
"step": 815
},
{
"epoch": 0.30356316520129567,
"grad_norm": 1.1175824255386086,
"learning_rate": 1.7585135833488692e-05,
"loss": 0.8352,
"step": 820
},
{
"epoch": 0.3054141601110597,
"grad_norm": 0.9758167149410084,
"learning_rate": 1.7542853340483175e-05,
"loss": 0.7525,
"step": 825
},
{
"epoch": 0.3072651550208237,
"grad_norm": 0.9971503389642156,
"learning_rate": 1.7500255665513112e-05,
"loss": 0.7279,
"step": 830
},
{
"epoch": 0.30911614993058767,
"grad_norm": 1.169901204414489,
"learning_rate": 1.7457344588544018e-05,
"loss": 0.7755,
"step": 835
},
{
"epoch": 0.3109671448403517,
"grad_norm": 1.0595523773603162,
"learning_rate": 1.7414121902637083e-05,
"loss": 0.7673,
"step": 840
},
{
"epoch": 0.3128181397501157,
"grad_norm": 1.1751687012571783,
"learning_rate": 1.7370589413874226e-05,
"loss": 0.741,
"step": 845
},
{
"epoch": 0.31466913465987967,
"grad_norm": 1.1835719361646306,
"learning_rate": 1.7326748941282638e-05,
"loss": 0.7788,
"step": 850
},
{
"epoch": 0.3165201295696437,
"grad_norm": 1.0111560919225886,
"learning_rate": 1.7282602316758774e-05,
"loss": 0.7678,
"step": 855
},
{
"epoch": 0.3183711244794077,
"grad_norm": 1.0835946252135704,
"learning_rate": 1.72381513849918e-05,
"loss": 0.7476,
"step": 860
},
{
"epoch": 0.32022211938917167,
"grad_norm": 1.1099985814534863,
"learning_rate": 1.7193398003386514e-05,
"loss": 0.779,
"step": 865
},
{
"epoch": 0.3220731142989357,
"grad_norm": 0.9066789182044317,
"learning_rate": 1.7148344041985736e-05,
"loss": 0.6901,
"step": 870
},
{
"epoch": 0.3239241092086997,
"grad_norm": 1.1422072479983323,
"learning_rate": 1.710299138339217e-05,
"loss": 0.806,
"step": 875
},
{
"epoch": 0.32577510411846367,
"grad_norm": 0.9742950243183045,
"learning_rate": 1.7057341922689733e-05,
"loss": 0.7739,
"step": 880
},
{
"epoch": 0.32762609902822765,
"grad_norm": 1.0227221722499922,
"learning_rate": 1.701139756736436e-05,
"loss": 0.7406,
"step": 885
},
{
"epoch": 0.3294770939379917,
"grad_norm": 1.0785612634689687,
"learning_rate": 1.696516023722431e-05,
"loss": 0.7728,
"step": 890
},
{
"epoch": 0.33132808884775566,
"grad_norm": 1.0880856661959288,
"learning_rate": 1.691863186431996e-05,
"loss": 0.7462,
"step": 895
},
{
"epoch": 0.33317908375751965,
"grad_norm": 1.0817764375919297,
"learning_rate": 1.6871814392863037e-05,
"loss": 0.8385,
"step": 900
},
{
"epoch": 0.33317908375751965,
"eval_loss": 0.7849360704421997,
"eval_runtime": 125.8514,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.509,
"step": 900
},
{
"epoch": 0.3350300786672837,
"grad_norm": 1.0848147512517947,
"learning_rate": 1.682470977914541e-05,
"loss": 0.7653,
"step": 905
},
{
"epoch": 0.33688107357704766,
"grad_norm": 1.1609482279651488,
"learning_rate": 1.6777319991457325e-05,
"loss": 0.786,
"step": 910
},
{
"epoch": 0.33873206848681164,
"grad_norm": 1.2171008251999773,
"learning_rate": 1.6729647010005175e-05,
"loss": 0.8124,
"step": 915
},
{
"epoch": 0.3405830633965757,
"grad_norm": 0.9446888141489528,
"learning_rate": 1.6681692826828743e-05,
"loss": 0.8249,
"step": 920
},
{
"epoch": 0.34243405830633966,
"grad_norm": 1.099582146653851,
"learning_rate": 1.6633459445717973e-05,
"loss": 0.7901,
"step": 925
},
{
"epoch": 0.34428505321610364,
"grad_norm": 0.9771996099399858,
"learning_rate": 1.6584948882129238e-05,
"loss": 0.759,
"step": 930
},
{
"epoch": 0.3461360481258677,
"grad_norm": 1.0185009245061947,
"learning_rate": 1.653616316310112e-05,
"loss": 0.7295,
"step": 935
},
{
"epoch": 0.34798704303563166,
"grad_norm": 1.0804952434925557,
"learning_rate": 1.6487104327169702e-05,
"loss": 0.755,
"step": 940
},
{
"epoch": 0.34983803794539564,
"grad_norm": 1.0894047826173676,
"learning_rate": 1.6437774424283414e-05,
"loss": 0.7916,
"step": 945
},
{
"epoch": 0.3516890328551596,
"grad_norm": 0.9805181725379623,
"learning_rate": 1.6388175515717336e-05,
"loss": 0.7285,
"step": 950
},
{
"epoch": 0.35354002776492366,
"grad_norm": 1.356858685263938,
"learning_rate": 1.63383096739871e-05,
"loss": 0.7864,
"step": 955
},
{
"epoch": 0.35539102267468764,
"grad_norm": 0.9525556743311706,
"learning_rate": 1.6288178982762287e-05,
"loss": 0.742,
"step": 960
},
{
"epoch": 0.3572420175844516,
"grad_norm": 1.0831204564195782,
"learning_rate": 1.6237785536779322e-05,
"loss": 0.7386,
"step": 965
},
{
"epoch": 0.35909301249421566,
"grad_norm": 0.9806687465833773,
"learning_rate": 1.618713144175399e-05,
"loss": 0.7399,
"step": 970
},
{
"epoch": 0.36094400740397964,
"grad_norm": 0.9492684447869085,
"learning_rate": 1.6136218814293422e-05,
"loss": 0.763,
"step": 975
},
{
"epoch": 0.3627950023137436,
"grad_norm": 1.05964205886519,
"learning_rate": 1.6085049781807656e-05,
"loss": 0.7658,
"step": 980
},
{
"epoch": 0.36464599722350766,
"grad_norm": 1.007330323233069,
"learning_rate": 1.603362648242076e-05,
"loss": 0.7371,
"step": 985
},
{
"epoch": 0.36649699213327164,
"grad_norm": 0.9531551141111623,
"learning_rate": 1.5981951064881456e-05,
"loss": 0.7538,
"step": 990
},
{
"epoch": 0.3683479870430356,
"grad_norm": 1.0737293889710073,
"learning_rate": 1.5930025688473353e-05,
"loss": 0.7896,
"step": 995
},
{
"epoch": 0.37019898195279966,
"grad_norm": 1.0915409073010398,
"learning_rate": 1.5877852522924733e-05,
"loss": 0.7926,
"step": 1000
},
{
"epoch": 0.37019898195279966,
"eval_loss": 0.7791726589202881,
"eval_runtime": 125.9827,
"eval_samples_per_second": 1.016,
"eval_steps_per_second": 0.508,
"step": 1000
},
{
"epoch": 0.37204997686256364,
"grad_norm": 1.1973665058190375,
"learning_rate": 1.5825433748317857e-05,
"loss": 0.7215,
"step": 1005
},
{
"epoch": 0.3739009717723276,
"grad_norm": 1.0701603196497034,
"learning_rate": 1.577277155499789e-05,
"loss": 0.7396,
"step": 1010
},
{
"epoch": 0.3757519666820916,
"grad_norm": 1.0987041230123638,
"learning_rate": 1.5719868143481385e-05,
"loss": 0.7429,
"step": 1015
},
{
"epoch": 0.37760296159185563,
"grad_norm": 1.0534343239222719,
"learning_rate": 1.5666725724364296e-05,
"loss": 0.7545,
"step": 1020
},
{
"epoch": 0.3794539565016196,
"grad_norm": 1.028797489541622,
"learning_rate": 1.5613346518229652e-05,
"loss": 0.7264,
"step": 1025
},
{
"epoch": 0.3813049514113836,
"grad_norm": 1.0217316773582072,
"learning_rate": 1.5559732755554734e-05,
"loss": 0.7177,
"step": 1030
},
{
"epoch": 0.38315594632114763,
"grad_norm": 1.047042319865492,
"learning_rate": 1.5505886676617902e-05,
"loss": 0.7714,
"step": 1035
},
{
"epoch": 0.3850069412309116,
"grad_norm": 1.0810668413004525,
"learning_rate": 1.545181053140495e-05,
"loss": 0.715,
"step": 1040
},
{
"epoch": 0.3868579361406756,
"grad_norm": 1.0363170821457612,
"learning_rate": 1.539750657951513e-05,
"loss": 0.7664,
"step": 1045
},
{
"epoch": 0.38870893105043963,
"grad_norm": 1.0452365868797338,
"learning_rate": 1.5342977090066696e-05,
"loss": 0.7236,
"step": 1050
},
{
"epoch": 0.3905599259602036,
"grad_norm": 0.9982499977285321,
"learning_rate": 1.52882243416021e-05,
"loss": 0.7731,
"step": 1055
},
{
"epoch": 0.3924109208699676,
"grad_norm": 1.030850130209849,
"learning_rate": 1.523325062199281e-05,
"loss": 0.7617,
"step": 1060
},
{
"epoch": 0.39426191577973163,
"grad_norm": 1.1101345309616182,
"learning_rate": 1.5178058228343662e-05,
"loss": 0.7552,
"step": 1065
},
{
"epoch": 0.3961129106894956,
"grad_norm": 0.9346869642461403,
"learning_rate": 1.5122649466896899e-05,
"loss": 0.7606,
"step": 1070
},
{
"epoch": 0.3979639055992596,
"grad_norm": 0.9976085317381093,
"learning_rate": 1.5067026652935823e-05,
"loss": 0.7303,
"step": 1075
},
{
"epoch": 0.3998149005090236,
"grad_norm": 0.8951591117484915,
"learning_rate": 1.5011192110688008e-05,
"loss": 0.6998,
"step": 1080
},
{
"epoch": 0.4016658954187876,
"grad_norm": 1.0337985910400393,
"learning_rate": 1.4955148173228209e-05,
"loss": 0.7814,
"step": 1085
},
{
"epoch": 0.4035168903285516,
"grad_norm": 1.0299336976844324,
"learning_rate": 1.4898897182380872e-05,
"loss": 0.7532,
"step": 1090
},
{
"epoch": 0.4053678852383156,
"grad_norm": 0.9541582795109266,
"learning_rate": 1.4842441488622266e-05,
"loss": 0.7609,
"step": 1095
},
{
"epoch": 0.4072188801480796,
"grad_norm": 1.1428468537841259,
"learning_rate": 1.4785783450982276e-05,
"loss": 0.7463,
"step": 1100
},
{
"epoch": 0.4072188801480796,
"eval_loss": 0.7768411636352539,
"eval_runtime": 125.9173,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 1100
},
{
"epoch": 0.4090698750578436,
"grad_norm": 0.9432900818713729,
"learning_rate": 1.4728925436945838e-05,
"loss": 0.7225,
"step": 1105
},
{
"epoch": 0.41092086996760757,
"grad_norm": 0.9568114894842158,
"learning_rate": 1.4671869822353993e-05,
"loss": 0.7414,
"step": 1110
},
{
"epoch": 0.4127718648773716,
"grad_norm": 1.1008061214147924,
"learning_rate": 1.4614618991304625e-05,
"loss": 0.78,
"step": 1115
},
{
"epoch": 0.4146228597871356,
"grad_norm": 1.0503790052814121,
"learning_rate": 1.4557175336052844e-05,
"loss": 0.7223,
"step": 1120
},
{
"epoch": 0.41647385469689957,
"grad_norm": 0.9583451918719977,
"learning_rate": 1.4499541256911008e-05,
"loss": 0.7495,
"step": 1125
},
{
"epoch": 0.4183248496066636,
"grad_norm": 1.0275134379520088,
"learning_rate": 1.4441719162148433e-05,
"loss": 0.7653,
"step": 1130
},
{
"epoch": 0.4201758445164276,
"grad_norm": 1.0425892623904864,
"learning_rate": 1.4383711467890776e-05,
"loss": 0.7055,
"step": 1135
},
{
"epoch": 0.42202683942619157,
"grad_norm": 1.069154922576832,
"learning_rate": 1.4325520598019049e-05,
"loss": 0.7837,
"step": 1140
},
{
"epoch": 0.42387783433595555,
"grad_norm": 0.9896521412449755,
"learning_rate": 1.4267148984068359e-05,
"loss": 0.7702,
"step": 1145
},
{
"epoch": 0.4257288292457196,
"grad_norm": 1.025563581357004,
"learning_rate": 1.4208599065126292e-05,
"loss": 0.7604,
"step": 1150
},
{
"epoch": 0.42757982415548357,
"grad_norm": 1.001282739654851,
"learning_rate": 1.4149873287731005e-05,
"loss": 0.7874,
"step": 1155
},
{
"epoch": 0.42943081906524755,
"grad_norm": 1.0858912388862354,
"learning_rate": 1.4090974105768977e-05,
"loss": 0.8066,
"step": 1160
},
{
"epoch": 0.4312818139750116,
"grad_norm": 0.9891342220670727,
"learning_rate": 1.4031903980372503e-05,
"loss": 0.7532,
"step": 1165
},
{
"epoch": 0.43313280888477557,
"grad_norm": 1.1089360136811275,
"learning_rate": 1.3972665379816821e-05,
"loss": 0.7344,
"step": 1170
},
{
"epoch": 0.43498380379453955,
"grad_norm": 1.0804929458235695,
"learning_rate": 1.3913260779416999e-05,
"loss": 0.7127,
"step": 1175
},
{
"epoch": 0.4368347987043036,
"grad_norm": 0.9487754961773092,
"learning_rate": 1.3853692661424485e-05,
"loss": 0.7544,
"step": 1180
},
{
"epoch": 0.43868579361406757,
"grad_norm": 1.1689376731656378,
"learning_rate": 1.3793963514923398e-05,
"loss": 0.7571,
"step": 1185
},
{
"epoch": 0.44053678852383155,
"grad_norm": 1.1079751308984782,
"learning_rate": 1.3734075835726515e-05,
"loss": 0.7632,
"step": 1190
},
{
"epoch": 0.4423877834335956,
"grad_norm": 0.9701136035154755,
"learning_rate": 1.3674032126270982e-05,
"loss": 0.7155,
"step": 1195
},
{
"epoch": 0.44423877834335956,
"grad_norm": 1.0783370064152846,
"learning_rate": 1.3613834895513748e-05,
"loss": 0.8017,
"step": 1200
},
{
"epoch": 0.44423877834335956,
"eval_loss": 0.7741294503211975,
"eval_runtime": 125.9108,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 1200
},
{
"epoch": 0.44608977325312354,
"grad_norm": 1.039233538517332,
"learning_rate": 1.355348665882673e-05,
"loss": 0.7332,
"step": 1205
},
{
"epoch": 0.4479407681628875,
"grad_norm": 1.1306337428774784,
"learning_rate": 1.3492989937891694e-05,
"loss": 0.7475,
"step": 1210
},
{
"epoch": 0.44979176307265156,
"grad_norm": 1.0587177357723303,
"learning_rate": 1.3432347260594911e-05,
"loss": 0.7603,
"step": 1215
},
{
"epoch": 0.45164275798241554,
"grad_norm": 1.0290685165693072,
"learning_rate": 1.3371561160921507e-05,
"loss": 0.7646,
"step": 1220
},
{
"epoch": 0.4534937528921795,
"grad_norm": 1.34640973701702,
"learning_rate": 1.3310634178849583e-05,
"loss": 0.6996,
"step": 1225
},
{
"epoch": 0.45534474780194356,
"grad_norm": 0.9941580643005229,
"learning_rate": 1.3249568860244081e-05,
"loss": 0.7254,
"step": 1230
},
{
"epoch": 0.45719574271170754,
"grad_norm": 1.0409471890052142,
"learning_rate": 1.3188367756750413e-05,
"loss": 0.7774,
"step": 1235
},
{
"epoch": 0.4590467376214715,
"grad_norm": 0.9723468054375276,
"learning_rate": 1.312703342568782e-05,
"loss": 0.7169,
"step": 1240
},
{
"epoch": 0.46089773253123556,
"grad_norm": 0.9902905902420863,
"learning_rate": 1.306556842994254e-05,
"loss": 0.7297,
"step": 1245
},
{
"epoch": 0.46274872744099954,
"grad_norm": 0.9601290745180161,
"learning_rate": 1.3003975337860686e-05,
"loss": 0.7538,
"step": 1250
},
{
"epoch": 0.4645997223507635,
"grad_norm": 0.9053050963379172,
"learning_rate": 1.2942256723140951e-05,
"loss": 0.7109,
"step": 1255
},
{
"epoch": 0.46645071726052756,
"grad_norm": 1.0874571881287778,
"learning_rate": 1.2880415164727058e-05,
"loss": 0.7919,
"step": 1260
},
{
"epoch": 0.46830171217029154,
"grad_norm": 1.1451121612817037,
"learning_rate": 1.2818453246699981e-05,
"loss": 0.7391,
"step": 1265
},
{
"epoch": 0.4701527070800555,
"grad_norm": 0.9681114397766631,
"learning_rate": 1.2756373558169992e-05,
"loss": 0.743,
"step": 1270
},
{
"epoch": 0.4720037019898195,
"grad_norm": 0.9223359776123135,
"learning_rate": 1.2694178693168468e-05,
"loss": 0.72,
"step": 1275
},
{
"epoch": 0.47385469689958354,
"grad_norm": 1.0469506863925069,
"learning_rate": 1.2631871250539478e-05,
"loss": 0.728,
"step": 1280
},
{
"epoch": 0.4757056918093475,
"grad_norm": 1.0014451232715824,
"learning_rate": 1.2569453833831222e-05,
"loss": 0.717,
"step": 1285
},
{
"epoch": 0.4775566867191115,
"grad_norm": 0.9639634997081045,
"learning_rate": 1.250692905118721e-05,
"loss": 0.7507,
"step": 1290
},
{
"epoch": 0.47940768162887554,
"grad_norm": 0.9929137343346536,
"learning_rate": 1.2444299515237298e-05,
"loss": 0.7643,
"step": 1295
},
{
"epoch": 0.4812586765386395,
"grad_norm": 1.0240687855090735,
"learning_rate": 1.238156784298851e-05,
"loss": 0.7072,
"step": 1300
},
{
"epoch": 0.4812586765386395,
"eval_loss": 0.7685373425483704,
"eval_runtime": 125.8483,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.509,
"step": 1300
},
{
"epoch": 0.4831096714484035,
"grad_norm": 1.0317937768510643,
"learning_rate": 1.2318736655715689e-05,
"loss": 0.7638,
"step": 1305
},
{
"epoch": 0.48496066635816754,
"grad_norm": 0.9255922281298357,
"learning_rate": 1.2255808578851958e-05,
"loss": 0.7467,
"step": 1310
},
{
"epoch": 0.4868116612679315,
"grad_norm": 0.9721165058296256,
"learning_rate": 1.2192786241879033e-05,
"loss": 0.7546,
"step": 1315
},
{
"epoch": 0.4886626561776955,
"grad_norm": 0.9644367734588111,
"learning_rate": 1.2129672278217328e-05,
"loss": 0.7169,
"step": 1320
},
{
"epoch": 0.49051365108745953,
"grad_norm": 1.0130538871266401,
"learning_rate": 1.2066469325115927e-05,
"loss": 0.7369,
"step": 1325
},
{
"epoch": 0.4923646459972235,
"grad_norm": 0.9834288208571581,
"learning_rate": 1.2003180023542375e-05,
"loss": 0.7325,
"step": 1330
},
{
"epoch": 0.4942156409069875,
"grad_norm": 0.960205484882582,
"learning_rate": 1.1939807018072345e-05,
"loss": 0.7277,
"step": 1335
},
{
"epoch": 0.4960666358167515,
"grad_norm": 1.000713485591568,
"learning_rate": 1.1876352956779114e-05,
"loss": 0.7671,
"step": 1340
},
{
"epoch": 0.4979176307265155,
"grad_norm": 0.9972457796952376,
"learning_rate": 1.1812820491122918e-05,
"loss": 0.7496,
"step": 1345
},
{
"epoch": 0.4997686256362795,
"grad_norm": 0.9176925753783508,
"learning_rate": 1.1749212275840159e-05,
"loss": 0.7719,
"step": 1350
},
{
"epoch": 0.5016196205460435,
"grad_norm": 0.9486097980237569,
"learning_rate": 1.1685530968832474e-05,
"loss": 0.7286,
"step": 1355
},
{
"epoch": 0.5034706154558075,
"grad_norm": 1.0669908596524535,
"learning_rate": 1.1621779231055677e-05,
"loss": 0.7279,
"step": 1360
},
{
"epoch": 0.5053216103655715,
"grad_norm": 1.0656718017579419,
"learning_rate": 1.1557959726408567e-05,
"loss": 0.7338,
"step": 1365
},
{
"epoch": 0.5071726052753355,
"grad_norm": 1.2390741577593356,
"learning_rate": 1.1494075121621618e-05,
"loss": 0.7997,
"step": 1370
},
{
"epoch": 0.5090236001850995,
"grad_norm": 1.0260656127221468,
"learning_rate": 1.1430128086145542e-05,
"loss": 0.7257,
"step": 1375
},
{
"epoch": 0.5108745950948634,
"grad_norm": 0.981403662935476,
"learning_rate": 1.1366121292039756e-05,
"loss": 0.6885,
"step": 1380
},
{
"epoch": 0.5127255900046275,
"grad_norm": 0.9506854040681342,
"learning_rate": 1.1302057413860717e-05,
"loss": 0.7523,
"step": 1385
},
{
"epoch": 0.5145765849143915,
"grad_norm": 0.9571399024989548,
"learning_rate": 1.1237939128550167e-05,
"loss": 0.7406,
"step": 1390
},
{
"epoch": 0.5164275798241554,
"grad_norm": 0.9379988174297818,
"learning_rate": 1.1173769115323276e-05,
"loss": 0.7152,
"step": 1395
},
{
"epoch": 0.5182785747339195,
"grad_norm": 0.9799140133111384,
"learning_rate": 1.1109550055556704e-05,
"loss": 0.6862,
"step": 1400
},
{
"epoch": 0.5182785747339195,
"eval_loss": 0.762523889541626,
"eval_runtime": 125.8824,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 1400
},
{
"epoch": 0.5201295696436835,
"grad_norm": 1.0128868126138035,
"learning_rate": 1.1045284632676535e-05,
"loss": 0.7491,
"step": 1405
},
{
"epoch": 0.5219805645534474,
"grad_norm": 1.1054871656808112,
"learning_rate": 1.098097553204616e-05,
"loss": 0.7193,
"step": 1410
},
{
"epoch": 0.5238315594632115,
"grad_norm": 1.0534251043687004,
"learning_rate": 1.0916625440854074e-05,
"loss": 0.7263,
"step": 1415
},
{
"epoch": 0.5256825543729755,
"grad_norm": 1.0586326058499738,
"learning_rate": 1.0852237048001568e-05,
"loss": 0.771,
"step": 1420
},
{
"epoch": 0.5275335492827394,
"grad_norm": 1.0689916226823393,
"learning_rate": 1.0787813043990405e-05,
"loss": 0.7936,
"step": 1425
},
{
"epoch": 0.5293845441925035,
"grad_norm": 0.9447698094104972,
"learning_rate": 1.0723356120810367e-05,
"loss": 0.7594,
"step": 1430
},
{
"epoch": 0.5312355391022675,
"grad_norm": 1.0832488716625996,
"learning_rate": 1.0658868971826785e-05,
"loss": 0.7371,
"step": 1435
},
{
"epoch": 0.5330865340120314,
"grad_norm": 1.0138170678877134,
"learning_rate": 1.0594354291667995e-05,
"loss": 0.7389,
"step": 1440
},
{
"epoch": 0.5349375289217955,
"grad_norm": 1.0133195189163617,
"learning_rate": 1.0529814776112724e-05,
"loss": 0.7443,
"step": 1445
},
{
"epoch": 0.5367885238315595,
"grad_norm": 1.062091180379715,
"learning_rate": 1.046525312197747e-05,
"loss": 0.7549,
"step": 1450
},
{
"epoch": 0.5386395187413234,
"grad_norm": 0.9945798468589808,
"learning_rate": 1.0400672027003795e-05,
"loss": 0.7033,
"step": 1455
},
{
"epoch": 0.5404905136510875,
"grad_norm": 1.0798176885934982,
"learning_rate": 1.0336074189745617e-05,
"loss": 0.7873,
"step": 1460
},
{
"epoch": 0.5423415085608515,
"grad_norm": 0.949930038696299,
"learning_rate": 1.027146230945643e-05,
"loss": 0.7655,
"step": 1465
},
{
"epoch": 0.5441925034706154,
"grad_norm": 0.9128607712688908,
"learning_rate": 1.0206839085976528e-05,
"loss": 0.7045,
"step": 1470
},
{
"epoch": 0.5460434983803795,
"grad_norm": 0.9894531926372522,
"learning_rate": 1.014220721962018e-05,
"loss": 0.7288,
"step": 1475
},
{
"epoch": 0.5478944932901435,
"grad_norm": 1.0508653514368567,
"learning_rate": 1.0077569411062804e-05,
"loss": 0.7393,
"step": 1480
},
{
"epoch": 0.5497454881999074,
"grad_norm": 1.0144705488797956,
"learning_rate": 1.001292836122812e-05,
"loss": 0.7388,
"step": 1485
},
{
"epoch": 0.5515964831096715,
"grad_norm": 1.0862811301549016,
"learning_rate": 9.948286771175288e-06,
"loss": 0.7666,
"step": 1490
},
{
"epoch": 0.5534474780194355,
"grad_norm": 0.9773752334499499,
"learning_rate": 9.883647341986032e-06,
"loss": 0.7572,
"step": 1495
},
{
"epoch": 0.5552984729291994,
"grad_norm": 0.965789789651361,
"learning_rate": 9.819012774651788e-06,
"loss": 0.7466,
"step": 1500
},
{
"epoch": 0.5552984729291994,
"eval_loss": 0.759286105632782,
"eval_runtime": 125.877,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 1500
},
{
"epoch": 0.5571494678389635,
"grad_norm": 1.0286512648148671,
"learning_rate": 9.75438576996084e-06,
"loss": 0.7419,
"step": 1505
},
{
"epoch": 0.5590004627487274,
"grad_norm": 0.9779662642913994,
"learning_rate": 9.689769028385463e-06,
"loss": 0.7594,
"step": 1510
},
{
"epoch": 0.5608514576584914,
"grad_norm": 0.9878313389871681,
"learning_rate": 9.625165249969075e-06,
"loss": 0.7236,
"step": 1515
},
{
"epoch": 0.5627024525682555,
"grad_norm": 1.020571296588044,
"learning_rate": 9.56057713421343e-06,
"loss": 0.7498,
"step": 1520
},
{
"epoch": 0.5645534474780194,
"grad_norm": 0.9995998586225845,
"learning_rate": 9.496007379965801e-06,
"loss": 0.7606,
"step": 1525
},
{
"epoch": 0.5664044423877834,
"grad_norm": 1.060622141268455,
"learning_rate": 9.431458685306227e-06,
"loss": 0.6984,
"step": 1530
},
{
"epoch": 0.5682554372975475,
"grad_norm": 0.9493673992150904,
"learning_rate": 9.366933747434758e-06,
"loss": 0.6911,
"step": 1535
},
{
"epoch": 0.5701064322073114,
"grad_norm": 1.0167667579449835,
"learning_rate": 9.302435262558748e-06,
"loss": 0.6778,
"step": 1540
},
{
"epoch": 0.5719574271170754,
"grad_norm": 1.0392115654580312,
"learning_rate": 9.237965925780207e-06,
"loss": 0.7691,
"step": 1545
},
{
"epoch": 0.5738084220268395,
"grad_norm": 1.0593025980336248,
"learning_rate": 9.173528430983167e-06,
"loss": 0.738,
"step": 1550
},
{
"epoch": 0.5756594169366034,
"grad_norm": 0.971968443552632,
"learning_rate": 9.109125470721141e-06,
"loss": 0.7421,
"step": 1555
},
{
"epoch": 0.5775104118463674,
"grad_norm": 0.932647291828516,
"learning_rate": 9.044759736104584e-06,
"loss": 0.7055,
"step": 1560
},
{
"epoch": 0.5793614067561315,
"grad_norm": 0.8632798438231859,
"learning_rate": 8.98043391668847e-06,
"loss": 0.6961,
"step": 1565
},
{
"epoch": 0.5812124016658954,
"grad_norm": 0.9441485921057498,
"learning_rate": 8.916150700359896e-06,
"loss": 0.7468,
"step": 1570
},
{
"epoch": 0.5830633965756594,
"grad_norm": 0.9939474443131038,
"learning_rate": 8.85191277322575e-06,
"loss": 0.7149,
"step": 1575
},
{
"epoch": 0.5849143914854235,
"grad_norm": 1.0783126296675078,
"learning_rate": 8.787722819500513e-06,
"loss": 0.7135,
"step": 1580
},
{
"epoch": 0.5867653863951874,
"grad_norm": 0.9705292869514378,
"learning_rate": 8.723583521394054e-06,
"loss": 0.7109,
"step": 1585
},
{
"epoch": 0.5886163813049514,
"grad_norm": 0.9367179465636712,
"learning_rate": 8.659497558999579e-06,
"loss": 0.7114,
"step": 1590
},
{
"epoch": 0.5904673762147155,
"grad_norm": 1.0408124857679764,
"learning_rate": 8.595467610181638e-06,
"loss": 0.7161,
"step": 1595
},
{
"epoch": 0.5923183711244794,
"grad_norm": 0.971797951572304,
"learning_rate": 8.53149635046421e-06,
"loss": 0.7508,
"step": 1600
},
{
"epoch": 0.5923183711244794,
"eval_loss": 0.7539975643157959,
"eval_runtime": 125.8675,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 1600
},
{
"epoch": 0.5941693660342434,
"grad_norm": 0.9800435183445545,
"learning_rate": 8.467586452918942e-06,
"loss": 0.715,
"step": 1605
},
{
"epoch": 0.5960203609440075,
"grad_norm": 1.0113663434618827,
"learning_rate": 8.403740588053408e-06,
"loss": 0.6989,
"step": 1610
},
{
"epoch": 0.5978713558537714,
"grad_norm": 1.0355947843868953,
"learning_rate": 8.339961423699563e-06,
"loss": 0.7193,
"step": 1615
},
{
"epoch": 0.5997223507635354,
"grad_norm": 1.0175465647016573,
"learning_rate": 8.276251624902235e-06,
"loss": 0.7059,
"step": 1620
},
{
"epoch": 0.6015733456732995,
"grad_norm": 1.0190101018280562,
"learning_rate": 8.212613853807783e-06,
"loss": 0.7302,
"step": 1625
},
{
"epoch": 0.6034243405830634,
"grad_norm": 1.053793043329184,
"learning_rate": 8.149050769552856e-06,
"loss": 0.7198,
"step": 1630
},
{
"epoch": 0.6052753354928274,
"grad_norm": 1.0096414981283535,
"learning_rate": 8.085565028153261e-06,
"loss": 0.7581,
"step": 1635
},
{
"epoch": 0.6071263304025913,
"grad_norm": 0.9812300345845569,
"learning_rate": 8.02215928239301e-06,
"loss": 0.7472,
"step": 1640
},
{
"epoch": 0.6089773253123554,
"grad_norm": 1.0007016748169104,
"learning_rate": 7.958836181713445e-06,
"loss": 0.7582,
"step": 1645
},
{
"epoch": 0.6108283202221194,
"grad_norm": 0.9909841301661106,
"learning_rate": 7.895598372102547e-06,
"loss": 0.7247,
"step": 1650
},
{
"epoch": 0.6126793151318833,
"grad_norm": 1.0341673384965404,
"learning_rate": 7.832448495984368e-06,
"loss": 0.6798,
"step": 1655
},
{
"epoch": 0.6145303100416474,
"grad_norm": 0.9331106658706707,
"learning_rate": 7.769389192108608e-06,
"loss": 0.7123,
"step": 1660
},
{
"epoch": 0.6163813049514114,
"grad_norm": 0.9435545677177092,
"learning_rate": 7.706423095440367e-06,
"loss": 0.7243,
"step": 1665
},
{
"epoch": 0.6182322998611753,
"grad_norm": 0.991710730848484,
"learning_rate": 7.643552837050026e-06,
"loss": 0.7433,
"step": 1670
},
{
"epoch": 0.6200832947709394,
"grad_norm": 1.0668190323716082,
"learning_rate": 7.580781044003324e-06,
"loss": 0.7375,
"step": 1675
},
{
"epoch": 0.6219342896807034,
"grad_norm": 1.0279605405376717,
"learning_rate": 7.518110339251568e-06,
"loss": 0.7234,
"step": 1680
},
{
"epoch": 0.6237852845904673,
"grad_norm": 0.9640733339122612,
"learning_rate": 7.455543341522042e-06,
"loss": 0.7168,
"step": 1685
},
{
"epoch": 0.6256362795002314,
"grad_norm": 0.9907817877992142,
"learning_rate": 7.393082665208587e-06,
"loss": 0.7053,
"step": 1690
},
{
"epoch": 0.6274872744099954,
"grad_norm": 1.0177580646696514,
"learning_rate": 7.33073092026233e-06,
"loss": 0.7772,
"step": 1695
},
{
"epoch": 0.6293382693197593,
"grad_norm": 1.0511047158796323,
"learning_rate": 7.268490712082667e-06,
"loss": 0.7027,
"step": 1700
},
{
"epoch": 0.6293382693197593,
"eval_loss": 0.7508743405342102,
"eval_runtime": 125.8608,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 1700
},
{
"epoch": 0.6311892642295234,
"grad_norm": 0.9786722954600374,
"learning_rate": 7.206364641408358e-06,
"loss": 0.754,
"step": 1705
},
{
"epoch": 0.6330402591392874,
"grad_norm": 1.001046944711236,
"learning_rate": 7.144355304208877e-06,
"loss": 0.7351,
"step": 1710
},
{
"epoch": 0.6348912540490513,
"grad_norm": 0.9068727644526208,
"learning_rate": 7.082465291575928e-06,
"loss": 0.7213,
"step": 1715
},
{
"epoch": 0.6367422489588154,
"grad_norm": 0.9817053840255833,
"learning_rate": 7.02069718961518e-06,
"loss": 0.7405,
"step": 1720
},
{
"epoch": 0.6385932438685794,
"grad_norm": 0.9963090460429577,
"learning_rate": 6.959053579338202e-06,
"loss": 0.7329,
"step": 1725
},
{
"epoch": 0.6404442387783433,
"grad_norm": 1.0063439847490308,
"learning_rate": 6.8975370365546055e-06,
"loss": 0.7466,
"step": 1730
},
{
"epoch": 0.6422952336881074,
"grad_norm": 1.0756259763530167,
"learning_rate": 6.836150131764434e-06,
"loss": 0.7328,
"step": 1735
},
{
"epoch": 0.6441462285978714,
"grad_norm": 1.0006380504007437,
"learning_rate": 6.7748954300507405e-06,
"loss": 0.7291,
"step": 1740
},
{
"epoch": 0.6459972235076353,
"grad_norm": 1.0476443249218472,
"learning_rate": 6.713775490972396e-06,
"loss": 0.722,
"step": 1745
},
{
"epoch": 0.6478482184173994,
"grad_norm": 0.9299325627763796,
"learning_rate": 6.652792868457159e-06,
"loss": 0.7259,
"step": 1750
},
{
"epoch": 0.6496992133271634,
"grad_norm": 1.0107437601362794,
"learning_rate": 6.591950110694928e-06,
"loss": 0.7313,
"step": 1755
},
{
"epoch": 0.6515502082369273,
"grad_norm": 1.0116149889736052,
"learning_rate": 6.531249760031304e-06,
"loss": 0.7495,
"step": 1760
},
{
"epoch": 0.6534012031466914,
"grad_norm": 1.060084766737352,
"learning_rate": 6.4706943528613135e-06,
"loss": 0.7406,
"step": 1765
},
{
"epoch": 0.6552521980564553,
"grad_norm": 0.9893182231305682,
"learning_rate": 6.410286419523457e-06,
"loss": 0.7236,
"step": 1770
},
{
"epoch": 0.6571031929662193,
"grad_norm": 1.0249584716250184,
"learning_rate": 6.350028484193971e-06,
"loss": 0.7735,
"step": 1775
},
{
"epoch": 0.6589541878759834,
"grad_norm": 0.9922083712983304,
"learning_rate": 6.2899230647813315e-06,
"loss": 0.7092,
"step": 1780
},
{
"epoch": 0.6608051827857473,
"grad_norm": 1.079323750602494,
"learning_rate": 6.229972672821081e-06,
"loss": 0.6999,
"step": 1785
},
{
"epoch": 0.6626561776955113,
"grad_norm": 0.9678468916444759,
"learning_rate": 6.170179813370838e-06,
"loss": 0.7215,
"step": 1790
},
{
"epoch": 0.6645071726052754,
"grad_norm": 0.9400337665409387,
"learning_rate": 6.110546984905661e-06,
"loss": 0.7189,
"step": 1795
},
{
"epoch": 0.6663581675150393,
"grad_norm": 0.9282443073802316,
"learning_rate": 6.051076679213632e-06,
"loss": 0.7228,
"step": 1800
},
{
"epoch": 0.6663581675150393,
"eval_loss": 0.7472089529037476,
"eval_runtime": 125.9422,
"eval_samples_per_second": 1.016,
"eval_steps_per_second": 0.508,
"step": 1800
},
{
"epoch": 0.6682091624248033,
"grad_norm": 0.9554957232923886,
"learning_rate": 5.991771381291727e-06,
"loss": 0.7502,
"step": 1805
},
{
"epoch": 0.6700601573345674,
"grad_norm": 1.0695199061812564,
"learning_rate": 5.932633569242e-06,
"loss": 0.7725,
"step": 1810
},
{
"epoch": 0.6719111522443313,
"grad_norm": 0.9775812687514566,
"learning_rate": 5.8736657141680066e-06,
"loss": 0.7201,
"step": 1815
},
{
"epoch": 0.6737621471540953,
"grad_norm": 1.0498313112511333,
"learning_rate": 5.814870280071581e-06,
"loss": 0.7066,
"step": 1820
},
{
"epoch": 0.6756131420638594,
"grad_norm": 0.8996562202890858,
"learning_rate": 5.756249723749847e-06,
"loss": 0.6839,
"step": 1825
},
{
"epoch": 0.6774641369736233,
"grad_norm": 0.9343357642622071,
"learning_rate": 5.697806494692575e-06,
"loss": 0.7006,
"step": 1830
},
{
"epoch": 0.6793151318833873,
"grad_norm": 0.9122261783600852,
"learning_rate": 5.6395430349798376e-06,
"loss": 0.71,
"step": 1835
},
{
"epoch": 0.6811661267931514,
"grad_norm": 1.0003802308196195,
"learning_rate": 5.581461779179924e-06,
"loss": 0.7352,
"step": 1840
},
{
"epoch": 0.6830171217029153,
"grad_norm": 0.9577445500508659,
"learning_rate": 5.5235651542476745e-06,
"loss": 0.7187,
"step": 1845
},
{
"epoch": 0.6848681166126793,
"grad_norm": 0.9048825285696591,
"learning_rate": 5.465855579423012e-06,
"loss": 0.6786,
"step": 1850
},
{
"epoch": 0.6867191115224434,
"grad_norm": 0.9658805694805159,
"learning_rate": 5.4083354661298816e-06,
"loss": 0.7012,
"step": 1855
},
{
"epoch": 0.6885701064322073,
"grad_norm": 0.9198866529754813,
"learning_rate": 5.351007217875493e-06,
"loss": 0.7193,
"step": 1860
},
{
"epoch": 0.6904211013419713,
"grad_norm": 0.9860185880901673,
"learning_rate": 5.293873230149851e-06,
"loss": 0.6928,
"step": 1865
},
{
"epoch": 0.6922720962517354,
"grad_norm": 0.9158051945254762,
"learning_rate": 5.236935890325717e-06,
"loss": 0.7283,
"step": 1870
},
{
"epoch": 0.6941230911614993,
"grad_norm": 1.0229541633123165,
"learning_rate": 5.180197577558792e-06,
"loss": 0.6934,
"step": 1875
},
{
"epoch": 0.6959740860712633,
"grad_norm": 1.058766779890904,
"learning_rate": 5.123660662688352e-06,
"loss": 0.7485,
"step": 1880
},
{
"epoch": 0.6978250809810274,
"grad_norm": 0.8745098195983263,
"learning_rate": 5.067327508138148e-06,
"loss": 0.7205,
"step": 1885
},
{
"epoch": 0.6996760758907913,
"grad_norm": 0.9669286665541216,
"learning_rate": 5.0112004678177e-06,
"loss": 0.7461,
"step": 1890
},
{
"epoch": 0.7015270708005553,
"grad_norm": 0.9439248557044554,
"learning_rate": 4.955281887023955e-06,
"loss": 0.7075,
"step": 1895
},
{
"epoch": 0.7033780657103192,
"grad_norm": 0.9612004193161535,
"learning_rate": 4.899574102343247e-06,
"loss": 0.7206,
"step": 1900
},
{
"epoch": 0.7033780657103192,
"eval_loss": 0.7448270320892334,
"eval_runtime": 125.9022,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 1900
},
{
"epoch": 0.7052290606200833,
"grad_norm": 0.9770375482265843,
"learning_rate": 4.844079441553717e-06,
"loss": 0.7312,
"step": 1905
},
{
"epoch": 0.7070800555298473,
"grad_norm": 0.9346202648428155,
"learning_rate": 4.7888002235279915e-06,
"loss": 0.7138,
"step": 1910
},
{
"epoch": 0.7089310504396112,
"grad_norm": 1.043008859124202,
"learning_rate": 4.733738758136327e-06,
"loss": 0.7201,
"step": 1915
},
{
"epoch": 0.7107820453493753,
"grad_norm": 0.9898196140644209,
"learning_rate": 4.678897346150067e-06,
"loss": 0.666,
"step": 1920
},
{
"epoch": 0.7126330402591393,
"grad_norm": 1.030447829309688,
"learning_rate": 4.624278279145509e-06,
"loss": 0.6968,
"step": 1925
},
{
"epoch": 0.7144840351689032,
"grad_norm": 1.0002038057708351,
"learning_rate": 4.56988383940817e-06,
"loss": 0.6815,
"step": 1930
},
{
"epoch": 0.7163350300786673,
"grad_norm": 0.8768349207452324,
"learning_rate": 4.515716299837376e-06,
"loss": 0.7143,
"step": 1935
},
{
"epoch": 0.7181860249884313,
"grad_norm": 0.9922854914505618,
"learning_rate": 4.461777923851337e-06,
"loss": 0.7168,
"step": 1940
},
{
"epoch": 0.7200370198981952,
"grad_norm": 0.9051677770762211,
"learning_rate": 4.408070965292534e-06,
"loss": 0.6871,
"step": 1945
},
{
"epoch": 0.7218880148079593,
"grad_norm": 0.9724233713765718,
"learning_rate": 4.354597668333551e-06,
"loss": 0.7079,
"step": 1950
},
{
"epoch": 0.7237390097177233,
"grad_norm": 1.0146841138791294,
"learning_rate": 4.30136026738332e-06,
"loss": 0.754,
"step": 1955
},
{
"epoch": 0.7255900046274872,
"grad_norm": 0.9260856251952512,
"learning_rate": 4.2483609869937115e-06,
"loss": 0.7366,
"step": 1960
},
{
"epoch": 0.7274409995372513,
"grad_norm": 0.9416530579594775,
"learning_rate": 4.195602041766638e-06,
"loss": 0.6743,
"step": 1965
},
{
"epoch": 0.7292919944470153,
"grad_norm": 0.9416273470678079,
"learning_rate": 4.143085636261452e-06,
"loss": 0.7548,
"step": 1970
},
{
"epoch": 0.7311429893567792,
"grad_norm": 0.9500826070023658,
"learning_rate": 4.090813964902889e-06,
"loss": 0.7318,
"step": 1975
},
{
"epoch": 0.7329939842665433,
"grad_norm": 1.044688650303936,
"learning_rate": 4.038789211889329e-06,
"loss": 0.691,
"step": 1980
},
{
"epoch": 0.7348449791763073,
"grad_norm": 0.9723998488803949,
"learning_rate": 3.987013551101543e-06,
"loss": 0.7205,
"step": 1985
},
{
"epoch": 0.7366959740860712,
"grad_norm": 1.0125026598661098,
"learning_rate": 3.9354891460118695e-06,
"loss": 0.7348,
"step": 1990
},
{
"epoch": 0.7385469689958353,
"grad_norm": 0.9786383589933646,
"learning_rate": 3.884218149593776e-06,
"loss": 0.6945,
"step": 1995
},
{
"epoch": 0.7403979639055993,
"grad_norm": 1.0768006764875049,
"learning_rate": 3.833202704231944e-06,
"loss": 0.7419,
"step": 2000
},
{
"epoch": 0.7403979639055993,
"eval_loss": 0.7422940135002136,
"eval_runtime": 125.9877,
"eval_samples_per_second": 1.016,
"eval_steps_per_second": 0.508,
"step": 2000
},
{
"epoch": 0.7422489588153632,
"grad_norm": 0.9714783620843155,
"learning_rate": 3.7824449416327123e-06,
"loss": 0.7486,
"step": 2005
},
{
"epoch": 0.7440999537251273,
"grad_norm": 0.9621418965839199,
"learning_rate": 3.7319469827350117e-06,
"loss": 0.7492,
"step": 2010
},
{
"epoch": 0.7459509486348913,
"grad_norm": 0.8838654671554304,
"learning_rate": 3.6817109376217574e-06,
"loss": 0.6973,
"step": 2015
},
{
"epoch": 0.7478019435446552,
"grad_norm": 1.0141612675604508,
"learning_rate": 3.631738905431641e-06,
"loss": 0.633,
"step": 2020
},
{
"epoch": 0.7496529384544193,
"grad_norm": 1.2478799462658503,
"learning_rate": 3.5820329742714666e-06,
"loss": 0.723,
"step": 2025
},
{
"epoch": 0.7515039333641832,
"grad_norm": 1.2113929317378798,
"learning_rate": 3.532595221128843e-06,
"loss": 0.7189,
"step": 2030
},
{
"epoch": 0.7533549282739472,
"grad_norm": 1.076623825378696,
"learning_rate": 3.483427711785449e-06,
"loss": 0.6935,
"step": 2035
},
{
"epoch": 0.7552059231837113,
"grad_norm": 1.0039638646776863,
"learning_rate": 3.4345325007306752e-06,
"loss": 0.7032,
"step": 2040
},
{
"epoch": 0.7570569180934752,
"grad_norm": 1.0481522906376972,
"learning_rate": 3.38591163107579e-06,
"loss": 0.7537,
"step": 2045
},
{
"epoch": 0.7589079130032392,
"grad_norm": 1.0167538112300836,
"learning_rate": 3.337567134468579e-06,
"loss": 0.7492,
"step": 2050
},
{
"epoch": 0.7607589079130033,
"grad_norm": 1.11799756175693,
"learning_rate": 3.2895010310084174e-06,
"loss": 0.687,
"step": 2055
},
{
"epoch": 0.7626099028227672,
"grad_norm": 1.026751398004723,
"learning_rate": 3.241715329161903e-06,
"loss": 0.7888,
"step": 2060
},
{
"epoch": 0.7644608977325312,
"grad_norm": 0.898992232633679,
"learning_rate": 3.1942120256788966e-06,
"loss": 0.7056,
"step": 2065
},
{
"epoch": 0.7663118926422953,
"grad_norm": 0.9452341682541248,
"learning_rate": 3.146993105509104e-06,
"loss": 0.6935,
"step": 2070
},
{
"epoch": 0.7681628875520592,
"grad_norm": 1.0381440177041223,
"learning_rate": 3.10006054171913e-06,
"loss": 0.6718,
"step": 2075
},
{
"epoch": 0.7700138824618232,
"grad_norm": 0.90151266081136,
"learning_rate": 3.0534162954100264e-06,
"loss": 0.734,
"step": 2080
},
{
"epoch": 0.7718648773715873,
"grad_norm": 0.9562956991505516,
"learning_rate": 3.0070623156353685e-06,
"loss": 0.7341,
"step": 2085
},
{
"epoch": 0.7737158722813512,
"grad_norm": 0.8801270739522954,
"learning_rate": 2.9610005393197707e-06,
"loss": 0.6526,
"step": 2090
},
{
"epoch": 0.7755668671911152,
"grad_norm": 0.9441163162257625,
"learning_rate": 2.9152328911780027e-06,
"loss": 0.6828,
"step": 2095
},
{
"epoch": 0.7774178621008793,
"grad_norm": 0.9634570912343101,
"learning_rate": 2.869761283634526e-06,
"loss": 0.7299,
"step": 2100
},
{
"epoch": 0.7774178621008793,
"eval_loss": 0.7405505180358887,
"eval_runtime": 125.9141,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 2100
},
{
"epoch": 0.7792688570106432,
"grad_norm": 0.9315181675274804,
"learning_rate": 2.8245876167435924e-06,
"loss": 0.7134,
"step": 2105
},
{
"epoch": 0.7811198519204072,
"grad_norm": 0.899902372156068,
"learning_rate": 2.779713778109867e-06,
"loss": 0.7112,
"step": 2110
},
{
"epoch": 0.7829708468301713,
"grad_norm": 0.9552279346547553,
"learning_rate": 2.7351416428095157e-06,
"loss": 0.7195,
"step": 2115
},
{
"epoch": 0.7848218417399352,
"grad_norm": 0.9400225445949076,
"learning_rate": 2.6908730733119025e-06,
"loss": 0.7465,
"step": 2120
},
{
"epoch": 0.7866728366496992,
"grad_norm": 0.9800712358563762,
"learning_rate": 2.6469099194017144e-06,
"loss": 0.739,
"step": 2125
},
{
"epoch": 0.7885238315594633,
"grad_norm": 0.9495123514350187,
"learning_rate": 2.603254018101715e-06,
"loss": 0.7729,
"step": 2130
},
{
"epoch": 0.7903748264692272,
"grad_norm": 0.981224001693797,
"learning_rate": 2.5599071935959495e-06,
"loss": 0.6977,
"step": 2135
},
{
"epoch": 0.7922258213789912,
"grad_norm": 0.979970394782048,
"learning_rate": 2.5168712571535305e-06,
"loss": 0.7,
"step": 2140
},
{
"epoch": 0.7940768162887553,
"grad_norm": 1.0931405844387208,
"learning_rate": 2.4741480070529657e-06,
"loss": 0.7212,
"step": 2145
},
{
"epoch": 0.7959278111985192,
"grad_norm": 0.9525278212316626,
"learning_rate": 2.4317392285069885e-06,
"loss": 0.7107,
"step": 2150
},
{
"epoch": 0.7977788061082832,
"grad_norm": 0.9587323014669643,
"learning_rate": 2.3896466935879957e-06,
"loss": 0.6903,
"step": 2155
},
{
"epoch": 0.7996298010180471,
"grad_norm": 0.8878613939635316,
"learning_rate": 2.3478721611539712e-06,
"loss": 0.755,
"step": 2160
},
{
"epoch": 0.8014807959278112,
"grad_norm": 0.9724146216671814,
"learning_rate": 2.3064173767750055e-06,
"loss": 0.695,
"step": 2165
},
{
"epoch": 0.8033317908375752,
"grad_norm": 1.0468722007497135,
"learning_rate": 2.265284072660362e-06,
"loss": 0.7233,
"step": 2170
},
{
"epoch": 0.8051827857473391,
"grad_norm": 1.0316210745516678,
"learning_rate": 2.224473967586073e-06,
"loss": 0.7607,
"step": 2175
},
{
"epoch": 0.8070337806571032,
"grad_norm": 1.0548409137457446,
"learning_rate": 2.1839887668231486e-06,
"loss": 0.697,
"step": 2180
},
{
"epoch": 0.8088847755668672,
"grad_norm": 1.0039170894073437,
"learning_rate": 2.1438301620662994e-06,
"loss": 0.7324,
"step": 2185
},
{
"epoch": 0.8107357704766311,
"grad_norm": 0.9588651745318834,
"learning_rate": 2.103999831363258e-06,
"loss": 0.6748,
"step": 2190
},
{
"epoch": 0.8125867653863952,
"grad_norm": 0.8996370636189378,
"learning_rate": 2.064499439044657e-06,
"loss": 0.73,
"step": 2195
},
{
"epoch": 0.8144377602961592,
"grad_norm": 0.9239380519912793,
"learning_rate": 2.0253306356544843e-06,
"loss": 0.7057,
"step": 2200
},
{
"epoch": 0.8144377602961592,
"eval_loss": 0.738842248916626,
"eval_runtime": 125.9296,
"eval_samples_per_second": 1.016,
"eval_steps_per_second": 0.508,
"step": 2200
},
{
"epoch": 0.8162887552059231,
"grad_norm": 0.9429244883085828,
"learning_rate": 1.986495057881126e-06,
"loss": 0.6772,
"step": 2205
},
{
"epoch": 0.8181397501156872,
"grad_norm": 1.0195318731984253,
"learning_rate": 1.947994328488949e-06,
"loss": 0.7356,
"step": 2210
},
{
"epoch": 0.8199907450254512,
"grad_norm": 0.9756633823952066,
"learning_rate": 1.9098300562505266e-06,
"loss": 0.7717,
"step": 2215
},
{
"epoch": 0.8218417399352151,
"grad_norm": 0.9865779961666936,
"learning_rate": 1.872003835879389e-06,
"loss": 0.6844,
"step": 2220
},
{
"epoch": 0.8236927348449792,
"grad_norm": 1.1283985725895185,
"learning_rate": 1.8345172479633977e-06,
"loss": 0.6513,
"step": 2225
},
{
"epoch": 0.8255437297547432,
"grad_norm": 1.008863839428332,
"learning_rate": 1.79737185889871e-06,
"loss": 0.6812,
"step": 2230
},
{
"epoch": 0.8273947246645071,
"grad_norm": 0.9300553415281679,
"learning_rate": 1.7605692208242953e-06,
"loss": 0.6909,
"step": 2235
},
{
"epoch": 0.8292457195742712,
"grad_norm": 0.9710044483167964,
"learning_rate": 1.7241108715571197e-06,
"loss": 0.718,
"step": 2240
},
{
"epoch": 0.8310967144840352,
"grad_norm": 1.015816894080233,
"learning_rate": 1.6879983345278528e-06,
"loss": 0.6821,
"step": 2245
},
{
"epoch": 0.8329477093937991,
"grad_norm": 0.9542344503119925,
"learning_rate": 1.652233118717229e-06,
"loss": 0.719,
"step": 2250
},
{
"epoch": 0.8347987043035632,
"grad_norm": 0.9708933601771149,
"learning_rate": 1.6168167185929883e-06,
"loss": 0.7041,
"step": 2255
},
{
"epoch": 0.8366496992133272,
"grad_norm": 1.005762211417554,
"learning_rate": 1.5817506140474248e-06,
"loss": 0.7491,
"step": 2260
},
{
"epoch": 0.8385006941230911,
"grad_norm": 1.1324500609372297,
"learning_rate": 1.5470362703355678e-06,
"loss": 0.7599,
"step": 2265
},
{
"epoch": 0.8403516890328552,
"grad_norm": 1.0774483423911148,
"learning_rate": 1.5126751380139203e-06,
"loss": 0.7304,
"step": 2270
},
{
"epoch": 0.8422026839426192,
"grad_norm": 0.9746075541108256,
"learning_rate": 1.4786686528798878e-06,
"loss": 0.7859,
"step": 2275
},
{
"epoch": 0.8440536788523831,
"grad_norm": 0.9132206341810313,
"learning_rate": 1.4450182359117493e-06,
"loss": 0.6567,
"step": 2280
},
{
"epoch": 0.8459046737621472,
"grad_norm": 0.9843969970761268,
"learning_rate": 1.4117252932093007e-06,
"loss": 0.7079,
"step": 2285
},
{
"epoch": 0.8477556686719111,
"grad_norm": 0.909900305880051,
"learning_rate": 1.3787912159350903e-06,
"loss": 0.7092,
"step": 2290
},
{
"epoch": 0.8496066635816751,
"grad_norm": 0.9862835059860517,
"learning_rate": 1.3462173802562917e-06,
"loss": 0.688,
"step": 2295
},
{
"epoch": 0.8514576584914392,
"grad_norm": 0.9817682414409103,
"learning_rate": 1.3140051472872062e-06,
"loss": 0.7146,
"step": 2300
},
{
"epoch": 0.8514576584914392,
"eval_loss": 0.7375185489654541,
"eval_runtime": 125.8616,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 2300
},
{
"epoch": 0.8533086534012031,
"grad_norm": 0.97635505354809,
"learning_rate": 1.282155863032377e-06,
"loss": 0.6785,
"step": 2305
},
{
"epoch": 0.8551596483109671,
"grad_norm": 0.9432652702625532,
"learning_rate": 1.2506708583303495e-06,
"loss": 0.6829,
"step": 2310
},
{
"epoch": 0.8570106432207312,
"grad_norm": 1.0500105672657278,
"learning_rate": 1.2195514487980643e-06,
"loss": 0.7097,
"step": 2315
},
{
"epoch": 0.8588616381304951,
"grad_norm": 0.9677123461437018,
"learning_rate": 1.188798934775881e-06,
"loss": 0.7271,
"step": 2320
},
{
"epoch": 0.8607126330402591,
"grad_norm": 0.9822692297793124,
"learning_rate": 1.158414601273251e-06,
"loss": 0.74,
"step": 2325
},
{
"epoch": 0.8625636279500232,
"grad_norm": 1.02828865110176,
"learning_rate": 1.1283997179149987e-06,
"loss": 0.7406,
"step": 2330
},
{
"epoch": 0.8644146228597871,
"grad_norm": 1.0596259638715508,
"learning_rate": 1.0987555388883042e-06,
"loss": 0.6883,
"step": 2335
},
{
"epoch": 0.8662656177695511,
"grad_norm": 0.9898076031142375,
"learning_rate": 1.0694833028902686e-06,
"loss": 0.7012,
"step": 2340
},
{
"epoch": 0.8681166126793152,
"grad_norm": 0.9639230942285071,
"learning_rate": 1.0405842330761651e-06,
"loss": 0.7491,
"step": 2345
},
{
"epoch": 0.8699676075890791,
"grad_norm": 0.9865850178730345,
"learning_rate": 1.012059537008332e-06,
"loss": 0.7295,
"step": 2350
},
{
"epoch": 0.8718186024988431,
"grad_norm": 0.9302298993217906,
"learning_rate": 9.839104066057025e-07,
"loss": 0.7505,
"step": 2355
},
{
"epoch": 0.8736695974086072,
"grad_norm": 0.9246047712950533,
"learning_rate": 9.561380180940182e-07,
"loss": 0.7165,
"step": 2360
},
{
"epoch": 0.8755205923183711,
"grad_norm": 0.9341328111976939,
"learning_rate": 9.287435319566618e-07,
"loss": 0.6928,
"step": 2365
},
{
"epoch": 0.8773715872281351,
"grad_norm": 1.0029919467384116,
"learning_rate": 9.017280928861727e-07,
"loss": 0.7503,
"step": 2370
},
{
"epoch": 0.8792225821378992,
"grad_norm": 0.924763857124508,
"learning_rate": 8.750928297364192e-07,
"loss": 0.6788,
"step": 2375
},
{
"epoch": 0.8810735770476631,
"grad_norm": 0.9276137090660882,
"learning_rate": 8.488388554754223e-07,
"loss": 0.6975,
"step": 2380
},
{
"epoch": 0.8829245719574271,
"grad_norm": 0.9511441358772591,
"learning_rate": 8.229672671388578e-07,
"loss": 0.7379,
"step": 2385
},
{
"epoch": 0.8847755668671912,
"grad_norm": 0.9703292177752127,
"learning_rate": 7.974791457842012e-07,
"loss": 0.6863,
"step": 2390
},
{
"epoch": 0.8866265617769551,
"grad_norm": 0.9665966427045252,
"learning_rate": 7.723755564455771e-07,
"loss": 0.7569,
"step": 2395
},
{
"epoch": 0.8884775566867191,
"grad_norm": 0.9199955573458448,
"learning_rate": 7.476575480892357e-07,
"loss": 0.7246,
"step": 2400
},
{
"epoch": 0.8884775566867191,
"eval_loss": 0.7366934418678284,
"eval_runtime": 125.8212,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.509,
"step": 2400
},
{
"epoch": 0.8903285515964832,
"grad_norm": 0.9866170842794822,
"learning_rate": 7.233261535697333e-07,
"loss": 0.6831,
"step": 2405
},
{
"epoch": 0.8921795465062471,
"grad_norm": 0.9020922896178015,
"learning_rate": 6.99382389586769e-07,
"loss": 0.7543,
"step": 2410
},
{
"epoch": 0.8940305414160111,
"grad_norm": 0.8891310119602659,
"learning_rate": 6.758272566427027e-07,
"loss": 0.6992,
"step": 2415
},
{
"epoch": 0.895881536325775,
"grad_norm": 1.0258150846316885,
"learning_rate": 6.526617390007506e-07,
"loss": 0.7159,
"step": 2420
},
{
"epoch": 0.8977325312355391,
"grad_norm": 0.9315994997755779,
"learning_rate": 6.298868046438533e-07,
"loss": 0.7339,
"step": 2425
},
{
"epoch": 0.8995835261453031,
"grad_norm": 0.8794790525509596,
"learning_rate": 6.075034052342288e-07,
"loss": 0.6877,
"step": 2430
},
{
"epoch": 0.901434521055067,
"grad_norm": 1.0216931119406394,
"learning_rate": 5.855124760736119e-07,
"loss": 0.7273,
"step": 2435
},
{
"epoch": 0.9032855159648311,
"grad_norm": 0.9162022348716465,
"learning_rate": 5.63914936064165e-07,
"loss": 0.6129,
"step": 2440
},
{
"epoch": 0.9051365108745951,
"grad_norm": 1.0548470483887789,
"learning_rate": 5.427116876700877e-07,
"loss": 0.7462,
"step": 2445
},
{
"epoch": 0.906987505784359,
"grad_norm": 0.9119109915842329,
"learning_rate": 5.219036168798986e-07,
"loss": 0.7091,
"step": 2450
},
{
"epoch": 0.9088385006941231,
"grad_norm": 0.9773112801791846,
"learning_rate": 5.014915931694253e-07,
"loss": 0.7402,
"step": 2455
},
{
"epoch": 0.9106894956038871,
"grad_norm": 0.967120980381721,
"learning_rate": 4.81476469465465e-07,
"loss": 0.7563,
"step": 2460
},
{
"epoch": 0.912540490513651,
"grad_norm": 1.0900216384603985,
"learning_rate": 4.618590821101432e-07,
"loss": 0.7257,
"step": 2465
},
{
"epoch": 0.9143914854234151,
"grad_norm": 0.9410281325856403,
"learning_rate": 4.4264025082597084e-07,
"loss": 0.726,
"step": 2470
},
{
"epoch": 0.9162424803331791,
"grad_norm": 0.9555645068242209,
"learning_rate": 4.2382077868159286e-07,
"loss": 0.6977,
"step": 2475
},
{
"epoch": 0.918093475242943,
"grad_norm": 1.1057620347138868,
"learning_rate": 4.054014520582283e-07,
"loss": 0.7611,
"step": 2480
},
{
"epoch": 0.9199444701527071,
"grad_norm": 0.9023240417887544,
"learning_rate": 3.8738304061681107e-07,
"loss": 0.7097,
"step": 2485
},
{
"epoch": 0.9217954650624711,
"grad_norm": 0.8961027836341736,
"learning_rate": 3.6976629726583115e-07,
"loss": 0.7533,
"step": 2490
},
{
"epoch": 0.923646459972235,
"grad_norm": 0.9463370594210976,
"learning_rate": 3.525519581298731e-07,
"loss": 0.7657,
"step": 2495
},
{
"epoch": 0.9254974548819991,
"grad_norm": 0.8857551112046717,
"learning_rate": 3.357407425188541e-07,
"loss": 0.6637,
"step": 2500
},
{
"epoch": 0.9254974548819991,
"eval_loss": 0.736148476600647,
"eval_runtime": 125.8161,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.509,
"step": 2500
},
{
"epoch": 0.9273484497917631,
"grad_norm": 0.9557036435732803,
"learning_rate": 3.1933335289797364e-07,
"loss": 0.7211,
"step": 2505
},
{
"epoch": 0.929199444701527,
"grad_norm": 0.9411968516873396,
"learning_rate": 3.033304748583543e-07,
"loss": 0.6711,
"step": 2510
},
{
"epoch": 0.9310504396112911,
"grad_norm": 0.9529612292006595,
"learning_rate": 2.877327770883964e-07,
"loss": 0.7352,
"step": 2515
},
{
"epoch": 0.9329014345210551,
"grad_norm": 0.9072504944314087,
"learning_rate": 2.7254091134583995e-07,
"loss": 0.702,
"step": 2520
},
{
"epoch": 0.934752429430819,
"grad_norm": 1.0216263771503715,
"learning_rate": 2.577555124305209e-07,
"loss": 0.6984,
"step": 2525
},
{
"epoch": 0.9366034243405831,
"grad_norm": 0.9271796888513735,
"learning_rate": 2.433771981578581e-07,
"loss": 0.7594,
"step": 2530
},
{
"epoch": 0.9384544192503471,
"grad_norm": 0.9325885923721386,
"learning_rate": 2.2940656933302607e-07,
"loss": 0.721,
"step": 2535
},
{
"epoch": 0.940305414160111,
"grad_norm": 1.0530858890236963,
"learning_rate": 2.1584420972586174e-07,
"loss": 0.7108,
"step": 2540
},
{
"epoch": 0.9421564090698751,
"grad_norm": 1.9244924879459369,
"learning_rate": 2.0269068604646058e-07,
"loss": 0.7149,
"step": 2545
},
{
"epoch": 0.944007403979639,
"grad_norm": 1.0361671451409022,
"learning_rate": 1.8994654792150125e-07,
"loss": 0.7285,
"step": 2550
},
{
"epoch": 0.945858398889403,
"grad_norm": 0.9757871793671443,
"learning_rate": 1.7761232787127936e-07,
"loss": 0.6992,
"step": 2555
},
{
"epoch": 0.9477093937991671,
"grad_norm": 0.9675283656651317,
"learning_rate": 1.6568854128745537e-07,
"loss": 0.7044,
"step": 2560
},
{
"epoch": 0.949560388708931,
"grad_norm": 0.9846806445819078,
"learning_rate": 1.5417568641151848e-07,
"loss": 0.6704,
"step": 2565
},
{
"epoch": 0.951411383618695,
"grad_norm": 0.9281209052786342,
"learning_rate": 1.4307424431396654e-07,
"loss": 0.7176,
"step": 2570
},
{
"epoch": 0.9532623785284591,
"grad_norm": 0.8999935681059046,
"learning_rate": 1.323846788742078e-07,
"loss": 0.7003,
"step": 2575
},
{
"epoch": 0.955113373438223,
"grad_norm": 0.9213878579995839,
"learning_rate": 1.2210743676117188e-07,
"loss": 0.6906,
"step": 2580
},
{
"epoch": 0.956964368347987,
"grad_norm": 0.9250556197998173,
"learning_rate": 1.1224294741464914e-07,
"loss": 0.6794,
"step": 2585
},
{
"epoch": 0.9588153632577511,
"grad_norm": 0.9221148649933307,
"learning_rate": 1.0279162302734624e-07,
"loss": 0.6645,
"step": 2590
},
{
"epoch": 0.960666358167515,
"grad_norm": 0.9308662826828088,
"learning_rate": 9.375385852765983e-08,
"loss": 0.7139,
"step": 2595
},
{
"epoch": 0.962517353077279,
"grad_norm": 0.9640193787486492,
"learning_rate": 8.513003156317978e-08,
"loss": 0.7416,
"step": 2600
},
{
"epoch": 0.962517353077279,
"eval_loss": 0.7359679937362671,
"eval_runtime": 125.8349,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.509,
"step": 2600
},
{
"epoch": 0.9643683479870431,
"grad_norm": 0.9534219021183425,
"learning_rate": 7.692050248490291e-08,
"loss": 0.7339,
"step": 2605
},
{
"epoch": 0.966219342896807,
"grad_norm": 0.9352372053413327,
"learning_rate": 6.912561433217946e-08,
"loss": 0.7221,
"step": 2610
},
{
"epoch": 0.968070337806571,
"grad_norm": 0.9840909103193446,
"learning_rate": 6.174569281837573e-08,
"loss": 0.7012,
"step": 2615
},
{
"epoch": 0.9699213327163351,
"grad_norm": 0.9413628052352356,
"learning_rate": 5.4781046317267103e-08,
"loss": 0.7021,
"step": 2620
},
{
"epoch": 0.971772327626099,
"grad_norm": 0.9862336464985173,
"learning_rate": 4.823196585015066e-08,
"loss": 0.6618,
"step": 2625
},
{
"epoch": 0.973623322535863,
"grad_norm": 0.9484184901249422,
"learning_rate": 4.209872507368706e-08,
"loss": 0.6719,
"step": 2630
},
{
"epoch": 0.9754743174456271,
"grad_norm": 1.0274897071656908,
"learning_rate": 3.6381580268463056e-08,
"loss": 0.7218,
"step": 2635
},
{
"epoch": 0.977325312355391,
"grad_norm": 0.9664338709751774,
"learning_rate": 3.108077032828116e-08,
"loss": 0.6997,
"step": 2640
},
{
"epoch": 0.979176307265155,
"grad_norm": 1.1238393315125719,
"learning_rate": 2.6196516750183198e-08,
"loss": 0.7338,
"step": 2645
},
{
"epoch": 0.9810273021749191,
"grad_norm": 0.8933118725313647,
"learning_rate": 2.1729023625189916e-08,
"loss": 0.7035,
"step": 2650
},
{
"epoch": 0.982878297084683,
"grad_norm": 0.9704540461821111,
"learning_rate": 1.767847762977337e-08,
"loss": 0.7668,
"step": 2655
},
{
"epoch": 0.984729291994447,
"grad_norm": 0.9466080459329079,
"learning_rate": 1.4045048018059837e-08,
"loss": 0.6771,
"step": 2660
},
{
"epoch": 0.9865802869042111,
"grad_norm": 0.9644726251473568,
"learning_rate": 1.0828886614754342e-08,
"loss": 0.7539,
"step": 2665
},
{
"epoch": 0.988431281813975,
"grad_norm": 0.8818176818424777,
"learning_rate": 8.030127808797972e-09,
"loss": 0.719,
"step": 2670
},
{
"epoch": 0.990282276723739,
"grad_norm": 0.8961966562928385,
"learning_rate": 5.648888547750142e-09,
"loss": 0.697,
"step": 2675
},
{
"epoch": 0.992133271633503,
"grad_norm": 1.2004341644367225,
"learning_rate": 3.6852683329058336e-09,
"loss": 0.7457,
"step": 2680
},
{
"epoch": 0.993984266543267,
"grad_norm": 0.9729347850621183,
"learning_rate": 2.1393492151333684e-09,
"loss": 0.6743,
"step": 2685
},
{
"epoch": 0.995835261453031,
"grad_norm": 0.9956973839014348,
"learning_rate": 1.0111957914515914e-09,
"loss": 0.6976,
"step": 2690
},
{
"epoch": 0.997686256362795,
"grad_norm": 0.9334887689919787,
"learning_rate": 3.008552023242572e-10,
"loss": 0.7241,
"step": 2695
},
{
"epoch": 0.999537251272559,
"grad_norm": 0.9781490083976876,
"learning_rate": 8.357129693825628e-12,
"loss": 0.7024,
"step": 2700
},
{
"epoch": 0.999537251272559,
"eval_loss": 0.7359436750411987,
"eval_runtime": 125.9053,
"eval_samples_per_second": 1.017,
"eval_steps_per_second": 0.508,
"step": 2700
},
{
"epoch": 0.9999074502545118,
"step": 2701,
"total_flos": 76938190258176.0,
"train_loss": 0.7557382643200741,
"train_runtime": 83163.2816,
"train_samples_per_second": 0.26,
"train_steps_per_second": 0.032
}
],
"logging_steps": 5,
"max_steps": 2701,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 76938190258176.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}