DANGDOCAO's picture
HVU_QA
bec6a02 verified
raw
history blame
193 kB
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.42472682342947604,
"eval_steps": 500,
"global_step": 11000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0003861152940267964,
"grad_norm": 8.003422737121582,
"learning_rate": 0.0001999768330823584,
"loss": 5.0107,
"step": 10
},
{
"epoch": 0.0007722305880535929,
"grad_norm": 11.965606689453125,
"learning_rate": 0.0001999510920627566,
"loss": 0.9013,
"step": 20
},
{
"epoch": 0.0011583458820803893,
"grad_norm": 9.783374786376953,
"learning_rate": 0.00019992535104315483,
"loss": 0.5506,
"step": 30
},
{
"epoch": 0.0015444611761071857,
"grad_norm": 5.720436096191406,
"learning_rate": 0.00019989961002355304,
"loss": 0.501,
"step": 40
},
{
"epoch": 0.001930576470133982,
"grad_norm": 11.18126392364502,
"learning_rate": 0.00019987386900395125,
"loss": 0.8176,
"step": 50
},
{
"epoch": 0.0023166917641607786,
"grad_norm": 8.88875675201416,
"learning_rate": 0.00019984812798434947,
"loss": 0.5404,
"step": 60
},
{
"epoch": 0.0027028070581875748,
"grad_norm": 2.9886066913604736,
"learning_rate": 0.00019982238696474768,
"loss": 0.5295,
"step": 70
},
{
"epoch": 0.0030889223522143714,
"grad_norm": 8.936307907104492,
"learning_rate": 0.0001997966459451459,
"loss": 0.7398,
"step": 80
},
{
"epoch": 0.0034750376462411676,
"grad_norm": 11.393534660339355,
"learning_rate": 0.0001997709049255441,
"loss": 0.6333,
"step": 90
},
{
"epoch": 0.003861152940267964,
"grad_norm": 8.725994110107422,
"learning_rate": 0.00019974516390594235,
"loss": 0.4689,
"step": 100
},
{
"epoch": 0.00424726823429476,
"grad_norm": 20.316652297973633,
"learning_rate": 0.00019971942288634053,
"loss": 0.8522,
"step": 110
},
{
"epoch": 0.004633383528321557,
"grad_norm": 4.232663631439209,
"learning_rate": 0.00019969368186673875,
"loss": 0.5041,
"step": 120
},
{
"epoch": 0.005019498822348353,
"grad_norm": 2.609255313873291,
"learning_rate": 0.00019966794084713696,
"loss": 0.7439,
"step": 130
},
{
"epoch": 0.0054056141163751495,
"grad_norm": 10.063919067382812,
"learning_rate": 0.00019964219982753517,
"loss": 0.6702,
"step": 140
},
{
"epoch": 0.005791729410401946,
"grad_norm": 5.799802303314209,
"learning_rate": 0.00019961645880793339,
"loss": 0.7356,
"step": 150
},
{
"epoch": 0.006177844704428743,
"grad_norm": 12.664258003234863,
"learning_rate": 0.0001995907177883316,
"loss": 0.4067,
"step": 160
},
{
"epoch": 0.006563959998455539,
"grad_norm": 9.42366886138916,
"learning_rate": 0.00019956497676872984,
"loss": 0.4767,
"step": 170
},
{
"epoch": 0.006950075292482335,
"grad_norm": 5.382272243499756,
"learning_rate": 0.00019953923574912803,
"loss": 0.536,
"step": 180
},
{
"epoch": 0.0073361905865091314,
"grad_norm": 9.797371864318848,
"learning_rate": 0.00019951349472952624,
"loss": 0.4735,
"step": 190
},
{
"epoch": 0.007722305880535928,
"grad_norm": 7.965329647064209,
"learning_rate": 0.00019948775370992445,
"loss": 0.3881,
"step": 200
},
{
"epoch": 0.008108421174562725,
"grad_norm": 4.075791835784912,
"learning_rate": 0.00019946201269032267,
"loss": 0.5564,
"step": 210
},
{
"epoch": 0.00849453646858952,
"grad_norm": 24.367305755615234,
"learning_rate": 0.0001994362716707209,
"loss": 0.9795,
"step": 220
},
{
"epoch": 0.008880651762616317,
"grad_norm": 9.627866744995117,
"learning_rate": 0.0001994105306511191,
"loss": 0.4528,
"step": 230
},
{
"epoch": 0.009266767056643114,
"grad_norm": 7.469555854797363,
"learning_rate": 0.00019938478963151733,
"loss": 0.447,
"step": 240
},
{
"epoch": 0.00965288235066991,
"grad_norm": 7.426730155944824,
"learning_rate": 0.00019935904861191552,
"loss": 0.6026,
"step": 250
},
{
"epoch": 0.010038997644696707,
"grad_norm": 6.999317169189453,
"learning_rate": 0.00019933330759231373,
"loss": 0.4962,
"step": 260
},
{
"epoch": 0.010425112938723502,
"grad_norm": 10.492286682128906,
"learning_rate": 0.00019930756657271194,
"loss": 0.7987,
"step": 270
},
{
"epoch": 0.010811228232750299,
"grad_norm": 7.079407215118408,
"learning_rate": 0.00019928182555311016,
"loss": 0.4395,
"step": 280
},
{
"epoch": 0.011197343526777096,
"grad_norm": 9.610014915466309,
"learning_rate": 0.0001992560845335084,
"loss": 0.748,
"step": 290
},
{
"epoch": 0.011583458820803891,
"grad_norm": 5.993048667907715,
"learning_rate": 0.00019923034351390658,
"loss": 0.4328,
"step": 300
},
{
"epoch": 0.011969574114830689,
"grad_norm": 7.336791515350342,
"learning_rate": 0.00019920460249430483,
"loss": 0.4104,
"step": 310
},
{
"epoch": 0.012355689408857486,
"grad_norm": 7.967221736907959,
"learning_rate": 0.000199178861474703,
"loss": 0.4662,
"step": 320
},
{
"epoch": 0.012741804702884281,
"grad_norm": 4.464987754821777,
"learning_rate": 0.00019915312045510125,
"loss": 0.725,
"step": 330
},
{
"epoch": 0.013127919996911078,
"grad_norm": 8.669449806213379,
"learning_rate": 0.00019912737943549944,
"loss": 0.4256,
"step": 340
},
{
"epoch": 0.013514035290937873,
"grad_norm": 4.114014148712158,
"learning_rate": 0.00019910163841589765,
"loss": 0.4477,
"step": 350
},
{
"epoch": 0.01390015058496467,
"grad_norm": 9.254106521606445,
"learning_rate": 0.0001990758973962959,
"loss": 0.514,
"step": 360
},
{
"epoch": 0.014286265878991468,
"grad_norm": 0.8039970993995667,
"learning_rate": 0.00019905015637669408,
"loss": 0.5802,
"step": 370
},
{
"epoch": 0.014672381173018263,
"grad_norm": 3.9931838512420654,
"learning_rate": 0.00019902441535709232,
"loss": 0.8973,
"step": 380
},
{
"epoch": 0.01505849646704506,
"grad_norm": 1.7645355463027954,
"learning_rate": 0.0001989986743374905,
"loss": 0.7108,
"step": 390
},
{
"epoch": 0.015444611761071855,
"grad_norm": 6.8542866706848145,
"learning_rate": 0.00019897293331788875,
"loss": 0.5796,
"step": 400
},
{
"epoch": 0.015830727055098654,
"grad_norm": 5.278103828430176,
"learning_rate": 0.00019894719229828696,
"loss": 0.3841,
"step": 410
},
{
"epoch": 0.01621684234912545,
"grad_norm": 9.00206184387207,
"learning_rate": 0.00019892145127868514,
"loss": 0.5891,
"step": 420
},
{
"epoch": 0.016602957643152245,
"grad_norm": 7.684702396392822,
"learning_rate": 0.00019889571025908339,
"loss": 0.4868,
"step": 430
},
{
"epoch": 0.01698907293717904,
"grad_norm": 4.198502540588379,
"learning_rate": 0.00019886996923948157,
"loss": 0.571,
"step": 440
},
{
"epoch": 0.01737518823120584,
"grad_norm": 7.454501628875732,
"learning_rate": 0.0001988442282198798,
"loss": 0.5133,
"step": 450
},
{
"epoch": 0.017761303525232634,
"grad_norm": 13.236722946166992,
"learning_rate": 0.000198818487200278,
"loss": 0.4139,
"step": 460
},
{
"epoch": 0.01814741881925943,
"grad_norm": 6.4592390060424805,
"learning_rate": 0.00019879274618067624,
"loss": 0.6078,
"step": 470
},
{
"epoch": 0.01853353411328623,
"grad_norm": 11.73417854309082,
"learning_rate": 0.00019876700516107445,
"loss": 0.5472,
"step": 480
},
{
"epoch": 0.018919649407313024,
"grad_norm": 2.5162808895111084,
"learning_rate": 0.00019874126414147264,
"loss": 0.6611,
"step": 490
},
{
"epoch": 0.01930576470133982,
"grad_norm": 4.9637837409973145,
"learning_rate": 0.00019871552312187088,
"loss": 0.6472,
"step": 500
},
{
"epoch": 0.019691879995366618,
"grad_norm": 11.545489311218262,
"learning_rate": 0.00019868978210226906,
"loss": 0.5304,
"step": 510
},
{
"epoch": 0.020077995289393413,
"grad_norm": 5.197858810424805,
"learning_rate": 0.0001986640410826673,
"loss": 0.605,
"step": 520
},
{
"epoch": 0.02046411058342021,
"grad_norm": 4.935055255889893,
"learning_rate": 0.0001986383000630655,
"loss": 0.6524,
"step": 530
},
{
"epoch": 0.020850225877447004,
"grad_norm": 5.838052749633789,
"learning_rate": 0.00019861255904346373,
"loss": 0.4957,
"step": 540
},
{
"epoch": 0.021236341171473803,
"grad_norm": 4.682408809661865,
"learning_rate": 0.00019858681802386194,
"loss": 0.8523,
"step": 550
},
{
"epoch": 0.021622456465500598,
"grad_norm": 10.720857620239258,
"learning_rate": 0.00019856107700426013,
"loss": 0.516,
"step": 560
},
{
"epoch": 0.022008571759527393,
"grad_norm": 6.515562534332275,
"learning_rate": 0.00019853533598465837,
"loss": 0.6095,
"step": 570
},
{
"epoch": 0.022394687053554192,
"grad_norm": 3.204960584640503,
"learning_rate": 0.00019850959496505656,
"loss": 0.6624,
"step": 580
},
{
"epoch": 0.022780802347580988,
"grad_norm": 2.305497884750366,
"learning_rate": 0.0001984838539454548,
"loss": 0.5986,
"step": 590
},
{
"epoch": 0.023166917641607783,
"grad_norm": 13.07105541229248,
"learning_rate": 0.000198458112925853,
"loss": 0.337,
"step": 600
},
{
"epoch": 0.023553032935634582,
"grad_norm": 3.1491329669952393,
"learning_rate": 0.00019843237190625122,
"loss": 0.5466,
"step": 610
},
{
"epoch": 0.023939148229661377,
"grad_norm": 1.92014479637146,
"learning_rate": 0.00019840663088664944,
"loss": 0.5525,
"step": 620
},
{
"epoch": 0.024325263523688172,
"grad_norm": 2.206550121307373,
"learning_rate": 0.00019838088986704762,
"loss": 0.5069,
"step": 630
},
{
"epoch": 0.02471137881771497,
"grad_norm": 2.386288642883301,
"learning_rate": 0.00019835514884744586,
"loss": 0.3749,
"step": 640
},
{
"epoch": 0.025097494111741767,
"grad_norm": 7.393959045410156,
"learning_rate": 0.00019832940782784405,
"loss": 0.7144,
"step": 650
},
{
"epoch": 0.025483609405768562,
"grad_norm": 5.7293171882629395,
"learning_rate": 0.0001983036668082423,
"loss": 0.5052,
"step": 660
},
{
"epoch": 0.025869724699795357,
"grad_norm": 6.440220832824707,
"learning_rate": 0.0001982779257886405,
"loss": 0.4343,
"step": 670
},
{
"epoch": 0.026255839993822156,
"grad_norm": 0.8553487658500671,
"learning_rate": 0.00019825218476903872,
"loss": 0.7562,
"step": 680
},
{
"epoch": 0.02664195528784895,
"grad_norm": 3.762784719467163,
"learning_rate": 0.00019822644374943693,
"loss": 0.5593,
"step": 690
},
{
"epoch": 0.027028070581875747,
"grad_norm": 11.674392700195312,
"learning_rate": 0.00019820070272983512,
"loss": 0.6069,
"step": 700
},
{
"epoch": 0.027414185875902546,
"grad_norm": 8.631232261657715,
"learning_rate": 0.00019817496171023336,
"loss": 0.3584,
"step": 710
},
{
"epoch": 0.02780030116992934,
"grad_norm": 5.7163920402526855,
"learning_rate": 0.00019814922069063157,
"loss": 0.5563,
"step": 720
},
{
"epoch": 0.028186416463956136,
"grad_norm": 8.186172485351562,
"learning_rate": 0.00019812347967102978,
"loss": 0.5289,
"step": 730
},
{
"epoch": 0.028572531757982935,
"grad_norm": 7.287814140319824,
"learning_rate": 0.000198097738651428,
"loss": 0.4543,
"step": 740
},
{
"epoch": 0.02895864705200973,
"grad_norm": 6.621245384216309,
"learning_rate": 0.0001980719976318262,
"loss": 0.3244,
"step": 750
},
{
"epoch": 0.029344762346036526,
"grad_norm": 3.5209403038024902,
"learning_rate": 0.00019804625661222442,
"loss": 0.5385,
"step": 760
},
{
"epoch": 0.029730877640063325,
"grad_norm": 2.562343120574951,
"learning_rate": 0.00019802051559262264,
"loss": 0.4868,
"step": 770
},
{
"epoch": 0.03011699293409012,
"grad_norm": 7.782780647277832,
"learning_rate": 0.00019799477457302085,
"loss": 0.5682,
"step": 780
},
{
"epoch": 0.030503108228116915,
"grad_norm": 8.173531532287598,
"learning_rate": 0.00019796903355341906,
"loss": 0.3557,
"step": 790
},
{
"epoch": 0.03088922352214371,
"grad_norm": 4.502675533294678,
"learning_rate": 0.00019794329253381728,
"loss": 0.517,
"step": 800
},
{
"epoch": 0.031275338816170506,
"grad_norm": 6.314894199371338,
"learning_rate": 0.0001979175515142155,
"loss": 0.6392,
"step": 810
},
{
"epoch": 0.03166145411019731,
"grad_norm": 7.054763317108154,
"learning_rate": 0.0001978918104946137,
"loss": 0.5823,
"step": 820
},
{
"epoch": 0.032047569404224104,
"grad_norm": 1.7847551107406616,
"learning_rate": 0.00019786606947501192,
"loss": 0.4495,
"step": 830
},
{
"epoch": 0.0324336846982509,
"grad_norm": 5.268612861633301,
"learning_rate": 0.00019784032845541013,
"loss": 0.4379,
"step": 840
},
{
"epoch": 0.032819799992277694,
"grad_norm": 4.834717273712158,
"learning_rate": 0.00019781458743580834,
"loss": 0.5379,
"step": 850
},
{
"epoch": 0.03320591528630449,
"grad_norm": 6.077203273773193,
"learning_rate": 0.00019778884641620656,
"loss": 0.5666,
"step": 860
},
{
"epoch": 0.033592030580331285,
"grad_norm": 0.9583851099014282,
"learning_rate": 0.00019776310539660477,
"loss": 0.8146,
"step": 870
},
{
"epoch": 0.03397814587435808,
"grad_norm": 0.8457469940185547,
"learning_rate": 0.00019773736437700298,
"loss": 0.3497,
"step": 880
},
{
"epoch": 0.03436426116838488,
"grad_norm": 5.251153945922852,
"learning_rate": 0.0001977116233574012,
"loss": 0.6299,
"step": 890
},
{
"epoch": 0.03475037646241168,
"grad_norm": 4.057605266571045,
"learning_rate": 0.0001976858823377994,
"loss": 0.5829,
"step": 900
},
{
"epoch": 0.03513649175643847,
"grad_norm": 7.625199794769287,
"learning_rate": 0.00019766014131819762,
"loss": 0.452,
"step": 910
},
{
"epoch": 0.03552260705046527,
"grad_norm": 4.618866443634033,
"learning_rate": 0.00019763440029859584,
"loss": 0.5693,
"step": 920
},
{
"epoch": 0.035908722344492064,
"grad_norm": 8.480955123901367,
"learning_rate": 0.00019760865927899405,
"loss": 0.4894,
"step": 930
},
{
"epoch": 0.03629483763851886,
"grad_norm": 6.537581920623779,
"learning_rate": 0.00019758291825939226,
"loss": 0.7396,
"step": 940
},
{
"epoch": 0.03668095293254566,
"grad_norm": 8.093205451965332,
"learning_rate": 0.00019755717723979048,
"loss": 0.3702,
"step": 950
},
{
"epoch": 0.03706706822657246,
"grad_norm": 2.523141860961914,
"learning_rate": 0.0001975314362201887,
"loss": 0.3746,
"step": 960
},
{
"epoch": 0.03745318352059925,
"grad_norm": 1.2707194089889526,
"learning_rate": 0.0001975056952005869,
"loss": 0.3211,
"step": 970
},
{
"epoch": 0.03783929881462605,
"grad_norm": 2.818798780441284,
"learning_rate": 0.00019747995418098512,
"loss": 0.6594,
"step": 980
},
{
"epoch": 0.03822541410865284,
"grad_norm": 7.373154640197754,
"learning_rate": 0.00019745421316138333,
"loss": 0.7825,
"step": 990
},
{
"epoch": 0.03861152940267964,
"grad_norm": 2.8925669193267822,
"learning_rate": 0.00019742847214178154,
"loss": 0.4847,
"step": 1000
},
{
"epoch": 0.038997644696706434,
"grad_norm": 10.87833023071289,
"learning_rate": 0.00019740273112217976,
"loss": 0.5098,
"step": 1010
},
{
"epoch": 0.039383759990733236,
"grad_norm": 3.8262317180633545,
"learning_rate": 0.00019737699010257797,
"loss": 0.5168,
"step": 1020
},
{
"epoch": 0.03976987528476003,
"grad_norm": 2.5567004680633545,
"learning_rate": 0.00019735124908297618,
"loss": 0.5597,
"step": 1030
},
{
"epoch": 0.04015599057878683,
"grad_norm": 4.385695457458496,
"learning_rate": 0.0001973255080633744,
"loss": 0.3187,
"step": 1040
},
{
"epoch": 0.04054210587281362,
"grad_norm": 1.8186907768249512,
"learning_rate": 0.0001972997670437726,
"loss": 0.6274,
"step": 1050
},
{
"epoch": 0.04092822116684042,
"grad_norm": 7.446481704711914,
"learning_rate": 0.00019727402602417082,
"loss": 0.4365,
"step": 1060
},
{
"epoch": 0.04131433646086721,
"grad_norm": 8.973576545715332,
"learning_rate": 0.00019724828500456904,
"loss": 0.6518,
"step": 1070
},
{
"epoch": 0.04170045175489401,
"grad_norm": 4.984101295471191,
"learning_rate": 0.00019722254398496725,
"loss": 0.4694,
"step": 1080
},
{
"epoch": 0.04208656704892081,
"grad_norm": 8.2625093460083,
"learning_rate": 0.00019719680296536546,
"loss": 0.5532,
"step": 1090
},
{
"epoch": 0.042472682342947606,
"grad_norm": 2.2425265312194824,
"learning_rate": 0.0001971710619457637,
"loss": 0.4011,
"step": 1100
},
{
"epoch": 0.0428587976369744,
"grad_norm": 1.4552969932556152,
"learning_rate": 0.0001971453209261619,
"loss": 0.396,
"step": 1110
},
{
"epoch": 0.043244912931001196,
"grad_norm": 2.0675857067108154,
"learning_rate": 0.0001971195799065601,
"loss": 0.3727,
"step": 1120
},
{
"epoch": 0.04363102822502799,
"grad_norm": 5.24460506439209,
"learning_rate": 0.00019709383888695832,
"loss": 0.5016,
"step": 1130
},
{
"epoch": 0.04401714351905479,
"grad_norm": 5.524955749511719,
"learning_rate": 0.00019706809786735653,
"loss": 0.5866,
"step": 1140
},
{
"epoch": 0.04440325881308159,
"grad_norm": 4.2749199867248535,
"learning_rate": 0.00019704235684775474,
"loss": 0.4591,
"step": 1150
},
{
"epoch": 0.044789374107108385,
"grad_norm": 3.460395336151123,
"learning_rate": 0.00019701661582815296,
"loss": 0.5275,
"step": 1160
},
{
"epoch": 0.04517548940113518,
"grad_norm": 1.3979772329330444,
"learning_rate": 0.0001969908748085512,
"loss": 0.3375,
"step": 1170
},
{
"epoch": 0.045561604695161975,
"grad_norm": 2.29923939704895,
"learning_rate": 0.00019696513378894938,
"loss": 0.5683,
"step": 1180
},
{
"epoch": 0.04594771998918877,
"grad_norm": 3.211496353149414,
"learning_rate": 0.0001969393927693476,
"loss": 0.7122,
"step": 1190
},
{
"epoch": 0.046333835283215566,
"grad_norm": 4.18447208404541,
"learning_rate": 0.0001969136517497458,
"loss": 0.5149,
"step": 1200
},
{
"epoch": 0.04671995057724236,
"grad_norm": 14.650918960571289,
"learning_rate": 0.00019688791073014402,
"loss": 0.6384,
"step": 1210
},
{
"epoch": 0.047106065871269164,
"grad_norm": 4.956239700317383,
"learning_rate": 0.00019686216971054226,
"loss": 0.3602,
"step": 1220
},
{
"epoch": 0.04749218116529596,
"grad_norm": 4.0486860275268555,
"learning_rate": 0.00019683642869094045,
"loss": 0.7719,
"step": 1230
},
{
"epoch": 0.047878296459322754,
"grad_norm": 4.587133407592773,
"learning_rate": 0.0001968106876713387,
"loss": 0.4079,
"step": 1240
},
{
"epoch": 0.04826441175334955,
"grad_norm": 0.7830008268356323,
"learning_rate": 0.00019678494665173688,
"loss": 0.5841,
"step": 1250
},
{
"epoch": 0.048650527047376345,
"grad_norm": 5.378068447113037,
"learning_rate": 0.0001967592056321351,
"loss": 0.3226,
"step": 1260
},
{
"epoch": 0.04903664234140314,
"grad_norm": 4.002605438232422,
"learning_rate": 0.0001967334646125333,
"loss": 0.4411,
"step": 1270
},
{
"epoch": 0.04942275763542994,
"grad_norm": 4.695134162902832,
"learning_rate": 0.00019670772359293152,
"loss": 0.415,
"step": 1280
},
{
"epoch": 0.04980887292945674,
"grad_norm": 6.046143054962158,
"learning_rate": 0.00019668198257332976,
"loss": 0.463,
"step": 1290
},
{
"epoch": 0.05019498822348353,
"grad_norm": 1.8300361633300781,
"learning_rate": 0.00019665624155372794,
"loss": 0.408,
"step": 1300
},
{
"epoch": 0.05058110351751033,
"grad_norm": 5.80141544342041,
"learning_rate": 0.00019663050053412618,
"loss": 0.4481,
"step": 1310
},
{
"epoch": 0.050967218811537124,
"grad_norm": 4.103593349456787,
"learning_rate": 0.00019660475951452437,
"loss": 0.3054,
"step": 1320
},
{
"epoch": 0.05135333410556392,
"grad_norm": 9.129929542541504,
"learning_rate": 0.00019657901849492258,
"loss": 0.5554,
"step": 1330
},
{
"epoch": 0.051739449399590715,
"grad_norm": 6.979663372039795,
"learning_rate": 0.0001965532774753208,
"loss": 0.388,
"step": 1340
},
{
"epoch": 0.05212556469361752,
"grad_norm": 6.329915523529053,
"learning_rate": 0.000196527536455719,
"loss": 0.8894,
"step": 1350
},
{
"epoch": 0.05251167998764431,
"grad_norm": 0.2686227262020111,
"learning_rate": 0.00019650179543611725,
"loss": 0.5032,
"step": 1360
},
{
"epoch": 0.05289779528167111,
"grad_norm": 4.818896770477295,
"learning_rate": 0.00019647605441651544,
"loss": 0.3417,
"step": 1370
},
{
"epoch": 0.0532839105756979,
"grad_norm": 6.252008438110352,
"learning_rate": 0.00019645031339691368,
"loss": 0.398,
"step": 1380
},
{
"epoch": 0.0536700258697247,
"grad_norm": 1.2734620571136475,
"learning_rate": 0.00019642457237731186,
"loss": 0.5369,
"step": 1390
},
{
"epoch": 0.054056141163751494,
"grad_norm": 6.511690616607666,
"learning_rate": 0.00019639883135771008,
"loss": 0.4646,
"step": 1400
},
{
"epoch": 0.054442256457778296,
"grad_norm": 3.2352371215820312,
"learning_rate": 0.00019637309033810832,
"loss": 0.3805,
"step": 1410
},
{
"epoch": 0.05482837175180509,
"grad_norm": 1.0574132204055786,
"learning_rate": 0.0001963473493185065,
"loss": 0.4572,
"step": 1420
},
{
"epoch": 0.055214487045831886,
"grad_norm": 2.8525452613830566,
"learning_rate": 0.00019632160829890474,
"loss": 0.4417,
"step": 1430
},
{
"epoch": 0.05560060233985868,
"grad_norm": 3.588179111480713,
"learning_rate": 0.00019629586727930293,
"loss": 0.6214,
"step": 1440
},
{
"epoch": 0.05598671763388548,
"grad_norm": 3.969320058822632,
"learning_rate": 0.00019627012625970117,
"loss": 0.6114,
"step": 1450
},
{
"epoch": 0.05637283292791227,
"grad_norm": 3.465053081512451,
"learning_rate": 0.00019624438524009936,
"loss": 0.6066,
"step": 1460
},
{
"epoch": 0.05675894822193907,
"grad_norm": 3.5419201850891113,
"learning_rate": 0.00019621864422049757,
"loss": 0.3906,
"step": 1470
},
{
"epoch": 0.05714506351596587,
"grad_norm": 3.4580233097076416,
"learning_rate": 0.0001961929032008958,
"loss": 0.9283,
"step": 1480
},
{
"epoch": 0.057531178809992665,
"grad_norm": 4.222144603729248,
"learning_rate": 0.000196167162181294,
"loss": 0.4225,
"step": 1490
},
{
"epoch": 0.05791729410401946,
"grad_norm": 0.8072681427001953,
"learning_rate": 0.00019614142116169224,
"loss": 0.5012,
"step": 1500
},
{
"epoch": 0.058303409398046256,
"grad_norm": 2.827258348464966,
"learning_rate": 0.00019611568014209042,
"loss": 0.4333,
"step": 1510
},
{
"epoch": 0.05868952469207305,
"grad_norm": 1.3494776487350464,
"learning_rate": 0.00019608993912248866,
"loss": 0.3019,
"step": 1520
},
{
"epoch": 0.05907563998609985,
"grad_norm": 4.3279900550842285,
"learning_rate": 0.00019606419810288688,
"loss": 0.4807,
"step": 1530
},
{
"epoch": 0.05946175528012665,
"grad_norm": 3.8996474742889404,
"learning_rate": 0.0001960384570832851,
"loss": 0.4876,
"step": 1540
},
{
"epoch": 0.059847870574153444,
"grad_norm": 5.255978584289551,
"learning_rate": 0.0001960127160636833,
"loss": 0.4661,
"step": 1550
},
{
"epoch": 0.06023398586818024,
"grad_norm": 5.172120094299316,
"learning_rate": 0.0001959869750440815,
"loss": 0.4885,
"step": 1560
},
{
"epoch": 0.060620101162207035,
"grad_norm": 5.385959625244141,
"learning_rate": 0.00019596123402447973,
"loss": 0.2995,
"step": 1570
},
{
"epoch": 0.06100621645623383,
"grad_norm": 3.9922871589660645,
"learning_rate": 0.00019593549300487792,
"loss": 0.4568,
"step": 1580
},
{
"epoch": 0.061392331750260626,
"grad_norm": 6.048642158508301,
"learning_rate": 0.00019590975198527616,
"loss": 0.4649,
"step": 1590
},
{
"epoch": 0.06177844704428742,
"grad_norm": 1.0315563678741455,
"learning_rate": 0.00019588401096567437,
"loss": 0.5175,
"step": 1600
},
{
"epoch": 0.06216456233831422,
"grad_norm": 12.403678894042969,
"learning_rate": 0.00019585826994607258,
"loss": 0.523,
"step": 1610
},
{
"epoch": 0.06255067763234101,
"grad_norm": 6.127188205718994,
"learning_rate": 0.0001958325289264708,
"loss": 0.5861,
"step": 1620
},
{
"epoch": 0.06293679292636781,
"grad_norm": 6.398592948913574,
"learning_rate": 0.00019580678790686898,
"loss": 0.7471,
"step": 1630
},
{
"epoch": 0.06332290822039462,
"grad_norm": 4.127200603485107,
"learning_rate": 0.00019578104688726722,
"loss": 0.4921,
"step": 1640
},
{
"epoch": 0.06370902351442141,
"grad_norm": 4.601541042327881,
"learning_rate": 0.0001957553058676654,
"loss": 0.3682,
"step": 1650
},
{
"epoch": 0.06409513880844821,
"grad_norm": 6.32781457901001,
"learning_rate": 0.00019572956484806365,
"loss": 0.3748,
"step": 1660
},
{
"epoch": 0.064481254102475,
"grad_norm": 3.7280173301696777,
"learning_rate": 0.00019570382382846186,
"loss": 0.5912,
"step": 1670
},
{
"epoch": 0.0648673693965018,
"grad_norm": 6.7821946144104,
"learning_rate": 0.00019567808280886008,
"loss": 0.4073,
"step": 1680
},
{
"epoch": 0.0652534846905286,
"grad_norm": 1.4645791053771973,
"learning_rate": 0.0001956523417892583,
"loss": 0.7164,
"step": 1690
},
{
"epoch": 0.06563959998455539,
"grad_norm": 2.367361545562744,
"learning_rate": 0.00019562660076965648,
"loss": 0.3859,
"step": 1700
},
{
"epoch": 0.06602571527858218,
"grad_norm": 2.198493480682373,
"learning_rate": 0.00019560085975005472,
"loss": 0.4928,
"step": 1710
},
{
"epoch": 0.06641183057260898,
"grad_norm": 1.882567048072815,
"learning_rate": 0.00019557511873045293,
"loss": 0.5861,
"step": 1720
},
{
"epoch": 0.06679794586663577,
"grad_norm": 6.324089527130127,
"learning_rate": 0.00019554937771085114,
"loss": 0.6249,
"step": 1730
},
{
"epoch": 0.06718406116066257,
"grad_norm": 4.283392906188965,
"learning_rate": 0.00019552363669124936,
"loss": 0.5403,
"step": 1740
},
{
"epoch": 0.06757017645468937,
"grad_norm": 4.464428424835205,
"learning_rate": 0.00019549789567164757,
"loss": 0.5815,
"step": 1750
},
{
"epoch": 0.06795629174871616,
"grad_norm": 0.32923218607902527,
"learning_rate": 0.00019547215465204578,
"loss": 0.3791,
"step": 1760
},
{
"epoch": 0.06834240704274297,
"grad_norm": 5.255763053894043,
"learning_rate": 0.00019544641363244397,
"loss": 0.4252,
"step": 1770
},
{
"epoch": 0.06872852233676977,
"grad_norm": 2.1615116596221924,
"learning_rate": 0.0001954206726128422,
"loss": 0.4457,
"step": 1780
},
{
"epoch": 0.06911463763079656,
"grad_norm": 2.0209217071533203,
"learning_rate": 0.00019539493159324042,
"loss": 0.4377,
"step": 1790
},
{
"epoch": 0.06950075292482336,
"grad_norm": 8.12317943572998,
"learning_rate": 0.00019536919057363864,
"loss": 0.4025,
"step": 1800
},
{
"epoch": 0.06988686821885015,
"grad_norm": 1.7034660577774048,
"learning_rate": 0.00019534344955403685,
"loss": 0.2915,
"step": 1810
},
{
"epoch": 0.07027298351287695,
"grad_norm": 4.640942096710205,
"learning_rate": 0.00019531770853443506,
"loss": 0.6588,
"step": 1820
},
{
"epoch": 0.07065909880690374,
"grad_norm": 4.202883243560791,
"learning_rate": 0.00019529196751483328,
"loss": 0.4442,
"step": 1830
},
{
"epoch": 0.07104521410093054,
"grad_norm": 3.26898193359375,
"learning_rate": 0.00019526622649523146,
"loss": 0.5065,
"step": 1840
},
{
"epoch": 0.07143132939495733,
"grad_norm": 8.189995765686035,
"learning_rate": 0.0001952404854756297,
"loss": 0.5258,
"step": 1850
},
{
"epoch": 0.07181744468898413,
"grad_norm": 3.2618284225463867,
"learning_rate": 0.00019521474445602792,
"loss": 0.5037,
"step": 1860
},
{
"epoch": 0.07220355998301092,
"grad_norm": 2.168548583984375,
"learning_rate": 0.00019518900343642613,
"loss": 0.4887,
"step": 1870
},
{
"epoch": 0.07258967527703772,
"grad_norm": 2.2029404640197754,
"learning_rate": 0.00019516326241682434,
"loss": 0.4646,
"step": 1880
},
{
"epoch": 0.07297579057106451,
"grad_norm": 1.561713695526123,
"learning_rate": 0.00019513752139722256,
"loss": 0.432,
"step": 1890
},
{
"epoch": 0.07336190586509132,
"grad_norm": 3.428372621536255,
"learning_rate": 0.00019511178037762077,
"loss": 0.4124,
"step": 1900
},
{
"epoch": 0.07374802115911812,
"grad_norm": 5.6706671714782715,
"learning_rate": 0.00019508603935801898,
"loss": 0.4431,
"step": 1910
},
{
"epoch": 0.07413413645314491,
"grad_norm": 8.872734069824219,
"learning_rate": 0.0001950602983384172,
"loss": 0.7001,
"step": 1920
},
{
"epoch": 0.07452025174717171,
"grad_norm": 1.6821974515914917,
"learning_rate": 0.0001950345573188154,
"loss": 0.4204,
"step": 1930
},
{
"epoch": 0.0749063670411985,
"grad_norm": 3.117480993270874,
"learning_rate": 0.00019500881629921362,
"loss": 0.3748,
"step": 1940
},
{
"epoch": 0.0752924823352253,
"grad_norm": 0.8384984731674194,
"learning_rate": 0.00019498307527961184,
"loss": 0.2636,
"step": 1950
},
{
"epoch": 0.0756785976292521,
"grad_norm": 2.8956708908081055,
"learning_rate": 0.00019495733426001005,
"loss": 0.4514,
"step": 1960
},
{
"epoch": 0.07606471292327889,
"grad_norm": 11.233087539672852,
"learning_rate": 0.00019493159324040826,
"loss": 0.4002,
"step": 1970
},
{
"epoch": 0.07645082821730569,
"grad_norm": 4.066893100738525,
"learning_rate": 0.00019490585222080648,
"loss": 0.4449,
"step": 1980
},
{
"epoch": 0.07683694351133248,
"grad_norm": 4.854077339172363,
"learning_rate": 0.0001948801112012047,
"loss": 0.4961,
"step": 1990
},
{
"epoch": 0.07722305880535928,
"grad_norm": 2.5722827911376953,
"learning_rate": 0.0001948543701816029,
"loss": 0.3743,
"step": 2000
},
{
"epoch": 0.07760917409938607,
"grad_norm": 5.842077255249023,
"learning_rate": 0.00019482862916200112,
"loss": 0.2906,
"step": 2010
},
{
"epoch": 0.07799528939341287,
"grad_norm": 6.163092136383057,
"learning_rate": 0.00019480288814239933,
"loss": 0.4374,
"step": 2020
},
{
"epoch": 0.07838140468743968,
"grad_norm": 4.589334487915039,
"learning_rate": 0.00019477714712279754,
"loss": 0.484,
"step": 2030
},
{
"epoch": 0.07876751998146647,
"grad_norm": 6.951212406158447,
"learning_rate": 0.00019475140610319576,
"loss": 0.5767,
"step": 2040
},
{
"epoch": 0.07915363527549327,
"grad_norm": 3.3662521839141846,
"learning_rate": 0.00019472566508359397,
"loss": 0.5566,
"step": 2050
},
{
"epoch": 0.07953975056952006,
"grad_norm": 1.6602391004562378,
"learning_rate": 0.00019469992406399218,
"loss": 0.1436,
"step": 2060
},
{
"epoch": 0.07992586586354686,
"grad_norm": 6.451857089996338,
"learning_rate": 0.0001946741830443904,
"loss": 0.3778,
"step": 2070
},
{
"epoch": 0.08031198115757365,
"grad_norm": 3.7249560356140137,
"learning_rate": 0.0001946484420247886,
"loss": 0.5391,
"step": 2080
},
{
"epoch": 0.08069809645160045,
"grad_norm": 4.138098239898682,
"learning_rate": 0.00019462270100518682,
"loss": 0.3598,
"step": 2090
},
{
"epoch": 0.08108421174562724,
"grad_norm": 3.224778175354004,
"learning_rate": 0.00019459695998558504,
"loss": 0.2967,
"step": 2100
},
{
"epoch": 0.08147032703965404,
"grad_norm": 1.3951358795166016,
"learning_rate": 0.00019457121896598325,
"loss": 0.2698,
"step": 2110
},
{
"epoch": 0.08185644233368083,
"grad_norm": 5.956802845001221,
"learning_rate": 0.00019454547794638146,
"loss": 0.451,
"step": 2120
},
{
"epoch": 0.08224255762770763,
"grad_norm": 3.456360101699829,
"learning_rate": 0.00019451973692677968,
"loss": 0.3365,
"step": 2130
},
{
"epoch": 0.08262867292173443,
"grad_norm": 2.9433653354644775,
"learning_rate": 0.0001944939959071779,
"loss": 0.4424,
"step": 2140
},
{
"epoch": 0.08301478821576122,
"grad_norm": 3.136000871658325,
"learning_rate": 0.0001944682548875761,
"loss": 0.6224,
"step": 2150
},
{
"epoch": 0.08340090350978802,
"grad_norm": 2.669084310531616,
"learning_rate": 0.00019444251386797432,
"loss": 0.4435,
"step": 2160
},
{
"epoch": 0.08378701880381483,
"grad_norm": 4.573731899261475,
"learning_rate": 0.00019441677284837253,
"loss": 0.5555,
"step": 2170
},
{
"epoch": 0.08417313409784162,
"grad_norm": 6.354156017303467,
"learning_rate": 0.00019439103182877074,
"loss": 0.4232,
"step": 2180
},
{
"epoch": 0.08455924939186842,
"grad_norm": 2.993691921234131,
"learning_rate": 0.00019436529080916895,
"loss": 0.51,
"step": 2190
},
{
"epoch": 0.08494536468589521,
"grad_norm": 3.6496782302856445,
"learning_rate": 0.00019433954978956717,
"loss": 0.4031,
"step": 2200
},
{
"epoch": 0.085331479979922,
"grad_norm": 1.9039051532745361,
"learning_rate": 0.00019431380876996538,
"loss": 0.4407,
"step": 2210
},
{
"epoch": 0.0857175952739488,
"grad_norm": 2.3211915493011475,
"learning_rate": 0.00019428806775036362,
"loss": 0.4057,
"step": 2220
},
{
"epoch": 0.0861037105679756,
"grad_norm": 4.883905410766602,
"learning_rate": 0.0001942623267307618,
"loss": 0.6223,
"step": 2230
},
{
"epoch": 0.08648982586200239,
"grad_norm": 2.164484977722168,
"learning_rate": 0.00019423658571116005,
"loss": 0.3141,
"step": 2240
},
{
"epoch": 0.08687594115602919,
"grad_norm": 2.2078909873962402,
"learning_rate": 0.00019421084469155823,
"loss": 0.3523,
"step": 2250
},
{
"epoch": 0.08726205645005598,
"grad_norm": 1.0987967252731323,
"learning_rate": 0.00019418510367195645,
"loss": 0.4013,
"step": 2260
},
{
"epoch": 0.08764817174408278,
"grad_norm": 2.418327569961548,
"learning_rate": 0.00019415936265235466,
"loss": 0.581,
"step": 2270
},
{
"epoch": 0.08803428703810957,
"grad_norm": 4.029652118682861,
"learning_rate": 0.00019413362163275287,
"loss": 0.5244,
"step": 2280
},
{
"epoch": 0.08842040233213637,
"grad_norm": 3.4661777019500732,
"learning_rate": 0.00019410788061315112,
"loss": 0.4531,
"step": 2290
},
{
"epoch": 0.08880651762616318,
"grad_norm": 1.0475856065750122,
"learning_rate": 0.0001940821395935493,
"loss": 0.4362,
"step": 2300
},
{
"epoch": 0.08919263292018997,
"grad_norm": 4.021854400634766,
"learning_rate": 0.00019405639857394754,
"loss": 0.4532,
"step": 2310
},
{
"epoch": 0.08957874821421677,
"grad_norm": 1.836438536643982,
"learning_rate": 0.00019403065755434573,
"loss": 0.6858,
"step": 2320
},
{
"epoch": 0.08996486350824356,
"grad_norm": 2.5251567363739014,
"learning_rate": 0.00019400491653474394,
"loss": 0.3619,
"step": 2330
},
{
"epoch": 0.09035097880227036,
"grad_norm": 3.067208766937256,
"learning_rate": 0.00019397917551514215,
"loss": 0.7376,
"step": 2340
},
{
"epoch": 0.09073709409629716,
"grad_norm": 0.9124518036842346,
"learning_rate": 0.00019395343449554037,
"loss": 0.4193,
"step": 2350
},
{
"epoch": 0.09112320939032395,
"grad_norm": 3.8170812129974365,
"learning_rate": 0.0001939276934759386,
"loss": 0.5393,
"step": 2360
},
{
"epoch": 0.09150932468435075,
"grad_norm": 8.19250202178955,
"learning_rate": 0.0001939019524563368,
"loss": 0.424,
"step": 2370
},
{
"epoch": 0.09189543997837754,
"grad_norm": 4.459112167358398,
"learning_rate": 0.00019387621143673503,
"loss": 0.3278,
"step": 2380
},
{
"epoch": 0.09228155527240434,
"grad_norm": 5.578339576721191,
"learning_rate": 0.00019385047041713322,
"loss": 0.5223,
"step": 2390
},
{
"epoch": 0.09266767056643113,
"grad_norm": 1.3707878589630127,
"learning_rate": 0.00019382472939753143,
"loss": 0.3004,
"step": 2400
},
{
"epoch": 0.09305378586045793,
"grad_norm": 5.0041184425354,
"learning_rate": 0.00019379898837792967,
"loss": 0.4378,
"step": 2410
},
{
"epoch": 0.09343990115448472,
"grad_norm": 5.668384552001953,
"learning_rate": 0.00019377324735832786,
"loss": 0.499,
"step": 2420
},
{
"epoch": 0.09382601644851153,
"grad_norm": 5.605838775634766,
"learning_rate": 0.0001937475063387261,
"loss": 0.5642,
"step": 2430
},
{
"epoch": 0.09421213174253833,
"grad_norm": 5.055904865264893,
"learning_rate": 0.0001937217653191243,
"loss": 0.6225,
"step": 2440
},
{
"epoch": 0.09459824703656512,
"grad_norm": 3.1779348850250244,
"learning_rate": 0.00019369602429952253,
"loss": 0.3673,
"step": 2450
},
{
"epoch": 0.09498436233059192,
"grad_norm": 2.540269136428833,
"learning_rate": 0.00019367028327992071,
"loss": 0.3499,
"step": 2460
},
{
"epoch": 0.09537047762461871,
"grad_norm": 2.284114122390747,
"learning_rate": 0.00019364454226031893,
"loss": 0.5405,
"step": 2470
},
{
"epoch": 0.09575659291864551,
"grad_norm": 6.752682685852051,
"learning_rate": 0.00019361880124071717,
"loss": 0.5594,
"step": 2480
},
{
"epoch": 0.0961427082126723,
"grad_norm": 2.5659310817718506,
"learning_rate": 0.00019359306022111535,
"loss": 0.5872,
"step": 2490
},
{
"epoch": 0.0965288235066991,
"grad_norm": 4.503110885620117,
"learning_rate": 0.0001935673192015136,
"loss": 0.5171,
"step": 2500
},
{
"epoch": 0.0969149388007259,
"grad_norm": 1.7715080976486206,
"learning_rate": 0.00019354157818191178,
"loss": 0.6131,
"step": 2510
},
{
"epoch": 0.09730105409475269,
"grad_norm": 4.479000568389893,
"learning_rate": 0.00019351583716231002,
"loss": 0.4396,
"step": 2520
},
{
"epoch": 0.09768716938877949,
"grad_norm": 0.9424387812614441,
"learning_rate": 0.00019349009614270823,
"loss": 0.3166,
"step": 2530
},
{
"epoch": 0.09807328468280628,
"grad_norm": 3.792689800262451,
"learning_rate": 0.00019346435512310642,
"loss": 0.2591,
"step": 2540
},
{
"epoch": 0.09845939997683308,
"grad_norm": 2.9132003784179688,
"learning_rate": 0.00019343861410350466,
"loss": 0.3523,
"step": 2550
},
{
"epoch": 0.09884551527085988,
"grad_norm": 1.6421749591827393,
"learning_rate": 0.00019341287308390285,
"loss": 0.2805,
"step": 2560
},
{
"epoch": 0.09923163056488668,
"grad_norm": 1.9469813108444214,
"learning_rate": 0.0001933871320643011,
"loss": 0.3929,
"step": 2570
},
{
"epoch": 0.09961774585891348,
"grad_norm": 3.081062078475952,
"learning_rate": 0.00019336139104469927,
"loss": 0.569,
"step": 2580
},
{
"epoch": 0.10000386115294027,
"grad_norm": 4.728143215179443,
"learning_rate": 0.00019333565002509751,
"loss": 0.3443,
"step": 2590
},
{
"epoch": 0.10038997644696707,
"grad_norm": 2.4117422103881836,
"learning_rate": 0.00019330990900549573,
"loss": 0.4492,
"step": 2600
},
{
"epoch": 0.10077609174099386,
"grad_norm": 5.794168472290039,
"learning_rate": 0.00019328416798589391,
"loss": 0.5088,
"step": 2610
},
{
"epoch": 0.10116220703502066,
"grad_norm": 1.0662094354629517,
"learning_rate": 0.00019325842696629215,
"loss": 0.4683,
"step": 2620
},
{
"epoch": 0.10154832232904745,
"grad_norm": 2.13590931892395,
"learning_rate": 0.00019323268594669034,
"loss": 0.5833,
"step": 2630
},
{
"epoch": 0.10193443762307425,
"grad_norm": 7.52834415435791,
"learning_rate": 0.00019320694492708858,
"loss": 0.54,
"step": 2640
},
{
"epoch": 0.10232055291710104,
"grad_norm": 6.155847072601318,
"learning_rate": 0.00019318120390748677,
"loss": 0.6809,
"step": 2650
},
{
"epoch": 0.10270666821112784,
"grad_norm": 6.527890205383301,
"learning_rate": 0.000193155462887885,
"loss": 0.5239,
"step": 2660
},
{
"epoch": 0.10309278350515463,
"grad_norm": 3.3918986320495605,
"learning_rate": 0.00019312972186828322,
"loss": 0.4636,
"step": 2670
},
{
"epoch": 0.10347889879918143,
"grad_norm": 2.0933191776275635,
"learning_rate": 0.0001931039808486814,
"loss": 0.5288,
"step": 2680
},
{
"epoch": 0.10386501409320824,
"grad_norm": 2.4386465549468994,
"learning_rate": 0.00019307823982907965,
"loss": 0.5496,
"step": 2690
},
{
"epoch": 0.10425112938723503,
"grad_norm": 2.885315418243408,
"learning_rate": 0.00019305249880947783,
"loss": 0.5928,
"step": 2700
},
{
"epoch": 0.10463724468126183,
"grad_norm": 4.986598968505859,
"learning_rate": 0.00019302675778987607,
"loss": 0.3513,
"step": 2710
},
{
"epoch": 0.10502335997528862,
"grad_norm": 2.7999277114868164,
"learning_rate": 0.0001930010167702743,
"loss": 0.4484,
"step": 2720
},
{
"epoch": 0.10540947526931542,
"grad_norm": 1.6467676162719727,
"learning_rate": 0.0001929752757506725,
"loss": 0.4729,
"step": 2730
},
{
"epoch": 0.10579559056334222,
"grad_norm": 2.168477773666382,
"learning_rate": 0.00019294953473107071,
"loss": 0.3579,
"step": 2740
},
{
"epoch": 0.10618170585736901,
"grad_norm": 1.5439807176589966,
"learning_rate": 0.00019292379371146893,
"loss": 0.5715,
"step": 2750
},
{
"epoch": 0.1065678211513958,
"grad_norm": 1.2175947427749634,
"learning_rate": 0.00019289805269186714,
"loss": 0.4059,
"step": 2760
},
{
"epoch": 0.1069539364454226,
"grad_norm": 2.649782419204712,
"learning_rate": 0.00019287231167226533,
"loss": 0.7864,
"step": 2770
},
{
"epoch": 0.1073400517394494,
"grad_norm": 3.698002815246582,
"learning_rate": 0.00019284657065266357,
"loss": 0.5107,
"step": 2780
},
{
"epoch": 0.10772616703347619,
"grad_norm": 1.7418729066848755,
"learning_rate": 0.00019282082963306178,
"loss": 0.332,
"step": 2790
},
{
"epoch": 0.10811228232750299,
"grad_norm": 2.2264151573181152,
"learning_rate": 0.00019279508861346,
"loss": 0.3944,
"step": 2800
},
{
"epoch": 0.10849839762152978,
"grad_norm": 1.3407092094421387,
"learning_rate": 0.0001927693475938582,
"loss": 0.4603,
"step": 2810
},
{
"epoch": 0.10888451291555659,
"grad_norm": 0.3719189763069153,
"learning_rate": 0.00019274360657425642,
"loss": 0.496,
"step": 2820
},
{
"epoch": 0.10927062820958339,
"grad_norm": 4.728814125061035,
"learning_rate": 0.00019271786555465463,
"loss": 0.4079,
"step": 2830
},
{
"epoch": 0.10965674350361018,
"grad_norm": 5.007620334625244,
"learning_rate": 0.00019269212453505285,
"loss": 0.5221,
"step": 2840
},
{
"epoch": 0.11004285879763698,
"grad_norm": 2.7476320266723633,
"learning_rate": 0.00019266638351545106,
"loss": 0.3705,
"step": 2850
},
{
"epoch": 0.11042897409166377,
"grad_norm": 2.2960126399993896,
"learning_rate": 0.00019264064249584927,
"loss": 0.6656,
"step": 2860
},
{
"epoch": 0.11081508938569057,
"grad_norm": 0.9589812755584717,
"learning_rate": 0.0001926149014762475,
"loss": 0.6977,
"step": 2870
},
{
"epoch": 0.11120120467971736,
"grad_norm": 2.274984121322632,
"learning_rate": 0.0001925891604566457,
"loss": 0.4237,
"step": 2880
},
{
"epoch": 0.11158731997374416,
"grad_norm": 1.8849111795425415,
"learning_rate": 0.00019256341943704391,
"loss": 0.2357,
"step": 2890
},
{
"epoch": 0.11197343526777095,
"grad_norm": 2.7264740467071533,
"learning_rate": 0.00019253767841744213,
"loss": 0.3422,
"step": 2900
},
{
"epoch": 0.11235955056179775,
"grad_norm": 3.832827568054199,
"learning_rate": 0.00019251193739784034,
"loss": 0.3861,
"step": 2910
},
{
"epoch": 0.11274566585582455,
"grad_norm": 2.3612313270568848,
"learning_rate": 0.00019248619637823855,
"loss": 0.3299,
"step": 2920
},
{
"epoch": 0.11313178114985134,
"grad_norm": 2.2509396076202393,
"learning_rate": 0.00019246045535863677,
"loss": 0.6027,
"step": 2930
},
{
"epoch": 0.11351789644387814,
"grad_norm": 2.7600464820861816,
"learning_rate": 0.00019243471433903498,
"loss": 0.3864,
"step": 2940
},
{
"epoch": 0.11390401173790494,
"grad_norm": 5.956289768218994,
"learning_rate": 0.0001924089733194332,
"loss": 0.4669,
"step": 2950
},
{
"epoch": 0.11429012703193174,
"grad_norm": 4.651761531829834,
"learning_rate": 0.0001923832322998314,
"loss": 0.5774,
"step": 2960
},
{
"epoch": 0.11467624232595854,
"grad_norm": 1.1770590543746948,
"learning_rate": 0.00019235749128022962,
"loss": 0.3951,
"step": 2970
},
{
"epoch": 0.11506235761998533,
"grad_norm": 0.8117956519126892,
"learning_rate": 0.00019233175026062783,
"loss": 0.3919,
"step": 2980
},
{
"epoch": 0.11544847291401213,
"grad_norm": 1.318812608718872,
"learning_rate": 0.00019230600924102605,
"loss": 0.2905,
"step": 2990
},
{
"epoch": 0.11583458820803892,
"grad_norm": 1.981382966041565,
"learning_rate": 0.00019228026822142426,
"loss": 0.5312,
"step": 3000
},
{
"epoch": 0.11622070350206572,
"grad_norm": 1.9083003997802734,
"learning_rate": 0.00019225452720182247,
"loss": 0.3129,
"step": 3010
},
{
"epoch": 0.11660681879609251,
"grad_norm": 2.7581653594970703,
"learning_rate": 0.0001922287861822207,
"loss": 0.3954,
"step": 3020
},
{
"epoch": 0.11699293409011931,
"grad_norm": 1.065090537071228,
"learning_rate": 0.0001922030451626189,
"loss": 0.2722,
"step": 3030
},
{
"epoch": 0.1173790493841461,
"grad_norm": 0.600864827632904,
"learning_rate": 0.0001921773041430171,
"loss": 0.493,
"step": 3040
},
{
"epoch": 0.1177651646781729,
"grad_norm": 4.4449052810668945,
"learning_rate": 0.00019215156312341533,
"loss": 0.4982,
"step": 3050
},
{
"epoch": 0.1181512799721997,
"grad_norm": 3.34476637840271,
"learning_rate": 0.00019212582210381354,
"loss": 0.3204,
"step": 3060
},
{
"epoch": 0.11853739526622649,
"grad_norm": 1.7432445287704468,
"learning_rate": 0.00019210008108421175,
"loss": 0.6601,
"step": 3070
},
{
"epoch": 0.1189235105602533,
"grad_norm": 1.908324956893921,
"learning_rate": 0.00019207434006460997,
"loss": 0.5947,
"step": 3080
},
{
"epoch": 0.1193096258542801,
"grad_norm": 5.373056888580322,
"learning_rate": 0.00019204859904500818,
"loss": 0.5169,
"step": 3090
},
{
"epoch": 0.11969574114830689,
"grad_norm": 0.861535370349884,
"learning_rate": 0.0001920228580254064,
"loss": 0.3829,
"step": 3100
},
{
"epoch": 0.12008185644233368,
"grad_norm": 1.2700462341308594,
"learning_rate": 0.0001919971170058046,
"loss": 0.4475,
"step": 3110
},
{
"epoch": 0.12046797173636048,
"grad_norm": 2.9959444999694824,
"learning_rate": 0.00019197137598620282,
"loss": 0.4704,
"step": 3120
},
{
"epoch": 0.12085408703038727,
"grad_norm": 0.280109167098999,
"learning_rate": 0.00019194563496660103,
"loss": 0.3732,
"step": 3130
},
{
"epoch": 0.12124020232441407,
"grad_norm": 0.9746024012565613,
"learning_rate": 0.00019191989394699925,
"loss": 0.4693,
"step": 3140
},
{
"epoch": 0.12162631761844087,
"grad_norm": 1.7267721891403198,
"learning_rate": 0.00019189415292739746,
"loss": 0.4509,
"step": 3150
},
{
"epoch": 0.12201243291246766,
"grad_norm": 2.1759033203125,
"learning_rate": 0.00019186841190779567,
"loss": 0.428,
"step": 3160
},
{
"epoch": 0.12239854820649446,
"grad_norm": 1.270711064338684,
"learning_rate": 0.0001918426708881939,
"loss": 0.4262,
"step": 3170
},
{
"epoch": 0.12278466350052125,
"grad_norm": 3.7549123764038086,
"learning_rate": 0.0001918169298685921,
"loss": 0.4758,
"step": 3180
},
{
"epoch": 0.12317077879454805,
"grad_norm": 1.6550017595291138,
"learning_rate": 0.0001917911888489903,
"loss": 0.452,
"step": 3190
},
{
"epoch": 0.12355689408857484,
"grad_norm": 3.7151713371276855,
"learning_rate": 0.00019176544782938853,
"loss": 0.4844,
"step": 3200
},
{
"epoch": 0.12394300938260165,
"grad_norm": 0.5354440808296204,
"learning_rate": 0.00019173970680978674,
"loss": 0.4432,
"step": 3210
},
{
"epoch": 0.12432912467662845,
"grad_norm": 3.2494261264801025,
"learning_rate": 0.00019171396579018498,
"loss": 0.587,
"step": 3220
},
{
"epoch": 0.12471523997065524,
"grad_norm": 1.2129877805709839,
"learning_rate": 0.00019168822477058317,
"loss": 0.4662,
"step": 3230
},
{
"epoch": 0.12510135526468202,
"grad_norm": 3.723402500152588,
"learning_rate": 0.00019166248375098138,
"loss": 0.5261,
"step": 3240
},
{
"epoch": 0.12548747055870882,
"grad_norm": 1.596259593963623,
"learning_rate": 0.0001916367427313796,
"loss": 0.2802,
"step": 3250
},
{
"epoch": 0.12587358585273561,
"grad_norm": 5.5710320472717285,
"learning_rate": 0.0001916110017117778,
"loss": 0.5246,
"step": 3260
},
{
"epoch": 0.1262597011467624,
"grad_norm": 4.490183353424072,
"learning_rate": 0.00019158526069217602,
"loss": 0.4929,
"step": 3270
},
{
"epoch": 0.12664581644078923,
"grad_norm": 2.482572555541992,
"learning_rate": 0.00019155951967257423,
"loss": 0.3677,
"step": 3280
},
{
"epoch": 0.12703193173481603,
"grad_norm": 3.348520517349243,
"learning_rate": 0.00019153377865297247,
"loss": 0.6471,
"step": 3290
},
{
"epoch": 0.12741804702884282,
"grad_norm": 7.735306262969971,
"learning_rate": 0.00019150803763337066,
"loss": 0.6057,
"step": 3300
},
{
"epoch": 0.12780416232286962,
"grad_norm": 2.120649576187134,
"learning_rate": 0.00019148229661376887,
"loss": 0.5408,
"step": 3310
},
{
"epoch": 0.12819027761689641,
"grad_norm": 10.259540557861328,
"learning_rate": 0.00019145655559416709,
"loss": 0.4753,
"step": 3320
},
{
"epoch": 0.1285763929109232,
"grad_norm": 4.094576358795166,
"learning_rate": 0.0001914308145745653,
"loss": 0.3832,
"step": 3330
},
{
"epoch": 0.12896250820495,
"grad_norm": 3.1248559951782227,
"learning_rate": 0.00019140507355496354,
"loss": 0.5631,
"step": 3340
},
{
"epoch": 0.1293486234989768,
"grad_norm": 1.2975168228149414,
"learning_rate": 0.00019137933253536173,
"loss": 0.5158,
"step": 3350
},
{
"epoch": 0.1297347387930036,
"grad_norm": 3.2515244483947754,
"learning_rate": 0.00019135359151575997,
"loss": 0.4176,
"step": 3360
},
{
"epoch": 0.1301208540870304,
"grad_norm": 2.287757396697998,
"learning_rate": 0.00019132785049615815,
"loss": 0.5316,
"step": 3370
},
{
"epoch": 0.1305069693810572,
"grad_norm": 8.668967247009277,
"learning_rate": 0.00019130210947655637,
"loss": 0.6653,
"step": 3380
},
{
"epoch": 0.13089308467508398,
"grad_norm": 4.751536846160889,
"learning_rate": 0.00019127636845695458,
"loss": 0.4508,
"step": 3390
},
{
"epoch": 0.13127919996911078,
"grad_norm": 3.240792751312256,
"learning_rate": 0.0001912506274373528,
"loss": 0.441,
"step": 3400
},
{
"epoch": 0.13166531526313757,
"grad_norm": 2.146261215209961,
"learning_rate": 0.00019122488641775103,
"loss": 0.3394,
"step": 3410
},
{
"epoch": 0.13205143055716437,
"grad_norm": 2.259693145751953,
"learning_rate": 0.00019119914539814922,
"loss": 0.4348,
"step": 3420
},
{
"epoch": 0.13243754585119116,
"grad_norm": 1.8136098384857178,
"learning_rate": 0.00019117340437854746,
"loss": 0.4441,
"step": 3430
},
{
"epoch": 0.13282366114521796,
"grad_norm": 1.7324503660202026,
"learning_rate": 0.00019114766335894565,
"loss": 0.4725,
"step": 3440
},
{
"epoch": 0.13320977643924475,
"grad_norm": 4.709383487701416,
"learning_rate": 0.00019112192233934389,
"loss": 0.5383,
"step": 3450
},
{
"epoch": 0.13359589173327155,
"grad_norm": 0.3468118906021118,
"learning_rate": 0.00019109618131974207,
"loss": 0.3228,
"step": 3460
},
{
"epoch": 0.13398200702729834,
"grad_norm": 6.66448974609375,
"learning_rate": 0.00019107044030014029,
"loss": 0.4128,
"step": 3470
},
{
"epoch": 0.13436812232132514,
"grad_norm": 1.6971935033798218,
"learning_rate": 0.00019104469928053853,
"loss": 0.5061,
"step": 3480
},
{
"epoch": 0.13475423761535194,
"grad_norm": 0.9180198311805725,
"learning_rate": 0.0001910189582609367,
"loss": 0.4445,
"step": 3490
},
{
"epoch": 0.13514035290937873,
"grad_norm": 3.1244235038757324,
"learning_rate": 0.00019099321724133495,
"loss": 0.4099,
"step": 3500
},
{
"epoch": 0.13552646820340553,
"grad_norm": 2.7192864418029785,
"learning_rate": 0.00019096747622173314,
"loss": 0.5004,
"step": 3510
},
{
"epoch": 0.13591258349743232,
"grad_norm": 2.0905699729919434,
"learning_rate": 0.00019094173520213138,
"loss": 0.4277,
"step": 3520
},
{
"epoch": 0.13629869879145912,
"grad_norm": 3.3753092288970947,
"learning_rate": 0.0001909159941825296,
"loss": 0.4129,
"step": 3530
},
{
"epoch": 0.13668481408548594,
"grad_norm": 4.199211120605469,
"learning_rate": 0.00019089025316292778,
"loss": 0.5734,
"step": 3540
},
{
"epoch": 0.13707092937951273,
"grad_norm": 2.0411245822906494,
"learning_rate": 0.00019086451214332602,
"loss": 0.3018,
"step": 3550
},
{
"epoch": 0.13745704467353953,
"grad_norm": 17.236717224121094,
"learning_rate": 0.0001908387711237242,
"loss": 0.4527,
"step": 3560
},
{
"epoch": 0.13784315996756633,
"grad_norm": 1.4575644731521606,
"learning_rate": 0.00019081303010412245,
"loss": 0.3773,
"step": 3570
},
{
"epoch": 0.13822927526159312,
"grad_norm": 3.926090717315674,
"learning_rate": 0.00019078728908452063,
"loss": 0.5316,
"step": 3580
},
{
"epoch": 0.13861539055561992,
"grad_norm": 3.1841864585876465,
"learning_rate": 0.00019076154806491887,
"loss": 0.3705,
"step": 3590
},
{
"epoch": 0.1390015058496467,
"grad_norm": 4.08506441116333,
"learning_rate": 0.00019073580704531709,
"loss": 0.4941,
"step": 3600
},
{
"epoch": 0.1393876211436735,
"grad_norm": 3.063154458999634,
"learning_rate": 0.00019071006602571527,
"loss": 0.4435,
"step": 3610
},
{
"epoch": 0.1397737364377003,
"grad_norm": 6.122230529785156,
"learning_rate": 0.0001906843250061135,
"loss": 0.5067,
"step": 3620
},
{
"epoch": 0.1401598517317271,
"grad_norm": 3.3089540004730225,
"learning_rate": 0.0001906585839865117,
"loss": 0.4329,
"step": 3630
},
{
"epoch": 0.1405459670257539,
"grad_norm": 1.7245008945465088,
"learning_rate": 0.00019063284296690994,
"loss": 0.4502,
"step": 3640
},
{
"epoch": 0.1409320823197807,
"grad_norm": 1.7759568691253662,
"learning_rate": 0.00019060710194730813,
"loss": 0.2379,
"step": 3650
},
{
"epoch": 0.14131819761380748,
"grad_norm": 0.432452529668808,
"learning_rate": 0.00019058136092770637,
"loss": 0.4277,
"step": 3660
},
{
"epoch": 0.14170431290783428,
"grad_norm": 3.311952829360962,
"learning_rate": 0.00019055561990810458,
"loss": 0.4558,
"step": 3670
},
{
"epoch": 0.14209042820186107,
"grad_norm": 1.9942964315414429,
"learning_rate": 0.00019052987888850277,
"loss": 0.3349,
"step": 3680
},
{
"epoch": 0.14247654349588787,
"grad_norm": 6.226424217224121,
"learning_rate": 0.000190504137868901,
"loss": 0.5809,
"step": 3690
},
{
"epoch": 0.14286265878991466,
"grad_norm": 6.223634719848633,
"learning_rate": 0.0001904783968492992,
"loss": 0.5788,
"step": 3700
},
{
"epoch": 0.14324877408394146,
"grad_norm": 0.7370914220809937,
"learning_rate": 0.00019045265582969743,
"loss": 0.4834,
"step": 3710
},
{
"epoch": 0.14363488937796826,
"grad_norm": 1.745880365371704,
"learning_rate": 0.00019042691481009565,
"loss": 0.6995,
"step": 3720
},
{
"epoch": 0.14402100467199505,
"grad_norm": 0.8839595913887024,
"learning_rate": 0.00019040117379049386,
"loss": 0.3526,
"step": 3730
},
{
"epoch": 0.14440711996602185,
"grad_norm": 1.1224008798599243,
"learning_rate": 0.00019037543277089207,
"loss": 0.3558,
"step": 3740
},
{
"epoch": 0.14479323526004864,
"grad_norm": 1.0473041534423828,
"learning_rate": 0.00019034969175129026,
"loss": 0.2465,
"step": 3750
},
{
"epoch": 0.14517935055407544,
"grad_norm": 3.83192777633667,
"learning_rate": 0.0001903239507316885,
"loss": 0.4832,
"step": 3760
},
{
"epoch": 0.14556546584810223,
"grad_norm": 3.323885440826416,
"learning_rate": 0.00019029820971208669,
"loss": 0.4924,
"step": 3770
},
{
"epoch": 0.14595158114212903,
"grad_norm": 3.2334187030792236,
"learning_rate": 0.00019027246869248493,
"loss": 0.5053,
"step": 3780
},
{
"epoch": 0.14633769643615582,
"grad_norm": 2.280498743057251,
"learning_rate": 0.00019024672767288314,
"loss": 0.554,
"step": 3790
},
{
"epoch": 0.14672381173018265,
"grad_norm": 4.546648979187012,
"learning_rate": 0.00019022098665328135,
"loss": 0.3999,
"step": 3800
},
{
"epoch": 0.14710992702420944,
"grad_norm": 0.6303244829177856,
"learning_rate": 0.00019019524563367957,
"loss": 0.4481,
"step": 3810
},
{
"epoch": 0.14749604231823624,
"grad_norm": 2.605196475982666,
"learning_rate": 0.00019016950461407775,
"loss": 0.3561,
"step": 3820
},
{
"epoch": 0.14788215761226303,
"grad_norm": 3.0562639236450195,
"learning_rate": 0.000190143763594476,
"loss": 0.5903,
"step": 3830
},
{
"epoch": 0.14826827290628983,
"grad_norm": 11.164155006408691,
"learning_rate": 0.0001901180225748742,
"loss": 0.4299,
"step": 3840
},
{
"epoch": 0.14865438820031662,
"grad_norm": 4.996811866760254,
"learning_rate": 0.00019009228155527242,
"loss": 0.4423,
"step": 3850
},
{
"epoch": 0.14904050349434342,
"grad_norm": 2.627272844314575,
"learning_rate": 0.00019006654053567063,
"loss": 0.4875,
"step": 3860
},
{
"epoch": 0.1494266187883702,
"grad_norm": 2.6532809734344482,
"learning_rate": 0.00019004079951606885,
"loss": 0.5221,
"step": 3870
},
{
"epoch": 0.149812734082397,
"grad_norm": 5.821976661682129,
"learning_rate": 0.00019001505849646706,
"loss": 0.4793,
"step": 3880
},
{
"epoch": 0.1501988493764238,
"grad_norm": 2.888029098510742,
"learning_rate": 0.00018998931747686524,
"loss": 0.5784,
"step": 3890
},
{
"epoch": 0.1505849646704506,
"grad_norm": 0.9147624969482422,
"learning_rate": 0.00018996357645726349,
"loss": 0.5533,
"step": 3900
},
{
"epoch": 0.1509710799644774,
"grad_norm": 2.6088199615478516,
"learning_rate": 0.0001899378354376617,
"loss": 0.5028,
"step": 3910
},
{
"epoch": 0.1513571952585042,
"grad_norm": 3.8208296298980713,
"learning_rate": 0.0001899120944180599,
"loss": 0.4934,
"step": 3920
},
{
"epoch": 0.15174331055253099,
"grad_norm": 2.8711328506469727,
"learning_rate": 0.00018988635339845813,
"loss": 0.4417,
"step": 3930
},
{
"epoch": 0.15212942584655778,
"grad_norm": 2.922855854034424,
"learning_rate": 0.00018986061237885634,
"loss": 0.5303,
"step": 3940
},
{
"epoch": 0.15251554114058458,
"grad_norm": 2.52575945854187,
"learning_rate": 0.00018983487135925455,
"loss": 0.397,
"step": 3950
},
{
"epoch": 0.15290165643461137,
"grad_norm": 3.3369996547698975,
"learning_rate": 0.00018980913033965277,
"loss": 0.4172,
"step": 3960
},
{
"epoch": 0.15328777172863817,
"grad_norm": 1.7678214311599731,
"learning_rate": 0.00018978338932005098,
"loss": 0.3122,
"step": 3970
},
{
"epoch": 0.15367388702266496,
"grad_norm": 3.3293211460113525,
"learning_rate": 0.0001897576483004492,
"loss": 0.6864,
"step": 3980
},
{
"epoch": 0.15406000231669176,
"grad_norm": 1.4911530017852783,
"learning_rate": 0.0001897319072808474,
"loss": 0.3888,
"step": 3990
},
{
"epoch": 0.15444611761071855,
"grad_norm": 1.4884055852890015,
"learning_rate": 0.00018970616626124562,
"loss": 0.3952,
"step": 4000
},
{
"epoch": 0.15483223290474535,
"grad_norm": 1.2745383977890015,
"learning_rate": 0.00018968042524164383,
"loss": 0.3647,
"step": 4010
},
{
"epoch": 0.15521834819877214,
"grad_norm": 7.799386024475098,
"learning_rate": 0.00018965468422204205,
"loss": 0.5554,
"step": 4020
},
{
"epoch": 0.15560446349279894,
"grad_norm": 2.4778294563293457,
"learning_rate": 0.00018962894320244026,
"loss": 0.662,
"step": 4030
},
{
"epoch": 0.15599057878682573,
"grad_norm": 0.8415629267692566,
"learning_rate": 0.00018960320218283847,
"loss": 0.4317,
"step": 4040
},
{
"epoch": 0.15637669408085253,
"grad_norm": 4.507715702056885,
"learning_rate": 0.00018957746116323669,
"loss": 0.4512,
"step": 4050
},
{
"epoch": 0.15676280937487935,
"grad_norm": 3.5790421962738037,
"learning_rate": 0.0001895517201436349,
"loss": 0.4022,
"step": 4060
},
{
"epoch": 0.15714892466890615,
"grad_norm": 3.7266156673431396,
"learning_rate": 0.0001895259791240331,
"loss": 0.3945,
"step": 4070
},
{
"epoch": 0.15753503996293294,
"grad_norm": 7.909580230712891,
"learning_rate": 0.00018950023810443133,
"loss": 0.3726,
"step": 4080
},
{
"epoch": 0.15792115525695974,
"grad_norm": 2.2439534664154053,
"learning_rate": 0.00018947449708482954,
"loss": 0.4157,
"step": 4090
},
{
"epoch": 0.15830727055098653,
"grad_norm": 1.6076972484588623,
"learning_rate": 0.00018944875606522775,
"loss": 0.2363,
"step": 4100
},
{
"epoch": 0.15869338584501333,
"grad_norm": 3.7495157718658447,
"learning_rate": 0.00018942301504562596,
"loss": 0.4908,
"step": 4110
},
{
"epoch": 0.15907950113904012,
"grad_norm": 0.2942291796207428,
"learning_rate": 0.00018939727402602418,
"loss": 0.4915,
"step": 4120
},
{
"epoch": 0.15946561643306692,
"grad_norm": 1.3951829671859741,
"learning_rate": 0.0001893715330064224,
"loss": 0.4585,
"step": 4130
},
{
"epoch": 0.15985173172709372,
"grad_norm": 0.4405671954154968,
"learning_rate": 0.0001893457919868206,
"loss": 0.2839,
"step": 4140
},
{
"epoch": 0.1602378470211205,
"grad_norm": 1.0917588472366333,
"learning_rate": 0.00018932005096721882,
"loss": 0.401,
"step": 4150
},
{
"epoch": 0.1606239623151473,
"grad_norm": 1.6183397769927979,
"learning_rate": 0.00018929430994761703,
"loss": 0.5555,
"step": 4160
},
{
"epoch": 0.1610100776091741,
"grad_norm": 2.0909583568573,
"learning_rate": 0.00018926856892801524,
"loss": 0.52,
"step": 4170
},
{
"epoch": 0.1613961929032009,
"grad_norm": 2.901456356048584,
"learning_rate": 0.00018924282790841346,
"loss": 0.603,
"step": 4180
},
{
"epoch": 0.1617823081972277,
"grad_norm": 7.230431079864502,
"learning_rate": 0.00018921708688881167,
"loss": 0.6189,
"step": 4190
},
{
"epoch": 0.1621684234912545,
"grad_norm": 6.773900508880615,
"learning_rate": 0.00018919134586920988,
"loss": 0.2494,
"step": 4200
},
{
"epoch": 0.16255453878528128,
"grad_norm": 0.8557988405227661,
"learning_rate": 0.0001891656048496081,
"loss": 0.2617,
"step": 4210
},
{
"epoch": 0.16294065407930808,
"grad_norm": 1.3747268915176392,
"learning_rate": 0.00018913986383000634,
"loss": 0.4189,
"step": 4220
},
{
"epoch": 0.16332676937333487,
"grad_norm": 4.072261810302734,
"learning_rate": 0.00018911412281040452,
"loss": 0.5473,
"step": 4230
},
{
"epoch": 0.16371288466736167,
"grad_norm": 2.7210185527801514,
"learning_rate": 0.00018908838179080274,
"loss": 0.3501,
"step": 4240
},
{
"epoch": 0.16409899996138846,
"grad_norm": 2.276454448699951,
"learning_rate": 0.00018906264077120095,
"loss": 0.3078,
"step": 4250
},
{
"epoch": 0.16448511525541526,
"grad_norm": 3.586536169052124,
"learning_rate": 0.00018903689975159916,
"loss": 0.3856,
"step": 4260
},
{
"epoch": 0.16487123054944205,
"grad_norm": 2.199673891067505,
"learning_rate": 0.00018901115873199738,
"loss": 0.3677,
"step": 4270
},
{
"epoch": 0.16525734584346885,
"grad_norm": 2.8410561084747314,
"learning_rate": 0.0001889854177123956,
"loss": 0.6101,
"step": 4280
},
{
"epoch": 0.16564346113749565,
"grad_norm": 3.9638853073120117,
"learning_rate": 0.00018895967669279383,
"loss": 0.5066,
"step": 4290
},
{
"epoch": 0.16602957643152244,
"grad_norm": 1.2070738077163696,
"learning_rate": 0.00018893393567319202,
"loss": 0.385,
"step": 4300
},
{
"epoch": 0.16641569172554924,
"grad_norm": 1.0531187057495117,
"learning_rate": 0.00018890819465359023,
"loss": 0.3608,
"step": 4310
},
{
"epoch": 0.16680180701957603,
"grad_norm": 1.1998246908187866,
"learning_rate": 0.00018888245363398844,
"loss": 0.4624,
"step": 4320
},
{
"epoch": 0.16718792231360285,
"grad_norm": 2.126063346862793,
"learning_rate": 0.00018885671261438666,
"loss": 0.6076,
"step": 4330
},
{
"epoch": 0.16757403760762965,
"grad_norm": 1.5854765176773071,
"learning_rate": 0.0001888309715947849,
"loss": 0.4817,
"step": 4340
},
{
"epoch": 0.16796015290165645,
"grad_norm": 6.630712509155273,
"learning_rate": 0.00018880523057518308,
"loss": 0.4098,
"step": 4350
},
{
"epoch": 0.16834626819568324,
"grad_norm": 2.060789108276367,
"learning_rate": 0.00018877948955558132,
"loss": 0.3523,
"step": 4360
},
{
"epoch": 0.16873238348971004,
"grad_norm": 2.2551252841949463,
"learning_rate": 0.0001887537485359795,
"loss": 0.3095,
"step": 4370
},
{
"epoch": 0.16911849878373683,
"grad_norm": 3.736640453338623,
"learning_rate": 0.00018872800751637772,
"loss": 0.3812,
"step": 4380
},
{
"epoch": 0.16950461407776363,
"grad_norm": 1.9971100091934204,
"learning_rate": 0.00018870226649677594,
"loss": 0.3422,
"step": 4390
},
{
"epoch": 0.16989072937179042,
"grad_norm": 3.6577255725860596,
"learning_rate": 0.00018867652547717415,
"loss": 0.7857,
"step": 4400
},
{
"epoch": 0.17027684466581722,
"grad_norm": 2.166538715362549,
"learning_rate": 0.0001886507844575724,
"loss": 0.5596,
"step": 4410
},
{
"epoch": 0.170662959959844,
"grad_norm": 2.0177736282348633,
"learning_rate": 0.00018862504343797058,
"loss": 0.3197,
"step": 4420
},
{
"epoch": 0.1710490752538708,
"grad_norm": 0.29447808861732483,
"learning_rate": 0.00018859930241836882,
"loss": 0.5284,
"step": 4430
},
{
"epoch": 0.1714351905478976,
"grad_norm": 2.17985200881958,
"learning_rate": 0.000188573561398767,
"loss": 0.5188,
"step": 4440
},
{
"epoch": 0.1718213058419244,
"grad_norm": 2.87449049949646,
"learning_rate": 0.00018854782037916522,
"loss": 0.554,
"step": 4450
},
{
"epoch": 0.1722074211359512,
"grad_norm": 1.8865265846252441,
"learning_rate": 0.00018852207935956343,
"loss": 0.4338,
"step": 4460
},
{
"epoch": 0.172593536429978,
"grad_norm": 2.042337417602539,
"learning_rate": 0.00018849633833996164,
"loss": 0.3924,
"step": 4470
},
{
"epoch": 0.17297965172400478,
"grad_norm": 1.4254354238510132,
"learning_rate": 0.00018847059732035988,
"loss": 0.2607,
"step": 4480
},
{
"epoch": 0.17336576701803158,
"grad_norm": 2.611560344696045,
"learning_rate": 0.00018844485630075807,
"loss": 0.4967,
"step": 4490
},
{
"epoch": 0.17375188231205838,
"grad_norm": 1.1008936166763306,
"learning_rate": 0.0001884191152811563,
"loss": 0.4109,
"step": 4500
},
{
"epoch": 0.17413799760608517,
"grad_norm": 0.8280178308486938,
"learning_rate": 0.0001883933742615545,
"loss": 0.6632,
"step": 4510
},
{
"epoch": 0.17452411290011197,
"grad_norm": 2.226020336151123,
"learning_rate": 0.0001883676332419527,
"loss": 0.4777,
"step": 4520
},
{
"epoch": 0.17491022819413876,
"grad_norm": 1.6062042713165283,
"learning_rate": 0.00018834189222235095,
"loss": 0.4671,
"step": 4530
},
{
"epoch": 0.17529634348816556,
"grad_norm": 3.9853012561798096,
"learning_rate": 0.00018831615120274914,
"loss": 0.4843,
"step": 4540
},
{
"epoch": 0.17568245878219235,
"grad_norm": 0.30268657207489014,
"learning_rate": 0.00018829041018314738,
"loss": 0.3922,
"step": 4550
},
{
"epoch": 0.17606857407621915,
"grad_norm": 6.283960342407227,
"learning_rate": 0.00018826466916354556,
"loss": 0.6106,
"step": 4560
},
{
"epoch": 0.17645468937024594,
"grad_norm": 1.4164658784866333,
"learning_rate": 0.0001882389281439438,
"loss": 0.3014,
"step": 4570
},
{
"epoch": 0.17684080466427274,
"grad_norm": 4.847668170928955,
"learning_rate": 0.000188213187124342,
"loss": 0.5216,
"step": 4580
},
{
"epoch": 0.17722691995829956,
"grad_norm": 3.683180332183838,
"learning_rate": 0.0001881874461047402,
"loss": 0.3268,
"step": 4590
},
{
"epoch": 0.17761303525232636,
"grad_norm": 1.053144097328186,
"learning_rate": 0.00018816170508513844,
"loss": 0.5229,
"step": 4600
},
{
"epoch": 0.17799915054635315,
"grad_norm": 0.29438719153404236,
"learning_rate": 0.00018813596406553663,
"loss": 0.4523,
"step": 4610
},
{
"epoch": 0.17838526584037995,
"grad_norm": 1.5682024955749512,
"learning_rate": 0.00018811022304593487,
"loss": 0.4367,
"step": 4620
},
{
"epoch": 0.17877138113440674,
"grad_norm": 1.462189793586731,
"learning_rate": 0.00018808448202633306,
"loss": 0.5086,
"step": 4630
},
{
"epoch": 0.17915749642843354,
"grad_norm": 0.7927210927009583,
"learning_rate": 0.0001880587410067313,
"loss": 0.4654,
"step": 4640
},
{
"epoch": 0.17954361172246033,
"grad_norm": 1.4543548822402954,
"learning_rate": 0.0001880329999871295,
"loss": 0.5005,
"step": 4650
},
{
"epoch": 0.17992972701648713,
"grad_norm": 1.5814868211746216,
"learning_rate": 0.00018800725896752772,
"loss": 0.4127,
"step": 4660
},
{
"epoch": 0.18031584231051392,
"grad_norm": 1.9244798421859741,
"learning_rate": 0.00018798151794792594,
"loss": 0.3796,
"step": 4670
},
{
"epoch": 0.18070195760454072,
"grad_norm": 1.8725996017456055,
"learning_rate": 0.00018795577692832412,
"loss": 0.4112,
"step": 4680
},
{
"epoch": 0.18108807289856751,
"grad_norm": 2.8138442039489746,
"learning_rate": 0.00018793003590872236,
"loss": 0.6117,
"step": 4690
},
{
"epoch": 0.1814741881925943,
"grad_norm": 3.4465060234069824,
"learning_rate": 0.00018790429488912055,
"loss": 0.4223,
"step": 4700
},
{
"epoch": 0.1818603034866211,
"grad_norm": 4.431785583496094,
"learning_rate": 0.0001878785538695188,
"loss": 0.54,
"step": 4710
},
{
"epoch": 0.1822464187806479,
"grad_norm": 6.951846599578857,
"learning_rate": 0.000187852812849917,
"loss": 0.3702,
"step": 4720
},
{
"epoch": 0.1826325340746747,
"grad_norm": 1.0188024044036865,
"learning_rate": 0.00018782707183031522,
"loss": 0.2715,
"step": 4730
},
{
"epoch": 0.1830186493687015,
"grad_norm": 0.3875834047794342,
"learning_rate": 0.00018780133081071343,
"loss": 0.4208,
"step": 4740
},
{
"epoch": 0.1834047646627283,
"grad_norm": 2.7475740909576416,
"learning_rate": 0.00018777558979111162,
"loss": 0.3613,
"step": 4750
},
{
"epoch": 0.18379087995675508,
"grad_norm": 2.553227186203003,
"learning_rate": 0.00018774984877150986,
"loss": 0.4781,
"step": 4760
},
{
"epoch": 0.18417699525078188,
"grad_norm": 2.005154609680176,
"learning_rate": 0.00018772410775190804,
"loss": 0.3805,
"step": 4770
},
{
"epoch": 0.18456311054480867,
"grad_norm": 0.7380127310752869,
"learning_rate": 0.00018769836673230628,
"loss": 0.3679,
"step": 4780
},
{
"epoch": 0.18494922583883547,
"grad_norm": 3.6547505855560303,
"learning_rate": 0.0001876726257127045,
"loss": 0.4502,
"step": 4790
},
{
"epoch": 0.18533534113286226,
"grad_norm": 2.232980728149414,
"learning_rate": 0.0001876468846931027,
"loss": 0.4628,
"step": 4800
},
{
"epoch": 0.18572145642688906,
"grad_norm": 6.521275043487549,
"learning_rate": 0.00018762114367350092,
"loss": 0.4765,
"step": 4810
},
{
"epoch": 0.18610757172091585,
"grad_norm": 1.6310979127883911,
"learning_rate": 0.0001875954026538991,
"loss": 0.4039,
"step": 4820
},
{
"epoch": 0.18649368701494265,
"grad_norm": 1.1469775438308716,
"learning_rate": 0.00018756966163429735,
"loss": 0.4195,
"step": 4830
},
{
"epoch": 0.18687980230896944,
"grad_norm": 0.7688332200050354,
"learning_rate": 0.00018754392061469556,
"loss": 0.264,
"step": 4840
},
{
"epoch": 0.18726591760299627,
"grad_norm": 3.3422155380249023,
"learning_rate": 0.00018751817959509378,
"loss": 0.5275,
"step": 4850
},
{
"epoch": 0.18765203289702306,
"grad_norm": 1.517876386642456,
"learning_rate": 0.000187492438575492,
"loss": 0.4567,
"step": 4860
},
{
"epoch": 0.18803814819104986,
"grad_norm": 1.2196050882339478,
"learning_rate": 0.0001874666975558902,
"loss": 0.4231,
"step": 4870
},
{
"epoch": 0.18842426348507665,
"grad_norm": 1.3325402736663818,
"learning_rate": 0.00018744095653628842,
"loss": 0.6325,
"step": 4880
},
{
"epoch": 0.18881037877910345,
"grad_norm": 6.098769664764404,
"learning_rate": 0.0001874152155166866,
"loss": 0.576,
"step": 4890
},
{
"epoch": 0.18919649407313024,
"grad_norm": 2.602363348007202,
"learning_rate": 0.00018738947449708484,
"loss": 0.3237,
"step": 4900
},
{
"epoch": 0.18958260936715704,
"grad_norm": 0.970106303691864,
"learning_rate": 0.00018736373347748306,
"loss": 0.409,
"step": 4910
},
{
"epoch": 0.18996872466118384,
"grad_norm": 3.2592012882232666,
"learning_rate": 0.00018733799245788127,
"loss": 0.408,
"step": 4920
},
{
"epoch": 0.19035483995521063,
"grad_norm": 0.31132128834724426,
"learning_rate": 0.00018731225143827948,
"loss": 0.2446,
"step": 4930
},
{
"epoch": 0.19074095524923743,
"grad_norm": 5.321741104125977,
"learning_rate": 0.0001872865104186777,
"loss": 0.4604,
"step": 4940
},
{
"epoch": 0.19112707054326422,
"grad_norm": 1.1165122985839844,
"learning_rate": 0.0001872607693990759,
"loss": 0.3605,
"step": 4950
},
{
"epoch": 0.19151318583729102,
"grad_norm": 0.8274110555648804,
"learning_rate": 0.0001872350283794741,
"loss": 0.2669,
"step": 4960
},
{
"epoch": 0.1918993011313178,
"grad_norm": 2.8668346405029297,
"learning_rate": 0.00018720928735987234,
"loss": 0.4055,
"step": 4970
},
{
"epoch": 0.1922854164253446,
"grad_norm": 3.411841630935669,
"learning_rate": 0.00018718354634027055,
"loss": 0.5989,
"step": 4980
},
{
"epoch": 0.1926715317193714,
"grad_norm": 0.18740829825401306,
"learning_rate": 0.00018715780532066876,
"loss": 0.3805,
"step": 4990
},
{
"epoch": 0.1930576470133982,
"grad_norm": 1.0823473930358887,
"learning_rate": 0.00018713206430106698,
"loss": 0.2854,
"step": 5000
},
{
"epoch": 0.193443762307425,
"grad_norm": 1.9816405773162842,
"learning_rate": 0.0001871063232814652,
"loss": 0.3771,
"step": 5010
},
{
"epoch": 0.1938298776014518,
"grad_norm": 5.267081260681152,
"learning_rate": 0.0001870805822618634,
"loss": 0.3085,
"step": 5020
},
{
"epoch": 0.19421599289547858,
"grad_norm": 5.706038475036621,
"learning_rate": 0.00018705484124226162,
"loss": 0.484,
"step": 5030
},
{
"epoch": 0.19460210818950538,
"grad_norm": 1.3357723951339722,
"learning_rate": 0.00018702910022265983,
"loss": 0.2161,
"step": 5040
},
{
"epoch": 0.19498822348353217,
"grad_norm": 1.0626447200775146,
"learning_rate": 0.00018700335920305804,
"loss": 0.3491,
"step": 5050
},
{
"epoch": 0.19537433877755897,
"grad_norm": 2.441228151321411,
"learning_rate": 0.00018697761818345626,
"loss": 0.3975,
"step": 5060
},
{
"epoch": 0.19576045407158577,
"grad_norm": 2.6739327907562256,
"learning_rate": 0.00018695187716385447,
"loss": 0.3418,
"step": 5070
},
{
"epoch": 0.19614656936561256,
"grad_norm": 2.3216919898986816,
"learning_rate": 0.00018692613614425268,
"loss": 0.6265,
"step": 5080
},
{
"epoch": 0.19653268465963936,
"grad_norm": 3.9119021892547607,
"learning_rate": 0.0001869003951246509,
"loss": 0.2982,
"step": 5090
},
{
"epoch": 0.19691879995366615,
"grad_norm": 5.744061470031738,
"learning_rate": 0.0001868746541050491,
"loss": 0.4048,
"step": 5100
},
{
"epoch": 0.19730491524769297,
"grad_norm": 8.512910842895508,
"learning_rate": 0.00018684891308544732,
"loss": 0.3598,
"step": 5110
},
{
"epoch": 0.19769103054171977,
"grad_norm": 1.6382296085357666,
"learning_rate": 0.00018682317206584554,
"loss": 0.2121,
"step": 5120
},
{
"epoch": 0.19807714583574657,
"grad_norm": 2.1593070030212402,
"learning_rate": 0.00018679743104624375,
"loss": 0.4914,
"step": 5130
},
{
"epoch": 0.19846326112977336,
"grad_norm": 3.067112445831299,
"learning_rate": 0.00018677169002664196,
"loss": 0.4171,
"step": 5140
},
{
"epoch": 0.19884937642380016,
"grad_norm": 1.9954415559768677,
"learning_rate": 0.00018674594900704018,
"loss": 0.5161,
"step": 5150
},
{
"epoch": 0.19923549171782695,
"grad_norm": 2.793346643447876,
"learning_rate": 0.0001867202079874384,
"loss": 0.4159,
"step": 5160
},
{
"epoch": 0.19962160701185375,
"grad_norm": 4.087403774261475,
"learning_rate": 0.0001866944669678366,
"loss": 0.3339,
"step": 5170
},
{
"epoch": 0.20000772230588054,
"grad_norm": 2.10153865814209,
"learning_rate": 0.00018666872594823482,
"loss": 0.4352,
"step": 5180
},
{
"epoch": 0.20039383759990734,
"grad_norm": 2.947117805480957,
"learning_rate": 0.00018664298492863303,
"loss": 0.305,
"step": 5190
},
{
"epoch": 0.20077995289393413,
"grad_norm": 1.2496302127838135,
"learning_rate": 0.00018661724390903124,
"loss": 0.4578,
"step": 5200
},
{
"epoch": 0.20116606818796093,
"grad_norm": 0.5246118903160095,
"learning_rate": 0.00018659150288942946,
"loss": 0.7531,
"step": 5210
},
{
"epoch": 0.20155218348198772,
"grad_norm": 4.099668502807617,
"learning_rate": 0.00018656576186982767,
"loss": 0.3809,
"step": 5220
},
{
"epoch": 0.20193829877601452,
"grad_norm": 4.237419128417969,
"learning_rate": 0.00018654002085022588,
"loss": 0.3169,
"step": 5230
},
{
"epoch": 0.20232441407004131,
"grad_norm": 1.6228466033935547,
"learning_rate": 0.0001865142798306241,
"loss": 0.5832,
"step": 5240
},
{
"epoch": 0.2027105293640681,
"grad_norm": 4.567386627197266,
"learning_rate": 0.0001864885388110223,
"loss": 0.2177,
"step": 5250
},
{
"epoch": 0.2030966446580949,
"grad_norm": 1.4991040229797363,
"learning_rate": 0.00018646279779142052,
"loss": 0.3851,
"step": 5260
},
{
"epoch": 0.2034827599521217,
"grad_norm": 2.127082586288452,
"learning_rate": 0.00018643705677181874,
"loss": 0.4721,
"step": 5270
},
{
"epoch": 0.2038688752461485,
"grad_norm": 2.9149303436279297,
"learning_rate": 0.00018641131575221695,
"loss": 0.2556,
"step": 5280
},
{
"epoch": 0.2042549905401753,
"grad_norm": 0.06375914812088013,
"learning_rate": 0.00018638557473261516,
"loss": 0.3599,
"step": 5290
},
{
"epoch": 0.20464110583420209,
"grad_norm": 3.338331699371338,
"learning_rate": 0.00018635983371301338,
"loss": 0.4062,
"step": 5300
},
{
"epoch": 0.20502722112822888,
"grad_norm": 4.006681442260742,
"learning_rate": 0.0001863340926934116,
"loss": 0.4538,
"step": 5310
},
{
"epoch": 0.20541333642225568,
"grad_norm": 1.1406009197235107,
"learning_rate": 0.0001863083516738098,
"loss": 0.8432,
"step": 5320
},
{
"epoch": 0.20579945171628247,
"grad_norm": 9.281437873840332,
"learning_rate": 0.00018628261065420802,
"loss": 0.4538,
"step": 5330
},
{
"epoch": 0.20618556701030927,
"grad_norm": 3.1884214878082275,
"learning_rate": 0.00018625686963460626,
"loss": 0.3361,
"step": 5340
},
{
"epoch": 0.20657168230433606,
"grad_norm": 1.4311977624893188,
"learning_rate": 0.00018623112861500444,
"loss": 0.5519,
"step": 5350
},
{
"epoch": 0.20695779759836286,
"grad_norm": 3.574361801147461,
"learning_rate": 0.00018620538759540266,
"loss": 0.518,
"step": 5360
},
{
"epoch": 0.20734391289238968,
"grad_norm": 3.0186073780059814,
"learning_rate": 0.00018617964657580087,
"loss": 0.4204,
"step": 5370
},
{
"epoch": 0.20773002818641648,
"grad_norm": 2.832859754562378,
"learning_rate": 0.00018615390555619908,
"loss": 0.5736,
"step": 5380
},
{
"epoch": 0.20811614348044327,
"grad_norm": 2.2258200645446777,
"learning_rate": 0.0001861281645365973,
"loss": 0.8194,
"step": 5390
},
{
"epoch": 0.20850225877447007,
"grad_norm": 1.0975148677825928,
"learning_rate": 0.0001861024235169955,
"loss": 0.5235,
"step": 5400
},
{
"epoch": 0.20888837406849686,
"grad_norm": 2.597329616546631,
"learning_rate": 0.00018607668249739375,
"loss": 0.2798,
"step": 5410
},
{
"epoch": 0.20927448936252366,
"grad_norm": 1.3780876398086548,
"learning_rate": 0.00018605094147779194,
"loss": 0.4046,
"step": 5420
},
{
"epoch": 0.20966060465655045,
"grad_norm": 2.409886598587036,
"learning_rate": 0.00018602520045819018,
"loss": 0.3243,
"step": 5430
},
{
"epoch": 0.21004671995057725,
"grad_norm": 1.0368077754974365,
"learning_rate": 0.00018599945943858836,
"loss": 0.4469,
"step": 5440
},
{
"epoch": 0.21043283524460404,
"grad_norm": 2.961658000946045,
"learning_rate": 0.00018597371841898658,
"loss": 0.5104,
"step": 5450
},
{
"epoch": 0.21081895053863084,
"grad_norm": 1.1599836349487305,
"learning_rate": 0.00018594797739938482,
"loss": 0.3422,
"step": 5460
},
{
"epoch": 0.21120506583265763,
"grad_norm": 3.293682336807251,
"learning_rate": 0.000185922236379783,
"loss": 0.3556,
"step": 5470
},
{
"epoch": 0.21159118112668443,
"grad_norm": 1.6923863887786865,
"learning_rate": 0.00018589649536018124,
"loss": 0.3084,
"step": 5480
},
{
"epoch": 0.21197729642071123,
"grad_norm": 3.7289531230926514,
"learning_rate": 0.00018587075434057943,
"loss": 0.4668,
"step": 5490
},
{
"epoch": 0.21236341171473802,
"grad_norm": 1.3744993209838867,
"learning_rate": 0.00018584501332097767,
"loss": 0.2984,
"step": 5500
},
{
"epoch": 0.21274952700876482,
"grad_norm": 1.4377775192260742,
"learning_rate": 0.00018581927230137586,
"loss": 0.2622,
"step": 5510
},
{
"epoch": 0.2131356423027916,
"grad_norm": 4.957859992980957,
"learning_rate": 0.00018579353128177407,
"loss": 0.5561,
"step": 5520
},
{
"epoch": 0.2135217575968184,
"grad_norm": 3.2645647525787354,
"learning_rate": 0.0001857677902621723,
"loss": 0.59,
"step": 5530
},
{
"epoch": 0.2139078728908452,
"grad_norm": 1.1365091800689697,
"learning_rate": 0.0001857420492425705,
"loss": 0.443,
"step": 5540
},
{
"epoch": 0.214293988184872,
"grad_norm": 3.187476396560669,
"learning_rate": 0.00018571630822296874,
"loss": 0.2612,
"step": 5550
},
{
"epoch": 0.2146801034788988,
"grad_norm": 2.6851940155029297,
"learning_rate": 0.00018569056720336692,
"loss": 0.4543,
"step": 5560
},
{
"epoch": 0.2150662187729256,
"grad_norm": 2.2613587379455566,
"learning_rate": 0.00018566482618376516,
"loss": 0.3185,
"step": 5570
},
{
"epoch": 0.21545233406695238,
"grad_norm": 1.292475700378418,
"learning_rate": 0.00018563908516416335,
"loss": 0.2794,
"step": 5580
},
{
"epoch": 0.21583844936097918,
"grad_norm": 2.0878446102142334,
"learning_rate": 0.00018561334414456156,
"loss": 0.3908,
"step": 5590
},
{
"epoch": 0.21622456465500597,
"grad_norm": 8.058819770812988,
"learning_rate": 0.0001855876031249598,
"loss": 0.6282,
"step": 5600
},
{
"epoch": 0.21661067994903277,
"grad_norm": 1.8231629133224487,
"learning_rate": 0.000185561862105358,
"loss": 0.4973,
"step": 5610
},
{
"epoch": 0.21699679524305956,
"grad_norm": 3.947242259979248,
"learning_rate": 0.00018553612108575623,
"loss": 0.4598,
"step": 5620
},
{
"epoch": 0.21738291053708636,
"grad_norm": 3.3258073329925537,
"learning_rate": 0.00018551038006615442,
"loss": 0.5266,
"step": 5630
},
{
"epoch": 0.21776902583111318,
"grad_norm": 2.301485300064087,
"learning_rate": 0.00018548463904655266,
"loss": 0.4339,
"step": 5640
},
{
"epoch": 0.21815514112513998,
"grad_norm": 4.4706878662109375,
"learning_rate": 0.00018545889802695087,
"loss": 0.5233,
"step": 5650
},
{
"epoch": 0.21854125641916677,
"grad_norm": 1.1203399896621704,
"learning_rate": 0.00018543315700734906,
"loss": 0.4547,
"step": 5660
},
{
"epoch": 0.21892737171319357,
"grad_norm": 0.3744584918022156,
"learning_rate": 0.0001854074159877473,
"loss": 0.2524,
"step": 5670
},
{
"epoch": 0.21931348700722036,
"grad_norm": 2.7888870239257812,
"learning_rate": 0.00018538167496814548,
"loss": 0.411,
"step": 5680
},
{
"epoch": 0.21969960230124716,
"grad_norm": 4.9972429275512695,
"learning_rate": 0.00018535593394854372,
"loss": 0.6359,
"step": 5690
},
{
"epoch": 0.22008571759527396,
"grad_norm": 1.1321420669555664,
"learning_rate": 0.0001853301929289419,
"loss": 0.4068,
"step": 5700
},
{
"epoch": 0.22047183288930075,
"grad_norm": 1.9291785955429077,
"learning_rate": 0.00018530445190934015,
"loss": 0.5428,
"step": 5710
},
{
"epoch": 0.22085794818332755,
"grad_norm": 0.8663263916969299,
"learning_rate": 0.00018527871088973836,
"loss": 0.4662,
"step": 5720
},
{
"epoch": 0.22124406347735434,
"grad_norm": 3.039782762527466,
"learning_rate": 0.00018525296987013655,
"loss": 0.3045,
"step": 5730
},
{
"epoch": 0.22163017877138114,
"grad_norm": 1.3552179336547852,
"learning_rate": 0.0001852272288505348,
"loss": 0.3411,
"step": 5740
},
{
"epoch": 0.22201629406540793,
"grad_norm": 1.4136948585510254,
"learning_rate": 0.00018520148783093298,
"loss": 0.5517,
"step": 5750
},
{
"epoch": 0.22240240935943473,
"grad_norm": 2.463942766189575,
"learning_rate": 0.00018517574681133122,
"loss": 0.4681,
"step": 5760
},
{
"epoch": 0.22278852465346152,
"grad_norm": 0.9063917994499207,
"learning_rate": 0.0001851500057917294,
"loss": 0.4537,
"step": 5770
},
{
"epoch": 0.22317463994748832,
"grad_norm": 2.352678060531616,
"learning_rate": 0.00018512426477212764,
"loss": 0.4245,
"step": 5780
},
{
"epoch": 0.2235607552415151,
"grad_norm": 2.0424869060516357,
"learning_rate": 0.00018509852375252586,
"loss": 0.2892,
"step": 5790
},
{
"epoch": 0.2239468705355419,
"grad_norm": 2.7604904174804688,
"learning_rate": 0.00018507278273292404,
"loss": 0.3606,
"step": 5800
},
{
"epoch": 0.2243329858295687,
"grad_norm": 2.827798366546631,
"learning_rate": 0.00018504704171332228,
"loss": 0.3212,
"step": 5810
},
{
"epoch": 0.2247191011235955,
"grad_norm": 3.1988680362701416,
"learning_rate": 0.00018502130069372047,
"loss": 0.5649,
"step": 5820
},
{
"epoch": 0.2251052164176223,
"grad_norm": 1.8216092586517334,
"learning_rate": 0.0001849955596741187,
"loss": 0.2871,
"step": 5830
},
{
"epoch": 0.2254913317116491,
"grad_norm": 2.7595627307891846,
"learning_rate": 0.00018496981865451692,
"loss": 0.665,
"step": 5840
},
{
"epoch": 0.22587744700567589,
"grad_norm": 1.2395098209381104,
"learning_rate": 0.00018494407763491514,
"loss": 0.2504,
"step": 5850
},
{
"epoch": 0.22626356229970268,
"grad_norm": 0.6991098523139954,
"learning_rate": 0.00018491833661531335,
"loss": 0.2263,
"step": 5860
},
{
"epoch": 0.22664967759372948,
"grad_norm": 11.053647994995117,
"learning_rate": 0.00018489259559571156,
"loss": 0.5919,
"step": 5870
},
{
"epoch": 0.22703579288775627,
"grad_norm": 2.8663880825042725,
"learning_rate": 0.00018486685457610978,
"loss": 0.3399,
"step": 5880
},
{
"epoch": 0.22742190818178307,
"grad_norm": 1.4995262622833252,
"learning_rate": 0.00018484111355650796,
"loss": 0.4474,
"step": 5890
},
{
"epoch": 0.2278080234758099,
"grad_norm": 3.275681972503662,
"learning_rate": 0.0001848153725369062,
"loss": 0.4347,
"step": 5900
},
{
"epoch": 0.22819413876983669,
"grad_norm": 14.772253036499023,
"learning_rate": 0.00018478963151730442,
"loss": 0.3705,
"step": 5910
},
{
"epoch": 0.22858025406386348,
"grad_norm": 3.184976816177368,
"learning_rate": 0.00018476389049770263,
"loss": 0.3866,
"step": 5920
},
{
"epoch": 0.22896636935789028,
"grad_norm": 2.310765504837036,
"learning_rate": 0.00018473814947810084,
"loss": 0.2717,
"step": 5930
},
{
"epoch": 0.22935248465191707,
"grad_norm": 2.061189889907837,
"learning_rate": 0.00018471240845849906,
"loss": 0.2054,
"step": 5940
},
{
"epoch": 0.22973859994594387,
"grad_norm": 10.815469741821289,
"learning_rate": 0.00018468666743889727,
"loss": 0.5868,
"step": 5950
},
{
"epoch": 0.23012471523997066,
"grad_norm": 1.7080497741699219,
"learning_rate": 0.00018466092641929548,
"loss": 0.236,
"step": 5960
},
{
"epoch": 0.23051083053399746,
"grad_norm": 7.389080047607422,
"learning_rate": 0.0001846351853996937,
"loss": 0.2752,
"step": 5970
},
{
"epoch": 0.23089694582802425,
"grad_norm": 2.9860422611236572,
"learning_rate": 0.0001846094443800919,
"loss": 0.3436,
"step": 5980
},
{
"epoch": 0.23128306112205105,
"grad_norm": 13.12328815460205,
"learning_rate": 0.00018458370336049012,
"loss": 0.3952,
"step": 5990
},
{
"epoch": 0.23166917641607784,
"grad_norm": 3.7130823135375977,
"learning_rate": 0.00018455796234088834,
"loss": 0.3658,
"step": 6000
},
{
"epoch": 0.23205529171010464,
"grad_norm": 1.8329843282699585,
"learning_rate": 0.00018453222132128655,
"loss": 0.4172,
"step": 6010
},
{
"epoch": 0.23244140700413143,
"grad_norm": 1.3583799600601196,
"learning_rate": 0.00018450648030168476,
"loss": 0.4005,
"step": 6020
},
{
"epoch": 0.23282752229815823,
"grad_norm": 3.1711816787719727,
"learning_rate": 0.00018448073928208297,
"loss": 0.3674,
"step": 6030
},
{
"epoch": 0.23321363759218502,
"grad_norm": 1.576937198638916,
"learning_rate": 0.0001844549982624812,
"loss": 0.3444,
"step": 6040
},
{
"epoch": 0.23359975288621182,
"grad_norm": 3.922267436981201,
"learning_rate": 0.0001844292572428794,
"loss": 0.5939,
"step": 6050
},
{
"epoch": 0.23398586818023862,
"grad_norm": 2.9851067066192627,
"learning_rate": 0.00018440351622327761,
"loss": 0.2387,
"step": 6060
},
{
"epoch": 0.2343719834742654,
"grad_norm": 2.1216888427734375,
"learning_rate": 0.00018437777520367583,
"loss": 0.3836,
"step": 6070
},
{
"epoch": 0.2347580987682922,
"grad_norm": 2.9788095951080322,
"learning_rate": 0.00018435203418407404,
"loss": 0.474,
"step": 6080
},
{
"epoch": 0.235144214062319,
"grad_norm": 1.0204919576644897,
"learning_rate": 0.00018432629316447225,
"loss": 0.2837,
"step": 6090
},
{
"epoch": 0.2355303293563458,
"grad_norm": 0.9091696739196777,
"learning_rate": 0.00018430055214487047,
"loss": 0.6203,
"step": 6100
},
{
"epoch": 0.2359164446503726,
"grad_norm": 0.25899162888526917,
"learning_rate": 0.00018427481112526868,
"loss": 0.4759,
"step": 6110
},
{
"epoch": 0.2363025599443994,
"grad_norm": 1.8625538349151611,
"learning_rate": 0.0001842490701056669,
"loss": 0.2992,
"step": 6120
},
{
"epoch": 0.23668867523842618,
"grad_norm": 1.586521863937378,
"learning_rate": 0.0001842233290860651,
"loss": 0.6122,
"step": 6130
},
{
"epoch": 0.23707479053245298,
"grad_norm": 2.387650966644287,
"learning_rate": 0.00018419758806646332,
"loss": 0.3276,
"step": 6140
},
{
"epoch": 0.23746090582647977,
"grad_norm": 4.840515613555908,
"learning_rate": 0.00018417184704686153,
"loss": 0.6295,
"step": 6150
},
{
"epoch": 0.2378470211205066,
"grad_norm": 1.70024836063385,
"learning_rate": 0.00018414610602725975,
"loss": 0.2047,
"step": 6160
},
{
"epoch": 0.2382331364145334,
"grad_norm": 2.791619062423706,
"learning_rate": 0.00018412036500765796,
"loss": 0.4364,
"step": 6170
},
{
"epoch": 0.2386192517085602,
"grad_norm": 3.710066318511963,
"learning_rate": 0.00018409462398805617,
"loss": 0.4564,
"step": 6180
},
{
"epoch": 0.23900536700258698,
"grad_norm": 2.564347982406616,
"learning_rate": 0.0001840688829684544,
"loss": 0.3156,
"step": 6190
},
{
"epoch": 0.23939148229661378,
"grad_norm": 2.3921267986297607,
"learning_rate": 0.0001840431419488526,
"loss": 0.3483,
"step": 6200
},
{
"epoch": 0.23977759759064057,
"grad_norm": 1.4785810708999634,
"learning_rate": 0.00018401740092925081,
"loss": 0.4338,
"step": 6210
},
{
"epoch": 0.24016371288466737,
"grad_norm": 3.624790906906128,
"learning_rate": 0.00018399165990964903,
"loss": 0.7156,
"step": 6220
},
{
"epoch": 0.24054982817869416,
"grad_norm": 3.942161798477173,
"learning_rate": 0.00018396591889004724,
"loss": 0.3932,
"step": 6230
},
{
"epoch": 0.24093594347272096,
"grad_norm": 3.2236740589141846,
"learning_rate": 0.00018394017787044545,
"loss": 0.3933,
"step": 6240
},
{
"epoch": 0.24132205876674775,
"grad_norm": 2.5040500164031982,
"learning_rate": 0.00018391443685084367,
"loss": 0.5711,
"step": 6250
},
{
"epoch": 0.24170817406077455,
"grad_norm": 1.9934203624725342,
"learning_rate": 0.00018388869583124188,
"loss": 0.3074,
"step": 6260
},
{
"epoch": 0.24209428935480135,
"grad_norm": 3.702509641647339,
"learning_rate": 0.0001838629548116401,
"loss": 0.3454,
"step": 6270
},
{
"epoch": 0.24248040464882814,
"grad_norm": 2.076802968978882,
"learning_rate": 0.0001838372137920383,
"loss": 0.3044,
"step": 6280
},
{
"epoch": 0.24286651994285494,
"grad_norm": 5.798679351806641,
"learning_rate": 0.00018381147277243652,
"loss": 0.3396,
"step": 6290
},
{
"epoch": 0.24325263523688173,
"grad_norm": 4.698869705200195,
"learning_rate": 0.00018378573175283473,
"loss": 0.3735,
"step": 6300
},
{
"epoch": 0.24363875053090853,
"grad_norm": 3.029979705810547,
"learning_rate": 0.00018375999073323295,
"loss": 0.3891,
"step": 6310
},
{
"epoch": 0.24402486582493532,
"grad_norm": 2.5507185459136963,
"learning_rate": 0.00018373424971363116,
"loss": 0.4854,
"step": 6320
},
{
"epoch": 0.24441098111896212,
"grad_norm": 3.2052571773529053,
"learning_rate": 0.00018370850869402937,
"loss": 0.6789,
"step": 6330
},
{
"epoch": 0.2447970964129889,
"grad_norm": 1.9265435934066772,
"learning_rate": 0.00018368276767442761,
"loss": 0.4505,
"step": 6340
},
{
"epoch": 0.2451832117070157,
"grad_norm": 0.8391959071159363,
"learning_rate": 0.0001836570266548258,
"loss": 0.3432,
"step": 6350
},
{
"epoch": 0.2455693270010425,
"grad_norm": 3.4653851985931396,
"learning_rate": 0.00018363128563522401,
"loss": 0.3571,
"step": 6360
},
{
"epoch": 0.2459554422950693,
"grad_norm": 2.3033368587493896,
"learning_rate": 0.00018360554461562223,
"loss": 0.3625,
"step": 6370
},
{
"epoch": 0.2463415575890961,
"grad_norm": 1.659408450126648,
"learning_rate": 0.00018357980359602044,
"loss": 0.5311,
"step": 6380
},
{
"epoch": 0.2467276728831229,
"grad_norm": 1.1839714050292969,
"learning_rate": 0.00018355406257641865,
"loss": 0.3905,
"step": 6390
},
{
"epoch": 0.24711378817714968,
"grad_norm": 0.49230822920799255,
"learning_rate": 0.00018352832155681687,
"loss": 0.4021,
"step": 6400
},
{
"epoch": 0.24749990347117648,
"grad_norm": 4.451594829559326,
"learning_rate": 0.0001835025805372151,
"loss": 0.4504,
"step": 6410
},
{
"epoch": 0.2478860187652033,
"grad_norm": 1.0058324337005615,
"learning_rate": 0.0001834768395176133,
"loss": 0.2636,
"step": 6420
},
{
"epoch": 0.2482721340592301,
"grad_norm": 2.7853894233703613,
"learning_rate": 0.0001834510984980115,
"loss": 0.47,
"step": 6430
},
{
"epoch": 0.2486582493532569,
"grad_norm": 2.730095148086548,
"learning_rate": 0.00018342535747840972,
"loss": 0.3941,
"step": 6440
},
{
"epoch": 0.2490443646472837,
"grad_norm": 2.4993178844451904,
"learning_rate": 0.00018339961645880793,
"loss": 0.5777,
"step": 6450
},
{
"epoch": 0.24943047994131048,
"grad_norm": 2.361525297164917,
"learning_rate": 0.00018337387543920617,
"loss": 0.3798,
"step": 6460
},
{
"epoch": 0.24981659523533728,
"grad_norm": 2.5558526515960693,
"learning_rate": 0.00018334813441960436,
"loss": 0.3113,
"step": 6470
},
{
"epoch": 0.25020271052936405,
"grad_norm": 0.8033503890037537,
"learning_rate": 0.0001833223934000026,
"loss": 0.5254,
"step": 6480
},
{
"epoch": 0.25058882582339087,
"grad_norm": 2.721090078353882,
"learning_rate": 0.0001832966523804008,
"loss": 0.393,
"step": 6490
},
{
"epoch": 0.25097494111741764,
"grad_norm": 1.7147916555404663,
"learning_rate": 0.000183270911360799,
"loss": 0.3225,
"step": 6500
},
{
"epoch": 0.25136105641144446,
"grad_norm": 2.388347864151001,
"learning_rate": 0.00018324517034119721,
"loss": 0.3519,
"step": 6510
},
{
"epoch": 0.25174717170547123,
"grad_norm": 2.470891237258911,
"learning_rate": 0.00018321942932159543,
"loss": 0.4384,
"step": 6520
},
{
"epoch": 0.25213328699949805,
"grad_norm": 1.4743351936340332,
"learning_rate": 0.00018319368830199367,
"loss": 0.2464,
"step": 6530
},
{
"epoch": 0.2525194022935248,
"grad_norm": 1.5889122486114502,
"learning_rate": 0.00018316794728239185,
"loss": 0.3149,
"step": 6540
},
{
"epoch": 0.25290551758755164,
"grad_norm": 4.900819778442383,
"learning_rate": 0.0001831422062627901,
"loss": 0.3978,
"step": 6550
},
{
"epoch": 0.25329163288157847,
"grad_norm": 5.22566556930542,
"learning_rate": 0.00018311646524318828,
"loss": 0.4473,
"step": 6560
},
{
"epoch": 0.25367774817560523,
"grad_norm": 4.7480363845825195,
"learning_rate": 0.0001830907242235865,
"loss": 0.3976,
"step": 6570
},
{
"epoch": 0.25406386346963206,
"grad_norm": 1.4711374044418335,
"learning_rate": 0.0001830649832039847,
"loss": 0.5183,
"step": 6580
},
{
"epoch": 0.2544499787636588,
"grad_norm": 2.237309217453003,
"learning_rate": 0.00018303924218438292,
"loss": 0.2171,
"step": 6590
},
{
"epoch": 0.25483609405768565,
"grad_norm": 4.107303619384766,
"learning_rate": 0.00018301350116478116,
"loss": 0.3918,
"step": 6600
},
{
"epoch": 0.2552222093517124,
"grad_norm": 4.7285003662109375,
"learning_rate": 0.00018298776014517935,
"loss": 0.2042,
"step": 6610
},
{
"epoch": 0.25560832464573924,
"grad_norm": 2.1333792209625244,
"learning_rate": 0.0001829620191255776,
"loss": 0.3502,
"step": 6620
},
{
"epoch": 0.255994439939766,
"grad_norm": 3.062173843383789,
"learning_rate": 0.00018293627810597577,
"loss": 0.3949,
"step": 6630
},
{
"epoch": 0.25638055523379283,
"grad_norm": 1.538854956626892,
"learning_rate": 0.00018291053708637401,
"loss": 0.4613,
"step": 6640
},
{
"epoch": 0.2567666705278196,
"grad_norm": 2.546586751937866,
"learning_rate": 0.00018288479606677223,
"loss": 0.5868,
"step": 6650
},
{
"epoch": 0.2571527858218464,
"grad_norm": 2.7282049655914307,
"learning_rate": 0.00018285905504717041,
"loss": 0.4186,
"step": 6660
},
{
"epoch": 0.2575389011158732,
"grad_norm": 3.204634189605713,
"learning_rate": 0.00018283331402756865,
"loss": 0.4072,
"step": 6670
},
{
"epoch": 0.2579250164099,
"grad_norm": 2.421846866607666,
"learning_rate": 0.00018280757300796684,
"loss": 0.306,
"step": 6680
},
{
"epoch": 0.2583111317039268,
"grad_norm": 4.243416786193848,
"learning_rate": 0.00018278183198836508,
"loss": 0.2631,
"step": 6690
},
{
"epoch": 0.2586972469979536,
"grad_norm": 1.0495362281799316,
"learning_rate": 0.00018275609096876327,
"loss": 0.3488,
"step": 6700
},
{
"epoch": 0.25908336229198037,
"grad_norm": 1.915279746055603,
"learning_rate": 0.0001827303499491615,
"loss": 0.2589,
"step": 6710
},
{
"epoch": 0.2594694775860072,
"grad_norm": 3.724299192428589,
"learning_rate": 0.00018270460892955972,
"loss": 0.5118,
"step": 6720
},
{
"epoch": 0.25985559288003396,
"grad_norm": 2.832204580307007,
"learning_rate": 0.0001826788679099579,
"loss": 0.2508,
"step": 6730
},
{
"epoch": 0.2602417081740608,
"grad_norm": 1.1942508220672607,
"learning_rate": 0.00018265312689035615,
"loss": 0.4328,
"step": 6740
},
{
"epoch": 0.26062782346808755,
"grad_norm": 1.0741711854934692,
"learning_rate": 0.00018262738587075433,
"loss": 0.3514,
"step": 6750
},
{
"epoch": 0.2610139387621144,
"grad_norm": 2.9918277263641357,
"learning_rate": 0.00018260164485115257,
"loss": 0.3528,
"step": 6760
},
{
"epoch": 0.26140005405614114,
"grad_norm": 1.3773655891418457,
"learning_rate": 0.0001825759038315508,
"loss": 0.365,
"step": 6770
},
{
"epoch": 0.26178616935016796,
"grad_norm": 3.5288615226745605,
"learning_rate": 0.000182550162811949,
"loss": 0.3645,
"step": 6780
},
{
"epoch": 0.26217228464419473,
"grad_norm": 1.2178785800933838,
"learning_rate": 0.00018252442179234721,
"loss": 0.3742,
"step": 6790
},
{
"epoch": 0.26255839993822155,
"grad_norm": 2.7981081008911133,
"learning_rate": 0.0001824986807727454,
"loss": 0.6174,
"step": 6800
},
{
"epoch": 0.2629445152322484,
"grad_norm": 1.6766215562820435,
"learning_rate": 0.00018247293975314364,
"loss": 0.3028,
"step": 6810
},
{
"epoch": 0.26333063052627514,
"grad_norm": 3.7797629833221436,
"learning_rate": 0.00018244719873354183,
"loss": 0.2633,
"step": 6820
},
{
"epoch": 0.26371674582030197,
"grad_norm": 7.794743537902832,
"learning_rate": 0.00018242145771394007,
"loss": 0.3586,
"step": 6830
},
{
"epoch": 0.26410286111432874,
"grad_norm": 0.5704814195632935,
"learning_rate": 0.00018239571669433828,
"loss": 0.3506,
"step": 6840
},
{
"epoch": 0.26448897640835556,
"grad_norm": 5.771059513092041,
"learning_rate": 0.0001823699756747365,
"loss": 0.3881,
"step": 6850
},
{
"epoch": 0.2648750917023823,
"grad_norm": 2.723592519760132,
"learning_rate": 0.0001823442346551347,
"loss": 0.3955,
"step": 6860
},
{
"epoch": 0.26526120699640915,
"grad_norm": 1.5448215007781982,
"learning_rate": 0.0001823184936355329,
"loss": 0.495,
"step": 6870
},
{
"epoch": 0.2656473222904359,
"grad_norm": 2.2980363368988037,
"learning_rate": 0.00018229275261593113,
"loss": 0.2695,
"step": 6880
},
{
"epoch": 0.26603343758446274,
"grad_norm": 1.959811806678772,
"learning_rate": 0.00018226701159632932,
"loss": 0.383,
"step": 6890
},
{
"epoch": 0.2664195528784895,
"grad_norm": 2.1491482257843018,
"learning_rate": 0.00018224127057672756,
"loss": 0.5655,
"step": 6900
},
{
"epoch": 0.26680566817251633,
"grad_norm": 6.472841262817383,
"learning_rate": 0.00018221552955712577,
"loss": 0.4757,
"step": 6910
},
{
"epoch": 0.2671917834665431,
"grad_norm": 7.878561496734619,
"learning_rate": 0.000182189788537524,
"loss": 0.3944,
"step": 6920
},
{
"epoch": 0.2675778987605699,
"grad_norm": 0.052701435983181,
"learning_rate": 0.0001821640475179222,
"loss": 0.382,
"step": 6930
},
{
"epoch": 0.2679640140545967,
"grad_norm": 2.294677972793579,
"learning_rate": 0.00018213830649832039,
"loss": 0.2932,
"step": 6940
},
{
"epoch": 0.2683501293486235,
"grad_norm": 1.6058757305145264,
"learning_rate": 0.00018211256547871863,
"loss": 0.4438,
"step": 6950
},
{
"epoch": 0.2687362446426503,
"grad_norm": 4.003495693206787,
"learning_rate": 0.00018208682445911684,
"loss": 0.5945,
"step": 6960
},
{
"epoch": 0.2691223599366771,
"grad_norm": 1.423017144203186,
"learning_rate": 0.00018206108343951505,
"loss": 0.4356,
"step": 6970
},
{
"epoch": 0.26950847523070387,
"grad_norm": 2.206341028213501,
"learning_rate": 0.00018203534241991327,
"loss": 0.344,
"step": 6980
},
{
"epoch": 0.2698945905247307,
"grad_norm": 0.6644784212112427,
"learning_rate": 0.00018200960140031148,
"loss": 0.4988,
"step": 6990
},
{
"epoch": 0.27028070581875746,
"grad_norm": 2.4569833278656006,
"learning_rate": 0.0001819838603807097,
"loss": 0.3689,
"step": 7000
},
{
"epoch": 0.2706668211127843,
"grad_norm": 1.554567575454712,
"learning_rate": 0.00018195811936110788,
"loss": 0.4684,
"step": 7010
},
{
"epoch": 0.27105293640681105,
"grad_norm": 3.2556328773498535,
"learning_rate": 0.00018193237834150612,
"loss": 0.611,
"step": 7020
},
{
"epoch": 0.2714390517008379,
"grad_norm": 2.9123427867889404,
"learning_rate": 0.00018190663732190433,
"loss": 0.4278,
"step": 7030
},
{
"epoch": 0.27182516699486464,
"grad_norm": 2.159273862838745,
"learning_rate": 0.00018188089630230255,
"loss": 0.2384,
"step": 7040
},
{
"epoch": 0.27221128228889147,
"grad_norm": 3.4977822303771973,
"learning_rate": 0.00018185515528270076,
"loss": 0.5459,
"step": 7050
},
{
"epoch": 0.27259739758291823,
"grad_norm": 1.1822031736373901,
"learning_rate": 0.00018182941426309897,
"loss": 0.4364,
"step": 7060
},
{
"epoch": 0.27298351287694506,
"grad_norm": 2.4467339515686035,
"learning_rate": 0.00018180367324349719,
"loss": 0.5198,
"step": 7070
},
{
"epoch": 0.2733696281709719,
"grad_norm": 1.0406467914581299,
"learning_rate": 0.0001817779322238954,
"loss": 0.2797,
"step": 7080
},
{
"epoch": 0.27375574346499865,
"grad_norm": 1.925830602645874,
"learning_rate": 0.0001817521912042936,
"loss": 0.4898,
"step": 7090
},
{
"epoch": 0.27414185875902547,
"grad_norm": 3.0385682582855225,
"learning_rate": 0.00018172645018469183,
"loss": 0.3867,
"step": 7100
},
{
"epoch": 0.27452797405305224,
"grad_norm": 1.5285695791244507,
"learning_rate": 0.00018170070916509004,
"loss": 0.4233,
"step": 7110
},
{
"epoch": 0.27491408934707906,
"grad_norm": 1.266693115234375,
"learning_rate": 0.00018167496814548825,
"loss": 0.4724,
"step": 7120
},
{
"epoch": 0.27530020464110583,
"grad_norm": 3.371323585510254,
"learning_rate": 0.00018164922712588647,
"loss": 0.533,
"step": 7130
},
{
"epoch": 0.27568631993513265,
"grad_norm": 2.662691116333008,
"learning_rate": 0.00018162348610628468,
"loss": 0.3134,
"step": 7140
},
{
"epoch": 0.2760724352291594,
"grad_norm": 1.8977057933807373,
"learning_rate": 0.0001815977450866829,
"loss": 0.3038,
"step": 7150
},
{
"epoch": 0.27645855052318624,
"grad_norm": 3.1027894020080566,
"learning_rate": 0.0001815720040670811,
"loss": 0.5074,
"step": 7160
},
{
"epoch": 0.276844665817213,
"grad_norm": 1.2112785577774048,
"learning_rate": 0.00018154626304747932,
"loss": 0.324,
"step": 7170
},
{
"epoch": 0.27723078111123983,
"grad_norm": 1.6500996351242065,
"learning_rate": 0.00018152052202787753,
"loss": 0.2856,
"step": 7180
},
{
"epoch": 0.2776168964052666,
"grad_norm": 3.215747833251953,
"learning_rate": 0.00018149478100827575,
"loss": 0.4522,
"step": 7190
},
{
"epoch": 0.2780030116992934,
"grad_norm": 4.8541059494018555,
"learning_rate": 0.00018146903998867396,
"loss": 0.4106,
"step": 7200
},
{
"epoch": 0.2783891269933202,
"grad_norm": 2.3697152137756348,
"learning_rate": 0.00018144329896907217,
"loss": 0.2673,
"step": 7210
},
{
"epoch": 0.278775242287347,
"grad_norm": 2.9693639278411865,
"learning_rate": 0.00018141755794947039,
"loss": 0.3949,
"step": 7220
},
{
"epoch": 0.2791613575813738,
"grad_norm": 2.691817283630371,
"learning_rate": 0.0001813918169298686,
"loss": 0.3427,
"step": 7230
},
{
"epoch": 0.2795474728754006,
"grad_norm": 5.197331428527832,
"learning_rate": 0.0001813660759102668,
"loss": 0.4331,
"step": 7240
},
{
"epoch": 0.27993358816942737,
"grad_norm": 1.5799933671951294,
"learning_rate": 0.00018134033489066503,
"loss": 0.3543,
"step": 7250
},
{
"epoch": 0.2803197034634542,
"grad_norm": 1.3614271879196167,
"learning_rate": 0.00018131459387106324,
"loss": 0.5289,
"step": 7260
},
{
"epoch": 0.28070581875748096,
"grad_norm": 2.2942802906036377,
"learning_rate": 0.00018128885285146145,
"loss": 0.4318,
"step": 7270
},
{
"epoch": 0.2810919340515078,
"grad_norm": 1.1805604696273804,
"learning_rate": 0.00018126311183185967,
"loss": 0.4754,
"step": 7280
},
{
"epoch": 0.28147804934553455,
"grad_norm": 0.5108867883682251,
"learning_rate": 0.00018123737081225788,
"loss": 0.4517,
"step": 7290
},
{
"epoch": 0.2818641646395614,
"grad_norm": 1.1736596822738647,
"learning_rate": 0.0001812116297926561,
"loss": 0.4538,
"step": 7300
},
{
"epoch": 0.28225027993358814,
"grad_norm": 5.497414588928223,
"learning_rate": 0.0001811858887730543,
"loss": 0.5116,
"step": 7310
},
{
"epoch": 0.28263639522761497,
"grad_norm": 1.1347368955612183,
"learning_rate": 0.00018116014775345252,
"loss": 0.3848,
"step": 7320
},
{
"epoch": 0.28302251052164173,
"grad_norm": 2.740715742111206,
"learning_rate": 0.00018113440673385073,
"loss": 0.3456,
"step": 7330
},
{
"epoch": 0.28340862581566856,
"grad_norm": 1.3853389024734497,
"learning_rate": 0.00018110866571424897,
"loss": 0.3398,
"step": 7340
},
{
"epoch": 0.2837947411096954,
"grad_norm": 7.493706703186035,
"learning_rate": 0.00018108292469464716,
"loss": 0.2726,
"step": 7350
},
{
"epoch": 0.28418085640372215,
"grad_norm": 1.81704843044281,
"learning_rate": 0.00018105718367504537,
"loss": 0.3818,
"step": 7360
},
{
"epoch": 0.28456697169774897,
"grad_norm": 2.4877755641937256,
"learning_rate": 0.00018103144265544359,
"loss": 0.3499,
"step": 7370
},
{
"epoch": 0.28495308699177574,
"grad_norm": 1.3704471588134766,
"learning_rate": 0.0001810057016358418,
"loss": 0.2346,
"step": 7380
},
{
"epoch": 0.28533920228580256,
"grad_norm": 2.664745569229126,
"learning_rate": 0.00018097996061624,
"loss": 0.4041,
"step": 7390
},
{
"epoch": 0.28572531757982933,
"grad_norm": 3.6539089679718018,
"learning_rate": 0.00018095421959663823,
"loss": 0.2885,
"step": 7400
},
{
"epoch": 0.28611143287385615,
"grad_norm": 0.8653857707977295,
"learning_rate": 0.00018092847857703647,
"loss": 0.3849,
"step": 7410
},
{
"epoch": 0.2864975481678829,
"grad_norm": 2.6319446563720703,
"learning_rate": 0.00018090273755743465,
"loss": 0.2728,
"step": 7420
},
{
"epoch": 0.28688366346190974,
"grad_norm": 2.3457818031311035,
"learning_rate": 0.00018087699653783287,
"loss": 0.446,
"step": 7430
},
{
"epoch": 0.2872697787559365,
"grad_norm": 0.8546158671379089,
"learning_rate": 0.00018085125551823108,
"loss": 0.2898,
"step": 7440
},
{
"epoch": 0.28765589404996333,
"grad_norm": 0.45937278866767883,
"learning_rate": 0.0001808255144986293,
"loss": 0.583,
"step": 7450
},
{
"epoch": 0.2880420093439901,
"grad_norm": 1.7129520177841187,
"learning_rate": 0.00018079977347902753,
"loss": 0.4908,
"step": 7460
},
{
"epoch": 0.2884281246380169,
"grad_norm": 4.106715679168701,
"learning_rate": 0.00018077403245942572,
"loss": 0.3373,
"step": 7470
},
{
"epoch": 0.2888142399320437,
"grad_norm": 3.8112800121307373,
"learning_rate": 0.00018074829143982396,
"loss": 0.392,
"step": 7480
},
{
"epoch": 0.2892003552260705,
"grad_norm": 0.5382593274116516,
"learning_rate": 0.00018072255042022215,
"loss": 0.2929,
"step": 7490
},
{
"epoch": 0.2895864705200973,
"grad_norm": 2.50888991355896,
"learning_rate": 0.00018069680940062036,
"loss": 0.3361,
"step": 7500
},
{
"epoch": 0.2899725858141241,
"grad_norm": 3.3544275760650635,
"learning_rate": 0.00018067106838101857,
"loss": 0.388,
"step": 7510
},
{
"epoch": 0.2903587011081509,
"grad_norm": 1.192386507987976,
"learning_rate": 0.00018064532736141679,
"loss": 0.4427,
"step": 7520
},
{
"epoch": 0.2907448164021777,
"grad_norm": 1.5527079105377197,
"learning_rate": 0.00018061958634181503,
"loss": 0.4023,
"step": 7530
},
{
"epoch": 0.29113093169620446,
"grad_norm": 0.67446368932724,
"learning_rate": 0.0001805938453222132,
"loss": 0.4949,
"step": 7540
},
{
"epoch": 0.2915170469902313,
"grad_norm": 1.6349838972091675,
"learning_rate": 0.00018056810430261145,
"loss": 0.3811,
"step": 7550
},
{
"epoch": 0.29190316228425806,
"grad_norm": 1.4848904609680176,
"learning_rate": 0.00018054236328300964,
"loss": 0.3851,
"step": 7560
},
{
"epoch": 0.2922892775782849,
"grad_norm": 0.9933151006698608,
"learning_rate": 0.00018051662226340785,
"loss": 0.4699,
"step": 7570
},
{
"epoch": 0.29267539287231165,
"grad_norm": 1.1026233434677124,
"learning_rate": 0.00018049088124380607,
"loss": 0.3287,
"step": 7580
},
{
"epoch": 0.29306150816633847,
"grad_norm": 1.232954740524292,
"learning_rate": 0.00018046514022420428,
"loss": 0.3722,
"step": 7590
},
{
"epoch": 0.2934476234603653,
"grad_norm": 3.8303146362304688,
"learning_rate": 0.00018043939920460252,
"loss": 0.2985,
"step": 7600
},
{
"epoch": 0.29383373875439206,
"grad_norm": 1.9358845949172974,
"learning_rate": 0.0001804136581850007,
"loss": 0.4361,
"step": 7610
},
{
"epoch": 0.2942198540484189,
"grad_norm": 1.8905962705612183,
"learning_rate": 0.00018038791716539895,
"loss": 0.2835,
"step": 7620
},
{
"epoch": 0.29460596934244565,
"grad_norm": 1.9965651035308838,
"learning_rate": 0.00018036217614579713,
"loss": 0.5387,
"step": 7630
},
{
"epoch": 0.2949920846364725,
"grad_norm": 4.204270839691162,
"learning_rate": 0.00018033643512619535,
"loss": 0.3498,
"step": 7640
},
{
"epoch": 0.29537819993049924,
"grad_norm": 1.4732340574264526,
"learning_rate": 0.00018031069410659359,
"loss": 0.315,
"step": 7650
},
{
"epoch": 0.29576431522452606,
"grad_norm": 1.0233594179153442,
"learning_rate": 0.00018028495308699177,
"loss": 0.1536,
"step": 7660
},
{
"epoch": 0.29615043051855283,
"grad_norm": 3.1531457901000977,
"learning_rate": 0.00018025921206739,
"loss": 0.3793,
"step": 7670
},
{
"epoch": 0.29653654581257965,
"grad_norm": 0.8080945014953613,
"learning_rate": 0.0001802334710477882,
"loss": 0.5589,
"step": 7680
},
{
"epoch": 0.2969226611066064,
"grad_norm": 3.1202728748321533,
"learning_rate": 0.00018020773002818644,
"loss": 0.4652,
"step": 7690
},
{
"epoch": 0.29730877640063325,
"grad_norm": 2.5934784412384033,
"learning_rate": 0.00018018198900858463,
"loss": 0.4921,
"step": 7700
},
{
"epoch": 0.29769489169466,
"grad_norm": 2.858642101287842,
"learning_rate": 0.00018015624798898284,
"loss": 0.2732,
"step": 7710
},
{
"epoch": 0.29808100698868684,
"grad_norm": 3.621229887008667,
"learning_rate": 0.00018013050696938108,
"loss": 0.5639,
"step": 7720
},
{
"epoch": 0.2984671222827136,
"grad_norm": 3.7943220138549805,
"learning_rate": 0.00018010476594977926,
"loss": 0.3177,
"step": 7730
},
{
"epoch": 0.2988532375767404,
"grad_norm": 1.6371623277664185,
"learning_rate": 0.0001800790249301775,
"loss": 0.4211,
"step": 7740
},
{
"epoch": 0.2992393528707672,
"grad_norm": 1.9557713270187378,
"learning_rate": 0.0001800532839105757,
"loss": 0.4351,
"step": 7750
},
{
"epoch": 0.299625468164794,
"grad_norm": 2.684964895248413,
"learning_rate": 0.00018002754289097393,
"loss": 0.39,
"step": 7760
},
{
"epoch": 0.3000115834588208,
"grad_norm": 1.7401316165924072,
"learning_rate": 0.00018000180187137215,
"loss": 0.2844,
"step": 7770
},
{
"epoch": 0.3003976987528476,
"grad_norm": 0.6305844187736511,
"learning_rate": 0.00017997606085177033,
"loss": 0.2472,
"step": 7780
},
{
"epoch": 0.3007838140468744,
"grad_norm": 2.2880289554595947,
"learning_rate": 0.00017995031983216857,
"loss": 0.3952,
"step": 7790
},
{
"epoch": 0.3011699293409012,
"grad_norm": 3.423980951309204,
"learning_rate": 0.00017992457881256676,
"loss": 0.4459,
"step": 7800
},
{
"epoch": 0.30155604463492797,
"grad_norm": 0.6920475363731384,
"learning_rate": 0.000179898837792965,
"loss": 0.2909,
"step": 7810
},
{
"epoch": 0.3019421599289548,
"grad_norm": 0.8905349373817444,
"learning_rate": 0.00017987309677336318,
"loss": 0.346,
"step": 7820
},
{
"epoch": 0.30232827522298156,
"grad_norm": 1.8836702108383179,
"learning_rate": 0.00017984735575376143,
"loss": 0.4038,
"step": 7830
},
{
"epoch": 0.3027143905170084,
"grad_norm": 2.6712753772735596,
"learning_rate": 0.00017982161473415964,
"loss": 0.3452,
"step": 7840
},
{
"epoch": 0.30310050581103515,
"grad_norm": 2.344122886657715,
"learning_rate": 0.00017979587371455785,
"loss": 0.5091,
"step": 7850
},
{
"epoch": 0.30348662110506197,
"grad_norm": 3.734415054321289,
"learning_rate": 0.00017977013269495607,
"loss": 0.3893,
"step": 7860
},
{
"epoch": 0.3038727363990888,
"grad_norm": 1.70572829246521,
"learning_rate": 0.00017974439167535425,
"loss": 0.4829,
"step": 7870
},
{
"epoch": 0.30425885169311556,
"grad_norm": 1.779189109802246,
"learning_rate": 0.0001797186506557525,
"loss": 0.5361,
"step": 7880
},
{
"epoch": 0.3046449669871424,
"grad_norm": 2.888803482055664,
"learning_rate": 0.00017969290963615068,
"loss": 0.4305,
"step": 7890
},
{
"epoch": 0.30503108228116915,
"grad_norm": 1.2247655391693115,
"learning_rate": 0.00017966716861654892,
"loss": 0.3817,
"step": 7900
},
{
"epoch": 0.305417197575196,
"grad_norm": 2.995152473449707,
"learning_rate": 0.00017964142759694713,
"loss": 0.4669,
"step": 7910
},
{
"epoch": 0.30580331286922274,
"grad_norm": 8.049060821533203,
"learning_rate": 0.00017961568657734535,
"loss": 0.6706,
"step": 7920
},
{
"epoch": 0.30618942816324957,
"grad_norm": 2.1181435585021973,
"learning_rate": 0.00017958994555774356,
"loss": 0.4353,
"step": 7930
},
{
"epoch": 0.30657554345727633,
"grad_norm": 8.394509315490723,
"learning_rate": 0.00017956420453814174,
"loss": 0.3497,
"step": 7940
},
{
"epoch": 0.30696165875130316,
"grad_norm": 2.5140750408172607,
"learning_rate": 0.00017953846351853998,
"loss": 0.5774,
"step": 7950
},
{
"epoch": 0.3073477740453299,
"grad_norm": 2.720942974090576,
"learning_rate": 0.0001795127224989382,
"loss": 0.4457,
"step": 7960
},
{
"epoch": 0.30773388933935675,
"grad_norm": 1.8155667781829834,
"learning_rate": 0.0001794869814793364,
"loss": 0.4155,
"step": 7970
},
{
"epoch": 0.3081200046333835,
"grad_norm": 1.9989752769470215,
"learning_rate": 0.00017946124045973462,
"loss": 0.3233,
"step": 7980
},
{
"epoch": 0.30850611992741034,
"grad_norm": 0.7483557462692261,
"learning_rate": 0.00017943549944013284,
"loss": 0.2932,
"step": 7990
},
{
"epoch": 0.3088922352214371,
"grad_norm": 0.5750642418861389,
"learning_rate": 0.00017940975842053105,
"loss": 0.401,
"step": 8000
},
{
"epoch": 0.30927835051546393,
"grad_norm": 1.2084500789642334,
"learning_rate": 0.00017938401740092924,
"loss": 0.3705,
"step": 8010
},
{
"epoch": 0.3096644658094907,
"grad_norm": 1.833434820175171,
"learning_rate": 0.00017935827638132748,
"loss": 0.3507,
"step": 8020
},
{
"epoch": 0.3100505811035175,
"grad_norm": 3.147508382797241,
"learning_rate": 0.0001793325353617257,
"loss": 0.3255,
"step": 8030
},
{
"epoch": 0.3104366963975443,
"grad_norm": 2.150932788848877,
"learning_rate": 0.0001793067943421239,
"loss": 0.3401,
"step": 8040
},
{
"epoch": 0.3108228116915711,
"grad_norm": 3.3340635299682617,
"learning_rate": 0.00017928105332252212,
"loss": 0.3606,
"step": 8050
},
{
"epoch": 0.3112089269855979,
"grad_norm": 5.173205375671387,
"learning_rate": 0.00017925531230292033,
"loss": 0.1695,
"step": 8060
},
{
"epoch": 0.3115950422796247,
"grad_norm": 1.0863877534866333,
"learning_rate": 0.00017922957128331854,
"loss": 0.3038,
"step": 8070
},
{
"epoch": 0.31198115757365147,
"grad_norm": 1.5977118015289307,
"learning_rate": 0.00017920383026371676,
"loss": 0.2291,
"step": 8080
},
{
"epoch": 0.3123672728676783,
"grad_norm": 4.040243625640869,
"learning_rate": 0.00017917808924411497,
"loss": 0.8538,
"step": 8090
},
{
"epoch": 0.31275338816170506,
"grad_norm": 1.5926854610443115,
"learning_rate": 0.00017915234822451318,
"loss": 0.4733,
"step": 8100
},
{
"epoch": 0.3131395034557319,
"grad_norm": 1.0959421396255493,
"learning_rate": 0.0001791266072049114,
"loss": 0.6076,
"step": 8110
},
{
"epoch": 0.3135256187497587,
"grad_norm": 2.786085367202759,
"learning_rate": 0.0001791008661853096,
"loss": 0.3229,
"step": 8120
},
{
"epoch": 0.3139117340437855,
"grad_norm": 2.2573914527893066,
"learning_rate": 0.00017907512516570782,
"loss": 0.3676,
"step": 8130
},
{
"epoch": 0.3142978493378123,
"grad_norm": 2.271852493286133,
"learning_rate": 0.00017904938414610604,
"loss": 0.6275,
"step": 8140
},
{
"epoch": 0.31468396463183906,
"grad_norm": 1.9762821197509766,
"learning_rate": 0.00017902364312650425,
"loss": 0.2232,
"step": 8150
},
{
"epoch": 0.3150700799258659,
"grad_norm": 2.9960873126983643,
"learning_rate": 0.00017899790210690246,
"loss": 0.4739,
"step": 8160
},
{
"epoch": 0.31545619521989265,
"grad_norm": 1.142216682434082,
"learning_rate": 0.00017897216108730068,
"loss": 0.5983,
"step": 8170
},
{
"epoch": 0.3158423105139195,
"grad_norm": 1.7127768993377686,
"learning_rate": 0.0001789464200676989,
"loss": 0.4131,
"step": 8180
},
{
"epoch": 0.31622842580794625,
"grad_norm": 1.579793095588684,
"learning_rate": 0.0001789206790480971,
"loss": 0.3119,
"step": 8190
},
{
"epoch": 0.31661454110197307,
"grad_norm": 0.9647886157035828,
"learning_rate": 0.00017889493802849532,
"loss": 0.398,
"step": 8200
},
{
"epoch": 0.31700065639599984,
"grad_norm": 3.435312032699585,
"learning_rate": 0.00017886919700889353,
"loss": 0.405,
"step": 8210
},
{
"epoch": 0.31738677169002666,
"grad_norm": 2.1500205993652344,
"learning_rate": 0.00017884345598929174,
"loss": 0.2519,
"step": 8220
},
{
"epoch": 0.3177728869840534,
"grad_norm": 1.3107216358184814,
"learning_rate": 0.00017881771496968996,
"loss": 0.2846,
"step": 8230
},
{
"epoch": 0.31815900227808025,
"grad_norm": 0.1899029165506363,
"learning_rate": 0.00017879197395008817,
"loss": 0.4597,
"step": 8240
},
{
"epoch": 0.318545117572107,
"grad_norm": 2.329299211502075,
"learning_rate": 0.00017876623293048638,
"loss": 0.6523,
"step": 8250
},
{
"epoch": 0.31893123286613384,
"grad_norm": 0.41523978114128113,
"learning_rate": 0.0001787404919108846,
"loss": 0.3266,
"step": 8260
},
{
"epoch": 0.3193173481601606,
"grad_norm": 0.7914639711380005,
"learning_rate": 0.0001787147508912828,
"loss": 0.4029,
"step": 8270
},
{
"epoch": 0.31970346345418743,
"grad_norm": 0.6159287691116333,
"learning_rate": 0.00017868900987168102,
"loss": 0.4426,
"step": 8280
},
{
"epoch": 0.3200895787482142,
"grad_norm": 1.3690640926361084,
"learning_rate": 0.00017866326885207924,
"loss": 0.2974,
"step": 8290
},
{
"epoch": 0.320475694042241,
"grad_norm": 0.8592869639396667,
"learning_rate": 0.00017863752783247745,
"loss": 0.232,
"step": 8300
},
{
"epoch": 0.3208618093362678,
"grad_norm": 0.43169018626213074,
"learning_rate": 0.00017861178681287566,
"loss": 0.4033,
"step": 8310
},
{
"epoch": 0.3212479246302946,
"grad_norm": 0.8405828475952148,
"learning_rate": 0.00017858604579327388,
"loss": 0.3339,
"step": 8320
},
{
"epoch": 0.3216340399243214,
"grad_norm": 2.3412604331970215,
"learning_rate": 0.0001785603047736721,
"loss": 0.2781,
"step": 8330
},
{
"epoch": 0.3220201552183482,
"grad_norm": 2.412045478820801,
"learning_rate": 0.0001785345637540703,
"loss": 0.4346,
"step": 8340
},
{
"epoch": 0.32240627051237497,
"grad_norm": 3.626305341720581,
"learning_rate": 0.00017850882273446852,
"loss": 0.327,
"step": 8350
},
{
"epoch": 0.3227923858064018,
"grad_norm": 0.5645825266838074,
"learning_rate": 0.00017848308171486673,
"loss": 0.234,
"step": 8360
},
{
"epoch": 0.32317850110042856,
"grad_norm": 4.27307653427124,
"learning_rate": 0.00017845734069526494,
"loss": 0.5493,
"step": 8370
},
{
"epoch": 0.3235646163944554,
"grad_norm": 0.4511154890060425,
"learning_rate": 0.00017843159967566316,
"loss": 0.3501,
"step": 8380
},
{
"epoch": 0.3239507316884822,
"grad_norm": 0.314996600151062,
"learning_rate": 0.00017840585865606137,
"loss": 0.3544,
"step": 8390
},
{
"epoch": 0.324336846982509,
"grad_norm": 1.6546530723571777,
"learning_rate": 0.00017838011763645958,
"loss": 0.2455,
"step": 8400
},
{
"epoch": 0.3247229622765358,
"grad_norm": 3.2812252044677734,
"learning_rate": 0.0001783543766168578,
"loss": 0.3333,
"step": 8410
},
{
"epoch": 0.32510907757056257,
"grad_norm": 3.5717616081237793,
"learning_rate": 0.000178328635597256,
"loss": 0.4679,
"step": 8420
},
{
"epoch": 0.3254951928645894,
"grad_norm": 1.12017023563385,
"learning_rate": 0.00017830289457765422,
"loss": 0.3481,
"step": 8430
},
{
"epoch": 0.32588130815861616,
"grad_norm": 1.869462490081787,
"learning_rate": 0.00017827715355805244,
"loss": 0.4566,
"step": 8440
},
{
"epoch": 0.326267423452643,
"grad_norm": 1.4613149166107178,
"learning_rate": 0.00017825141253845065,
"loss": 0.5456,
"step": 8450
},
{
"epoch": 0.32665353874666975,
"grad_norm": 0.6842670440673828,
"learning_rate": 0.0001782256715188489,
"loss": 0.2776,
"step": 8460
},
{
"epoch": 0.32703965404069657,
"grad_norm": 2.9485504627227783,
"learning_rate": 0.00017819993049924708,
"loss": 0.3204,
"step": 8470
},
{
"epoch": 0.32742576933472334,
"grad_norm": 3.2084853649139404,
"learning_rate": 0.0001781741894796453,
"loss": 0.3391,
"step": 8480
},
{
"epoch": 0.32781188462875016,
"grad_norm": 1.9732774496078491,
"learning_rate": 0.0001781484484600435,
"loss": 0.3283,
"step": 8490
},
{
"epoch": 0.32819799992277693,
"grad_norm": 0.6378610134124756,
"learning_rate": 0.00017812270744044172,
"loss": 0.4519,
"step": 8500
},
{
"epoch": 0.32858411521680375,
"grad_norm": 4.108947277069092,
"learning_rate": 0.00017809696642083993,
"loss": 0.4933,
"step": 8510
},
{
"epoch": 0.3289702305108305,
"grad_norm": 2.7623212337493896,
"learning_rate": 0.00017807122540123814,
"loss": 0.4197,
"step": 8520
},
{
"epoch": 0.32935634580485734,
"grad_norm": 1.8904645442962646,
"learning_rate": 0.00017804548438163638,
"loss": 0.325,
"step": 8530
},
{
"epoch": 0.3297424610988841,
"grad_norm": 0.5131659507751465,
"learning_rate": 0.00017801974336203457,
"loss": 0.2403,
"step": 8540
},
{
"epoch": 0.33012857639291093,
"grad_norm": 2.965916633605957,
"learning_rate": 0.0001779940023424328,
"loss": 0.4597,
"step": 8550
},
{
"epoch": 0.3305146916869377,
"grad_norm": 1.5409698486328125,
"learning_rate": 0.000177968261322831,
"loss": 0.4594,
"step": 8560
},
{
"epoch": 0.3309008069809645,
"grad_norm": 1.1746805906295776,
"learning_rate": 0.0001779425203032292,
"loss": 0.4581,
"step": 8570
},
{
"epoch": 0.3312869222749913,
"grad_norm": 4.493356227874756,
"learning_rate": 0.00017791677928362745,
"loss": 0.4699,
"step": 8580
},
{
"epoch": 0.3316730375690181,
"grad_norm": 3.506526470184326,
"learning_rate": 0.00017789103826402564,
"loss": 0.3974,
"step": 8590
},
{
"epoch": 0.3320591528630449,
"grad_norm": 2.3893234729766846,
"learning_rate": 0.00017786529724442388,
"loss": 0.2823,
"step": 8600
},
{
"epoch": 0.3324452681570717,
"grad_norm": 1.6228163242340088,
"learning_rate": 0.00017783955622482206,
"loss": 0.4199,
"step": 8610
},
{
"epoch": 0.3328313834510985,
"grad_norm": 3.2869131565093994,
"learning_rate": 0.0001778138152052203,
"loss": 0.3173,
"step": 8620
},
{
"epoch": 0.3332174987451253,
"grad_norm": 5.547116279602051,
"learning_rate": 0.0001777880741856185,
"loss": 0.4584,
"step": 8630
},
{
"epoch": 0.33360361403915206,
"grad_norm": 1.3338594436645508,
"learning_rate": 0.0001777623331660167,
"loss": 0.4235,
"step": 8640
},
{
"epoch": 0.3339897293331789,
"grad_norm": 1.9165093898773193,
"learning_rate": 0.00017773659214641494,
"loss": 0.2989,
"step": 8650
},
{
"epoch": 0.3343758446272057,
"grad_norm": 1.968935251235962,
"learning_rate": 0.00017771085112681313,
"loss": 0.4194,
"step": 8660
},
{
"epoch": 0.3347619599212325,
"grad_norm": 9.66997241973877,
"learning_rate": 0.00017768511010721137,
"loss": 0.5818,
"step": 8670
},
{
"epoch": 0.3351480752152593,
"grad_norm": 2.3636281490325928,
"learning_rate": 0.00017765936908760956,
"loss": 0.3317,
"step": 8680
},
{
"epoch": 0.33553419050928607,
"grad_norm": 3.3569977283477783,
"learning_rate": 0.0001776336280680078,
"loss": 0.4388,
"step": 8690
},
{
"epoch": 0.3359203058033129,
"grad_norm": 1.2452306747436523,
"learning_rate": 0.00017760788704840598,
"loss": 0.1368,
"step": 8700
},
{
"epoch": 0.33630642109733966,
"grad_norm": 0.0380173958837986,
"learning_rate": 0.0001775821460288042,
"loss": 0.3264,
"step": 8710
},
{
"epoch": 0.3366925363913665,
"grad_norm": 1.5271002054214478,
"learning_rate": 0.00017755640500920244,
"loss": 0.2943,
"step": 8720
},
{
"epoch": 0.33707865168539325,
"grad_norm": 0.9701687693595886,
"learning_rate": 0.00017753066398960062,
"loss": 0.353,
"step": 8730
},
{
"epoch": 0.33746476697942007,
"grad_norm": 1.9296154975891113,
"learning_rate": 0.00017750492296999886,
"loss": 0.3776,
"step": 8740
},
{
"epoch": 0.33785088227344684,
"grad_norm": 1.2136276960372925,
"learning_rate": 0.00017747918195039705,
"loss": 0.5126,
"step": 8750
},
{
"epoch": 0.33823699756747366,
"grad_norm": 1.7323212623596191,
"learning_rate": 0.0001774534409307953,
"loss": 0.3477,
"step": 8760
},
{
"epoch": 0.33862311286150043,
"grad_norm": 1.164534091949463,
"learning_rate": 0.0001774276999111935,
"loss": 0.4053,
"step": 8770
},
{
"epoch": 0.33900922815552725,
"grad_norm": 0.42989471554756165,
"learning_rate": 0.0001774019588915917,
"loss": 0.3026,
"step": 8780
},
{
"epoch": 0.339395343449554,
"grad_norm": 2.357590436935425,
"learning_rate": 0.00017737621787198993,
"loss": 0.3869,
"step": 8790
},
{
"epoch": 0.33978145874358084,
"grad_norm": 1.9374550580978394,
"learning_rate": 0.00017735047685238812,
"loss": 0.2975,
"step": 8800
},
{
"epoch": 0.3401675740376076,
"grad_norm": 4.8107428550720215,
"learning_rate": 0.00017732473583278636,
"loss": 0.3959,
"step": 8810
},
{
"epoch": 0.34055368933163443,
"grad_norm": 1.938700556755066,
"learning_rate": 0.00017729899481318454,
"loss": 0.3726,
"step": 8820
},
{
"epoch": 0.3409398046256612,
"grad_norm": 3.147167682647705,
"learning_rate": 0.00017727325379358278,
"loss": 0.1828,
"step": 8830
},
{
"epoch": 0.341325919919688,
"grad_norm": 1.8921313285827637,
"learning_rate": 0.000177247512773981,
"loss": 0.2038,
"step": 8840
},
{
"epoch": 0.3417120352137148,
"grad_norm": 0.9098349213600159,
"learning_rate": 0.00017722177175437918,
"loss": 0.2853,
"step": 8850
},
{
"epoch": 0.3420981505077416,
"grad_norm": 2.4006853103637695,
"learning_rate": 0.00017719603073477742,
"loss": 0.6054,
"step": 8860
},
{
"epoch": 0.3424842658017684,
"grad_norm": 1.9303867816925049,
"learning_rate": 0.0001771702897151756,
"loss": 0.3507,
"step": 8870
},
{
"epoch": 0.3428703810957952,
"grad_norm": 0.49361029267311096,
"learning_rate": 0.00017714454869557385,
"loss": 0.4661,
"step": 8880
},
{
"epoch": 0.343256496389822,
"grad_norm": 2.542618751525879,
"learning_rate": 0.00017711880767597204,
"loss": 0.6924,
"step": 8890
},
{
"epoch": 0.3436426116838488,
"grad_norm": 0.5868918895721436,
"learning_rate": 0.00017709306665637028,
"loss": 0.4507,
"step": 8900
},
{
"epoch": 0.3440287269778756,
"grad_norm": 2.4685137271881104,
"learning_rate": 0.0001770673256367685,
"loss": 0.4538,
"step": 8910
},
{
"epoch": 0.3444148422719024,
"grad_norm": 2.6662702560424805,
"learning_rate": 0.00017704158461716668,
"loss": 0.6181,
"step": 8920
},
{
"epoch": 0.3448009575659292,
"grad_norm": 1.705103874206543,
"learning_rate": 0.00017701584359756492,
"loss": 0.481,
"step": 8930
},
{
"epoch": 0.345187072859956,
"grad_norm": 2.0710952281951904,
"learning_rate": 0.0001769901025779631,
"loss": 0.4357,
"step": 8940
},
{
"epoch": 0.3455731881539828,
"grad_norm": 3.487117290496826,
"learning_rate": 0.00017696436155836134,
"loss": 0.4572,
"step": 8950
},
{
"epoch": 0.34595930344800957,
"grad_norm": 3.03472900390625,
"learning_rate": 0.00017693862053875956,
"loss": 0.4437,
"step": 8960
},
{
"epoch": 0.3463454187420364,
"grad_norm": 1.310692548751831,
"learning_rate": 0.00017691287951915777,
"loss": 0.4218,
"step": 8970
},
{
"epoch": 0.34673153403606316,
"grad_norm": 4.131219387054443,
"learning_rate": 0.00017688713849955598,
"loss": 0.445,
"step": 8980
},
{
"epoch": 0.34711764933009,
"grad_norm": 1.4199285507202148,
"learning_rate": 0.00017686139747995417,
"loss": 0.3093,
"step": 8990
},
{
"epoch": 0.34750376462411675,
"grad_norm": 1.9338914155960083,
"learning_rate": 0.0001768356564603524,
"loss": 0.275,
"step": 9000
},
{
"epoch": 0.3478898799181436,
"grad_norm": 3.425877332687378,
"learning_rate": 0.0001768099154407506,
"loss": 0.4621,
"step": 9010
},
{
"epoch": 0.34827599521217034,
"grad_norm": 3.6350486278533936,
"learning_rate": 0.00017678417442114884,
"loss": 0.3313,
"step": 9020
},
{
"epoch": 0.34866211050619716,
"grad_norm": 3.339202880859375,
"learning_rate": 0.00017675843340154705,
"loss": 0.5958,
"step": 9030
},
{
"epoch": 0.34904822580022393,
"grad_norm": 4.148682117462158,
"learning_rate": 0.00017673269238194526,
"loss": 0.2384,
"step": 9040
},
{
"epoch": 0.34943434109425076,
"grad_norm": 0.9697182178497314,
"learning_rate": 0.00017670695136234348,
"loss": 0.3119,
"step": 9050
},
{
"epoch": 0.3498204563882775,
"grad_norm": 0.53201824426651,
"learning_rate": 0.0001766812103427417,
"loss": 0.4339,
"step": 9060
},
{
"epoch": 0.35020657168230435,
"grad_norm": 0.9727185368537903,
"learning_rate": 0.0001766554693231399,
"loss": 0.3289,
"step": 9070
},
{
"epoch": 0.3505926869763311,
"grad_norm": 4.32904052734375,
"learning_rate": 0.00017662972830353812,
"loss": 0.4673,
"step": 9080
},
{
"epoch": 0.35097880227035794,
"grad_norm": 2.511558771133423,
"learning_rate": 0.00017660398728393633,
"loss": 0.2257,
"step": 9090
},
{
"epoch": 0.3513649175643847,
"grad_norm": 1.8378714323043823,
"learning_rate": 0.00017657824626433454,
"loss": 0.3977,
"step": 9100
},
{
"epoch": 0.3517510328584115,
"grad_norm": 1.3297137022018433,
"learning_rate": 0.00017655250524473276,
"loss": 0.3541,
"step": 9110
},
{
"epoch": 0.3521371481524383,
"grad_norm": 3.253089666366577,
"learning_rate": 0.00017652676422513097,
"loss": 0.6326,
"step": 9120
},
{
"epoch": 0.3525232634464651,
"grad_norm": 0.9691923260688782,
"learning_rate": 0.00017650102320552918,
"loss": 0.2206,
"step": 9130
},
{
"epoch": 0.3529093787404919,
"grad_norm": 1.570204496383667,
"learning_rate": 0.0001764752821859274,
"loss": 0.2769,
"step": 9140
},
{
"epoch": 0.3532954940345187,
"grad_norm": 1.9307161569595337,
"learning_rate": 0.0001764495411663256,
"loss": 0.3149,
"step": 9150
},
{
"epoch": 0.3536816093285455,
"grad_norm": 2.783297300338745,
"learning_rate": 0.00017642380014672382,
"loss": 0.3912,
"step": 9160
},
{
"epoch": 0.3540677246225723,
"grad_norm": 2.193371057510376,
"learning_rate": 0.00017639805912712204,
"loss": 0.3782,
"step": 9170
},
{
"epoch": 0.3544538399165991,
"grad_norm": 2.3460335731506348,
"learning_rate": 0.00017637231810752025,
"loss": 0.5051,
"step": 9180
},
{
"epoch": 0.3548399552106259,
"grad_norm": 2.4668326377868652,
"learning_rate": 0.00017634657708791846,
"loss": 0.2899,
"step": 9190
},
{
"epoch": 0.3552260705046527,
"grad_norm": 2.004683017730713,
"learning_rate": 0.00017632083606831668,
"loss": 0.3137,
"step": 9200
},
{
"epoch": 0.3556121857986795,
"grad_norm": 6.333971977233887,
"learning_rate": 0.0001762950950487149,
"loss": 0.5027,
"step": 9210
},
{
"epoch": 0.3559983010927063,
"grad_norm": 1.7840352058410645,
"learning_rate": 0.0001762693540291131,
"loss": 0.3988,
"step": 9220
},
{
"epoch": 0.35638441638673307,
"grad_norm": 0.9257024526596069,
"learning_rate": 0.00017624361300951132,
"loss": 0.3662,
"step": 9230
},
{
"epoch": 0.3567705316807599,
"grad_norm": 2.582887887954712,
"learning_rate": 0.00017621787198990953,
"loss": 0.2863,
"step": 9240
},
{
"epoch": 0.35715664697478666,
"grad_norm": 3.119943380355835,
"learning_rate": 0.00017619213097030774,
"loss": 0.4041,
"step": 9250
},
{
"epoch": 0.3575427622688135,
"grad_norm": 2.2561371326446533,
"learning_rate": 0.00017616638995070596,
"loss": 0.3969,
"step": 9260
},
{
"epoch": 0.35792887756284025,
"grad_norm": 2.104891538619995,
"learning_rate": 0.00017614064893110417,
"loss": 0.3216,
"step": 9270
},
{
"epoch": 0.3583149928568671,
"grad_norm": 1.6922805309295654,
"learning_rate": 0.00017611490791150238,
"loss": 0.3828,
"step": 9280
},
{
"epoch": 0.35870110815089384,
"grad_norm": 1.0928469896316528,
"learning_rate": 0.0001760891668919006,
"loss": 0.3225,
"step": 9290
},
{
"epoch": 0.35908722344492067,
"grad_norm": 2.4089863300323486,
"learning_rate": 0.0001760634258722988,
"loss": 0.4143,
"step": 9300
},
{
"epoch": 0.35947333873894743,
"grad_norm": 0.5562119483947754,
"learning_rate": 0.00017603768485269702,
"loss": 0.4597,
"step": 9310
},
{
"epoch": 0.35985945403297426,
"grad_norm": 1.3904486894607544,
"learning_rate": 0.00017601194383309524,
"loss": 0.4462,
"step": 9320
},
{
"epoch": 0.360245569327001,
"grad_norm": 2.1393306255340576,
"learning_rate": 0.00017598620281349345,
"loss": 0.2613,
"step": 9330
},
{
"epoch": 0.36063168462102785,
"grad_norm": 1.3657029867172241,
"learning_rate": 0.00017596046179389166,
"loss": 0.4968,
"step": 9340
},
{
"epoch": 0.3610177999150546,
"grad_norm": 2.424880027770996,
"learning_rate": 0.00017593472077428988,
"loss": 0.5982,
"step": 9350
},
{
"epoch": 0.36140391520908144,
"grad_norm": 6.178807735443115,
"learning_rate": 0.0001759089797546881,
"loss": 0.5355,
"step": 9360
},
{
"epoch": 0.3617900305031082,
"grad_norm": 1.5572419166564941,
"learning_rate": 0.0001758832387350863,
"loss": 0.4435,
"step": 9370
},
{
"epoch": 0.36217614579713503,
"grad_norm": 0.46649104356765747,
"learning_rate": 0.00017585749771548452,
"loss": 0.352,
"step": 9380
},
{
"epoch": 0.3625622610911618,
"grad_norm": 1.9611142873764038,
"learning_rate": 0.00017583175669588276,
"loss": 0.2684,
"step": 9390
},
{
"epoch": 0.3629483763851886,
"grad_norm": 1.7648595571517944,
"learning_rate": 0.00017580601567628094,
"loss": 0.3186,
"step": 9400
},
{
"epoch": 0.3633344916792154,
"grad_norm": 1.7970843315124512,
"learning_rate": 0.00017578027465667916,
"loss": 0.5339,
"step": 9410
},
{
"epoch": 0.3637206069732422,
"grad_norm": 3.084897994995117,
"learning_rate": 0.00017575453363707737,
"loss": 0.5143,
"step": 9420
},
{
"epoch": 0.36410672226726903,
"grad_norm": 1.440626621246338,
"learning_rate": 0.00017572879261747558,
"loss": 0.4067,
"step": 9430
},
{
"epoch": 0.3644928375612958,
"grad_norm": 0.44918450713157654,
"learning_rate": 0.0001757030515978738,
"loss": 0.2306,
"step": 9440
},
{
"epoch": 0.3648789528553226,
"grad_norm": 2.617272138595581,
"learning_rate": 0.000175677310578272,
"loss": 0.3166,
"step": 9450
},
{
"epoch": 0.3652650681493494,
"grad_norm": 2.575073719024658,
"learning_rate": 0.00017565156955867025,
"loss": 0.6645,
"step": 9460
},
{
"epoch": 0.3656511834433762,
"grad_norm": 0.9430664777755737,
"learning_rate": 0.00017562582853906844,
"loss": 0.2753,
"step": 9470
},
{
"epoch": 0.366037298737403,
"grad_norm": 1.9400445222854614,
"learning_rate": 0.00017560008751946665,
"loss": 0.4689,
"step": 9480
},
{
"epoch": 0.3664234140314298,
"grad_norm": 4.0443220138549805,
"learning_rate": 0.00017557434649986486,
"loss": 0.5373,
"step": 9490
},
{
"epoch": 0.3668095293254566,
"grad_norm": 3.4999184608459473,
"learning_rate": 0.00017554860548026308,
"loss": 0.3412,
"step": 9500
},
{
"epoch": 0.3671956446194834,
"grad_norm": 2.2023515701293945,
"learning_rate": 0.0001755228644606613,
"loss": 0.3385,
"step": 9510
},
{
"epoch": 0.36758175991351016,
"grad_norm": 1.213641881942749,
"learning_rate": 0.0001754971234410595,
"loss": 0.4785,
"step": 9520
},
{
"epoch": 0.367967875207537,
"grad_norm": 0.4178420603275299,
"learning_rate": 0.00017547138242145774,
"loss": 0.2605,
"step": 9530
},
{
"epoch": 0.36835399050156376,
"grad_norm": 2.676564931869507,
"learning_rate": 0.00017544564140185593,
"loss": 0.5297,
"step": 9540
},
{
"epoch": 0.3687401057955906,
"grad_norm": 0.8604353070259094,
"learning_rate": 0.00017541990038225414,
"loss": 0.3983,
"step": 9550
},
{
"epoch": 0.36912622108961735,
"grad_norm": 1.298893690109253,
"learning_rate": 0.00017539415936265236,
"loss": 0.3229,
"step": 9560
},
{
"epoch": 0.36951233638364417,
"grad_norm": 4.109025478363037,
"learning_rate": 0.00017536841834305057,
"loss": 0.519,
"step": 9570
},
{
"epoch": 0.36989845167767094,
"grad_norm": 3.440915584564209,
"learning_rate": 0.0001753426773234488,
"loss": 0.4061,
"step": 9580
},
{
"epoch": 0.37028456697169776,
"grad_norm": 0.1484186202287674,
"learning_rate": 0.000175316936303847,
"loss": 0.3443,
"step": 9590
},
{
"epoch": 0.3706706822657245,
"grad_norm": 3.114328145980835,
"learning_rate": 0.00017529119528424524,
"loss": 0.2451,
"step": 9600
},
{
"epoch": 0.37105679755975135,
"grad_norm": 1.8218796253204346,
"learning_rate": 0.00017526545426464342,
"loss": 0.2511,
"step": 9610
},
{
"epoch": 0.3714429128537781,
"grad_norm": 1.0732795000076294,
"learning_rate": 0.00017523971324504164,
"loss": 0.1581,
"step": 9620
},
{
"epoch": 0.37182902814780494,
"grad_norm": 1.0567959547042847,
"learning_rate": 0.00017521397222543985,
"loss": 0.1924,
"step": 9630
},
{
"epoch": 0.3722151434418317,
"grad_norm": 0.3467637896537781,
"learning_rate": 0.00017518823120583806,
"loss": 0.3571,
"step": 9640
},
{
"epoch": 0.37260125873585853,
"grad_norm": 2.6293838024139404,
"learning_rate": 0.0001751624901862363,
"loss": 0.3282,
"step": 9650
},
{
"epoch": 0.3729873740298853,
"grad_norm": 1.159696102142334,
"learning_rate": 0.0001751367491666345,
"loss": 0.2636,
"step": 9660
},
{
"epoch": 0.3733734893239121,
"grad_norm": 0.6884826421737671,
"learning_rate": 0.00017511100814703273,
"loss": 0.2842,
"step": 9670
},
{
"epoch": 0.3737596046179389,
"grad_norm": 3.789825201034546,
"learning_rate": 0.00017508526712743091,
"loss": 0.599,
"step": 9680
},
{
"epoch": 0.3741457199119657,
"grad_norm": 1.0705493688583374,
"learning_rate": 0.00017505952610782913,
"loss": 0.1746,
"step": 9690
},
{
"epoch": 0.37453183520599254,
"grad_norm": 1.8735803365707397,
"learning_rate": 0.00017503378508822734,
"loss": 0.3259,
"step": 9700
},
{
"epoch": 0.3749179505000193,
"grad_norm": 1.2987112998962402,
"learning_rate": 0.00017500804406862555,
"loss": 0.5738,
"step": 9710
},
{
"epoch": 0.3753040657940461,
"grad_norm": 1.5362507104873657,
"learning_rate": 0.0001749823030490238,
"loss": 0.3815,
"step": 9720
},
{
"epoch": 0.3756901810880729,
"grad_norm": 0.1640123724937439,
"learning_rate": 0.00017495656202942198,
"loss": 0.3672,
"step": 9730
},
{
"epoch": 0.3760762963820997,
"grad_norm": 0.6714594960212708,
"learning_rate": 0.00017493082100982022,
"loss": 0.2849,
"step": 9740
},
{
"epoch": 0.3764624116761265,
"grad_norm": 4.330246448516846,
"learning_rate": 0.0001749050799902184,
"loss": 0.4023,
"step": 9750
},
{
"epoch": 0.3768485269701533,
"grad_norm": 0.8616659641265869,
"learning_rate": 0.00017487933897061665,
"loss": 0.4434,
"step": 9760
},
{
"epoch": 0.3772346422641801,
"grad_norm": 2.6581578254699707,
"learning_rate": 0.00017485359795101486,
"loss": 0.4854,
"step": 9770
},
{
"epoch": 0.3776207575582069,
"grad_norm": 1.8269850015640259,
"learning_rate": 0.00017482785693141305,
"loss": 0.6033,
"step": 9780
},
{
"epoch": 0.37800687285223367,
"grad_norm": 2.256073236465454,
"learning_rate": 0.0001748021159118113,
"loss": 0.5317,
"step": 9790
},
{
"epoch": 0.3783929881462605,
"grad_norm": 0.8793076872825623,
"learning_rate": 0.00017477637489220947,
"loss": 0.3883,
"step": 9800
},
{
"epoch": 0.37877910344028726,
"grad_norm": 1.71831214427948,
"learning_rate": 0.00017475063387260772,
"loss": 0.2473,
"step": 9810
},
{
"epoch": 0.3791652187343141,
"grad_norm": 3.4802069664001465,
"learning_rate": 0.0001747248928530059,
"loss": 0.4847,
"step": 9820
},
{
"epoch": 0.37955133402834085,
"grad_norm": 5.419053077697754,
"learning_rate": 0.00017469915183340414,
"loss": 0.3668,
"step": 9830
},
{
"epoch": 0.37993744932236767,
"grad_norm": 1.567060112953186,
"learning_rate": 0.00017467341081380236,
"loss": 0.3342,
"step": 9840
},
{
"epoch": 0.38032356461639444,
"grad_norm": 3.0100274085998535,
"learning_rate": 0.00017464766979420054,
"loss": 0.476,
"step": 9850
},
{
"epoch": 0.38070967991042126,
"grad_norm": 0.7659344673156738,
"learning_rate": 0.00017462192877459878,
"loss": 0.2608,
"step": 9860
},
{
"epoch": 0.38109579520444803,
"grad_norm": 3.9540984630584717,
"learning_rate": 0.00017459618775499697,
"loss": 0.763,
"step": 9870
},
{
"epoch": 0.38148191049847485,
"grad_norm": 0.8768689036369324,
"learning_rate": 0.0001745704467353952,
"loss": 0.3365,
"step": 9880
},
{
"epoch": 0.3818680257925016,
"grad_norm": 0.9985928535461426,
"learning_rate": 0.00017454470571579342,
"loss": 0.3116,
"step": 9890
},
{
"epoch": 0.38225414108652844,
"grad_norm": 2.0326671600341797,
"learning_rate": 0.00017451896469619163,
"loss": 0.289,
"step": 9900
},
{
"epoch": 0.3826402563805552,
"grad_norm": 3.2696290016174316,
"learning_rate": 0.00017449322367658985,
"loss": 0.4097,
"step": 9910
},
{
"epoch": 0.38302637167458203,
"grad_norm": 3.048860788345337,
"learning_rate": 0.00017446748265698803,
"loss": 0.5181,
"step": 9920
},
{
"epoch": 0.3834124869686088,
"grad_norm": 1.7899913787841797,
"learning_rate": 0.00017444174163738627,
"loss": 0.2166,
"step": 9930
},
{
"epoch": 0.3837986022626356,
"grad_norm": 3.6762959957122803,
"learning_rate": 0.00017441600061778446,
"loss": 0.4971,
"step": 9940
},
{
"epoch": 0.3841847175566624,
"grad_norm": 0.9108519554138184,
"learning_rate": 0.0001743902595981827,
"loss": 0.4974,
"step": 9950
},
{
"epoch": 0.3845708328506892,
"grad_norm": 4.062527656555176,
"learning_rate": 0.00017436451857858091,
"loss": 0.4448,
"step": 9960
},
{
"epoch": 0.38495694814471604,
"grad_norm": 3.230902671813965,
"learning_rate": 0.00017433877755897913,
"loss": 0.2977,
"step": 9970
},
{
"epoch": 0.3853430634387428,
"grad_norm": 3.8190758228302,
"learning_rate": 0.00017431303653937734,
"loss": 0.4887,
"step": 9980
},
{
"epoch": 0.38572917873276963,
"grad_norm": 0.9079695343971252,
"learning_rate": 0.00017428729551977553,
"loss": 0.271,
"step": 9990
},
{
"epoch": 0.3861152940267964,
"grad_norm": 3.3730807304382324,
"learning_rate": 0.00017426155450017377,
"loss": 0.3782,
"step": 10000
},
{
"epoch": 0.3865014093208232,
"grad_norm": 1.07533860206604,
"learning_rate": 0.00017423581348057195,
"loss": 0.3905,
"step": 10010
},
{
"epoch": 0.38688752461485,
"grad_norm": 1.3856415748596191,
"learning_rate": 0.0001742100724609702,
"loss": 0.3757,
"step": 10020
},
{
"epoch": 0.3872736399088768,
"grad_norm": 5.751671314239502,
"learning_rate": 0.0001741843314413684,
"loss": 0.6657,
"step": 10030
},
{
"epoch": 0.3876597552029036,
"grad_norm": 0.6837680339813232,
"learning_rate": 0.00017415859042176662,
"loss": 0.2318,
"step": 10040
},
{
"epoch": 0.3880458704969304,
"grad_norm": 2.770787239074707,
"learning_rate": 0.00017413284940216483,
"loss": 0.3706,
"step": 10050
},
{
"epoch": 0.38843198579095717,
"grad_norm": 2.3058855533599854,
"learning_rate": 0.00017410710838256302,
"loss": 0.1641,
"step": 10060
},
{
"epoch": 0.388818101084984,
"grad_norm": 1.894718885421753,
"learning_rate": 0.00017408136736296126,
"loss": 0.4752,
"step": 10070
},
{
"epoch": 0.38920421637901076,
"grad_norm": 1.8346868753433228,
"learning_rate": 0.00017405562634335947,
"loss": 0.5007,
"step": 10080
},
{
"epoch": 0.3895903316730376,
"grad_norm": 5.277680397033691,
"learning_rate": 0.0001740298853237577,
"loss": 0.4399,
"step": 10090
},
{
"epoch": 0.38997644696706435,
"grad_norm": 1.306093692779541,
"learning_rate": 0.0001740041443041559,
"loss": 0.371,
"step": 10100
},
{
"epoch": 0.3903625622610912,
"grad_norm": 3.0306456089019775,
"learning_rate": 0.00017397840328455411,
"loss": 0.2515,
"step": 10110
},
{
"epoch": 0.39074867755511794,
"grad_norm": 0.7951543927192688,
"learning_rate": 0.00017395266226495233,
"loss": 0.3775,
"step": 10120
},
{
"epoch": 0.39113479284914476,
"grad_norm": 5.185150146484375,
"learning_rate": 0.00017392692124535051,
"loss": 0.3591,
"step": 10130
},
{
"epoch": 0.39152090814317153,
"grad_norm": 1.1718593835830688,
"learning_rate": 0.00017390118022574875,
"loss": 0.5484,
"step": 10140
},
{
"epoch": 0.39190702343719835,
"grad_norm": 1.6352128982543945,
"learning_rate": 0.00017387543920614697,
"loss": 0.2817,
"step": 10150
},
{
"epoch": 0.3922931387312251,
"grad_norm": 2.4863786697387695,
"learning_rate": 0.00017384969818654518,
"loss": 0.4027,
"step": 10160
},
{
"epoch": 0.39267925402525194,
"grad_norm": 2.069805383682251,
"learning_rate": 0.0001738239571669434,
"loss": 0.3559,
"step": 10170
},
{
"epoch": 0.3930653693192787,
"grad_norm": 1.671980619430542,
"learning_rate": 0.0001737982161473416,
"loss": 0.4405,
"step": 10180
},
{
"epoch": 0.39345148461330554,
"grad_norm": 4.298947334289551,
"learning_rate": 0.00017377247512773982,
"loss": 0.3005,
"step": 10190
},
{
"epoch": 0.3938375999073323,
"grad_norm": 0.4142851233482361,
"learning_rate": 0.000173746734108138,
"loss": 0.4248,
"step": 10200
},
{
"epoch": 0.3942237152013591,
"grad_norm": 3.5962865352630615,
"learning_rate": 0.00017372099308853625,
"loss": 0.27,
"step": 10210
},
{
"epoch": 0.39460983049538595,
"grad_norm": 2.20154070854187,
"learning_rate": 0.00017369525206893446,
"loss": 0.2858,
"step": 10220
},
{
"epoch": 0.3949959457894127,
"grad_norm": 0.2400553673505783,
"learning_rate": 0.00017366951104933267,
"loss": 0.2806,
"step": 10230
},
{
"epoch": 0.39538206108343954,
"grad_norm": 1.817741870880127,
"learning_rate": 0.0001736437700297309,
"loss": 0.3647,
"step": 10240
},
{
"epoch": 0.3957681763774663,
"grad_norm": 4.890044689178467,
"learning_rate": 0.0001736180290101291,
"loss": 0.4435,
"step": 10250
},
{
"epoch": 0.39615429167149313,
"grad_norm": 0.3407624065876007,
"learning_rate": 0.00017359228799052731,
"loss": 0.4857,
"step": 10260
},
{
"epoch": 0.3965404069655199,
"grad_norm": 2.4883463382720947,
"learning_rate": 0.00017356654697092553,
"loss": 0.2667,
"step": 10270
},
{
"epoch": 0.3969265222595467,
"grad_norm": 2.343823194503784,
"learning_rate": 0.00017354080595132374,
"loss": 0.3711,
"step": 10280
},
{
"epoch": 0.3973126375535735,
"grad_norm": 0.2056214064359665,
"learning_rate": 0.00017351506493172195,
"loss": 0.2695,
"step": 10290
},
{
"epoch": 0.3976987528476003,
"grad_norm": 0.20321065187454224,
"learning_rate": 0.00017348932391212017,
"loss": 0.3079,
"step": 10300
},
{
"epoch": 0.3980848681416271,
"grad_norm": 0.7993821501731873,
"learning_rate": 0.00017346358289251838,
"loss": 0.3599,
"step": 10310
},
{
"epoch": 0.3984709834356539,
"grad_norm": 2.0987348556518555,
"learning_rate": 0.0001734378418729166,
"loss": 0.3259,
"step": 10320
},
{
"epoch": 0.39885709872968067,
"grad_norm": 2.474246025085449,
"learning_rate": 0.0001734121008533148,
"loss": 0.3398,
"step": 10330
},
{
"epoch": 0.3992432140237075,
"grad_norm": 2.341064214706421,
"learning_rate": 0.00017338635983371302,
"loss": 0.5264,
"step": 10340
},
{
"epoch": 0.39962932931773426,
"grad_norm": 1.587437629699707,
"learning_rate": 0.00017336061881411123,
"loss": 0.4228,
"step": 10350
},
{
"epoch": 0.4000154446117611,
"grad_norm": 0.6692029237747192,
"learning_rate": 0.00017333487779450945,
"loss": 0.3576,
"step": 10360
},
{
"epoch": 0.40040155990578785,
"grad_norm": 2.088212251663208,
"learning_rate": 0.00017330913677490766,
"loss": 0.3096,
"step": 10370
},
{
"epoch": 0.4007876751998147,
"grad_norm": 1.5051954984664917,
"learning_rate": 0.00017328339575530587,
"loss": 0.3753,
"step": 10380
},
{
"epoch": 0.40117379049384144,
"grad_norm": 2.02595591545105,
"learning_rate": 0.0001732576547357041,
"loss": 0.3339,
"step": 10390
},
{
"epoch": 0.40155990578786827,
"grad_norm": 1.3062909841537476,
"learning_rate": 0.0001732319137161023,
"loss": 0.4301,
"step": 10400
},
{
"epoch": 0.40194602108189503,
"grad_norm": 2.5890421867370605,
"learning_rate": 0.00017320617269650051,
"loss": 0.3047,
"step": 10410
},
{
"epoch": 0.40233213637592186,
"grad_norm": 1.5994844436645508,
"learning_rate": 0.00017318043167689873,
"loss": 0.4158,
"step": 10420
},
{
"epoch": 0.4027182516699486,
"grad_norm": 0.5470211505889893,
"learning_rate": 0.00017315469065729694,
"loss": 0.4513,
"step": 10430
},
{
"epoch": 0.40310436696397545,
"grad_norm": 2.216935634613037,
"learning_rate": 0.00017312894963769515,
"loss": 0.5123,
"step": 10440
},
{
"epoch": 0.4034904822580022,
"grad_norm": 2.354724645614624,
"learning_rate": 0.00017310320861809337,
"loss": 0.2804,
"step": 10450
},
{
"epoch": 0.40387659755202904,
"grad_norm": 4.514159202575684,
"learning_rate": 0.00017307746759849158,
"loss": 0.3317,
"step": 10460
},
{
"epoch": 0.4042627128460558,
"grad_norm": 0.9874318242073059,
"learning_rate": 0.0001730517265788898,
"loss": 0.1948,
"step": 10470
},
{
"epoch": 0.40464882814008263,
"grad_norm": 2.0725696086883545,
"learning_rate": 0.000173025985559288,
"loss": 0.3627,
"step": 10480
},
{
"epoch": 0.40503494343410945,
"grad_norm": 2.4061577320098877,
"learning_rate": 0.00017300024453968622,
"loss": 0.3074,
"step": 10490
},
{
"epoch": 0.4054210587281362,
"grad_norm": 1.3369660377502441,
"learning_rate": 0.00017297450352008443,
"loss": 0.533,
"step": 10500
},
{
"epoch": 0.40580717402216304,
"grad_norm": 1.2730306386947632,
"learning_rate": 0.00017294876250048265,
"loss": 0.4688,
"step": 10510
},
{
"epoch": 0.4061932893161898,
"grad_norm": 0.6753021478652954,
"learning_rate": 0.00017292302148088086,
"loss": 0.4427,
"step": 10520
},
{
"epoch": 0.40657940461021663,
"grad_norm": 1.7279945611953735,
"learning_rate": 0.0001728972804612791,
"loss": 0.4921,
"step": 10530
},
{
"epoch": 0.4069655199042434,
"grad_norm": 0.9288708567619324,
"learning_rate": 0.0001728715394416773,
"loss": 0.363,
"step": 10540
},
{
"epoch": 0.4073516351982702,
"grad_norm": 0.5325084924697876,
"learning_rate": 0.0001728457984220755,
"loss": 0.4095,
"step": 10550
},
{
"epoch": 0.407737750492297,
"grad_norm": 1.2030489444732666,
"learning_rate": 0.0001728200574024737,
"loss": 0.3499,
"step": 10560
},
{
"epoch": 0.4081238657863238,
"grad_norm": 3.8157269954681396,
"learning_rate": 0.00017279431638287193,
"loss": 0.1622,
"step": 10570
},
{
"epoch": 0.4085099810803506,
"grad_norm": 0.6373336911201477,
"learning_rate": 0.00017276857536327017,
"loss": 0.4657,
"step": 10580
},
{
"epoch": 0.4088960963743774,
"grad_norm": 2.2850074768066406,
"learning_rate": 0.00017274283434366835,
"loss": 0.3585,
"step": 10590
},
{
"epoch": 0.40928221166840417,
"grad_norm": 0.8831659555435181,
"learning_rate": 0.0001727170933240666,
"loss": 0.293,
"step": 10600
},
{
"epoch": 0.409668326962431,
"grad_norm": 5.1165995597839355,
"learning_rate": 0.00017269135230446478,
"loss": 0.6539,
"step": 10610
},
{
"epoch": 0.41005444225645776,
"grad_norm": 4.901204586029053,
"learning_rate": 0.000172665611284863,
"loss": 0.4628,
"step": 10620
},
{
"epoch": 0.4104405575504846,
"grad_norm": 2.1492419242858887,
"learning_rate": 0.0001726398702652612,
"loss": 0.277,
"step": 10630
},
{
"epoch": 0.41082667284451135,
"grad_norm": 3.56510853767395,
"learning_rate": 0.00017261412924565942,
"loss": 0.4696,
"step": 10640
},
{
"epoch": 0.4112127881385382,
"grad_norm": 2.054769992828369,
"learning_rate": 0.00017258838822605766,
"loss": 0.4093,
"step": 10650
},
{
"epoch": 0.41159890343256494,
"grad_norm": 2.133474826812744,
"learning_rate": 0.00017256264720645585,
"loss": 0.3604,
"step": 10660
},
{
"epoch": 0.41198501872659177,
"grad_norm": 2.5062367916107178,
"learning_rate": 0.0001725369061868541,
"loss": 0.3916,
"step": 10670
},
{
"epoch": 0.41237113402061853,
"grad_norm": 0.431570827960968,
"learning_rate": 0.00017251116516725227,
"loss": 0.4048,
"step": 10680
},
{
"epoch": 0.41275724931464536,
"grad_norm": 1.2092580795288086,
"learning_rate": 0.0001724854241476505,
"loss": 0.602,
"step": 10690
},
{
"epoch": 0.4131433646086721,
"grad_norm": 2.712398052215576,
"learning_rate": 0.00017245968312804873,
"loss": 0.4172,
"step": 10700
},
{
"epoch": 0.41352947990269895,
"grad_norm": 3.914670467376709,
"learning_rate": 0.0001724339421084469,
"loss": 0.3843,
"step": 10710
},
{
"epoch": 0.4139155951967257,
"grad_norm": 1.7062132358551025,
"learning_rate": 0.00017240820108884515,
"loss": 0.343,
"step": 10720
},
{
"epoch": 0.41430171049075254,
"grad_norm": 0.5837095379829407,
"learning_rate": 0.00017238246006924334,
"loss": 0.3872,
"step": 10730
},
{
"epoch": 0.41468782578477936,
"grad_norm": 1.098900556564331,
"learning_rate": 0.00017235671904964158,
"loss": 0.2062,
"step": 10740
},
{
"epoch": 0.41507394107880613,
"grad_norm": 1.2533438205718994,
"learning_rate": 0.00017233097803003977,
"loss": 0.141,
"step": 10750
},
{
"epoch": 0.41546005637283295,
"grad_norm": 0.8688085079193115,
"learning_rate": 0.00017230523701043798,
"loss": 0.3686,
"step": 10760
},
{
"epoch": 0.4158461716668597,
"grad_norm": 1.868402361869812,
"learning_rate": 0.00017227949599083622,
"loss": 0.449,
"step": 10770
},
{
"epoch": 0.41623228696088654,
"grad_norm": 0.7168850898742676,
"learning_rate": 0.0001722537549712344,
"loss": 0.2317,
"step": 10780
},
{
"epoch": 0.4166184022549133,
"grad_norm": 3.1062309741973877,
"learning_rate": 0.00017222801395163265,
"loss": 0.4655,
"step": 10790
},
{
"epoch": 0.41700451754894013,
"grad_norm": 2.7296605110168457,
"learning_rate": 0.00017220227293203083,
"loss": 0.3934,
"step": 10800
},
{
"epoch": 0.4173906328429669,
"grad_norm": 2.3148224353790283,
"learning_rate": 0.00017217653191242907,
"loss": 0.2367,
"step": 10810
},
{
"epoch": 0.4177767481369937,
"grad_norm": 0.7049677968025208,
"learning_rate": 0.00017215079089282726,
"loss": 0.3157,
"step": 10820
},
{
"epoch": 0.4181628634310205,
"grad_norm": 3.3960344791412354,
"learning_rate": 0.00017212504987322547,
"loss": 0.4945,
"step": 10830
},
{
"epoch": 0.4185489787250473,
"grad_norm": 2.606316566467285,
"learning_rate": 0.0001720993088536237,
"loss": 0.4056,
"step": 10840
},
{
"epoch": 0.4189350940190741,
"grad_norm": 1.7469319105148315,
"learning_rate": 0.0001720735678340219,
"loss": 0.4176,
"step": 10850
},
{
"epoch": 0.4193212093131009,
"grad_norm": 0.8538552522659302,
"learning_rate": 0.00017204782681442014,
"loss": 0.3025,
"step": 10860
},
{
"epoch": 0.4197073246071277,
"grad_norm": 1.9576159715652466,
"learning_rate": 0.00017202208579481833,
"loss": 0.5626,
"step": 10870
},
{
"epoch": 0.4200934399011545,
"grad_norm": 0.8435356616973877,
"learning_rate": 0.00017199634477521657,
"loss": 0.2397,
"step": 10880
},
{
"epoch": 0.42047955519518126,
"grad_norm": 1.3026552200317383,
"learning_rate": 0.00017197060375561478,
"loss": 0.4793,
"step": 10890
},
{
"epoch": 0.4208656704892081,
"grad_norm": 1.8935116529464722,
"learning_rate": 0.00017194486273601297,
"loss": 0.2459,
"step": 10900
},
{
"epoch": 0.42125178578323486,
"grad_norm": 0.7297415137290955,
"learning_rate": 0.0001719191217164112,
"loss": 0.4115,
"step": 10910
},
{
"epoch": 0.4216379010772617,
"grad_norm": 2.730445146560669,
"learning_rate": 0.0001718933806968094,
"loss": 0.3467,
"step": 10920
},
{
"epoch": 0.42202401637128845,
"grad_norm": 1.5462249517440796,
"learning_rate": 0.00017186763967720763,
"loss": 0.2319,
"step": 10930
},
{
"epoch": 0.42241013166531527,
"grad_norm": 2.173388957977295,
"learning_rate": 0.00017184189865760582,
"loss": 0.3664,
"step": 10940
},
{
"epoch": 0.42279624695934204,
"grad_norm": 0.9086957573890686,
"learning_rate": 0.00017181615763800406,
"loss": 0.3928,
"step": 10950
},
{
"epoch": 0.42318236225336886,
"grad_norm": 1.6344754695892334,
"learning_rate": 0.00017179041661840227,
"loss": 0.32,
"step": 10960
},
{
"epoch": 0.4235684775473956,
"grad_norm": 3.7620887756347656,
"learning_rate": 0.00017176467559880049,
"loss": 0.3998,
"step": 10970
},
{
"epoch": 0.42395459284142245,
"grad_norm": 2.3914058208465576,
"learning_rate": 0.0001717389345791987,
"loss": 0.3003,
"step": 10980
},
{
"epoch": 0.4243407081354492,
"grad_norm": 1.1183325052261353,
"learning_rate": 0.00017171319355959689,
"loss": 0.2408,
"step": 10990
},
{
"epoch": 0.42472682342947604,
"grad_norm": 1.5570834875106812,
"learning_rate": 0.00017168745253999513,
"loss": 0.2638,
"step": 11000
}
],
"logging_steps": 10,
"max_steps": 77697,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 6698536796160000.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}