{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 5.541516245487364, "eval_steps": 770, "global_step": 770, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.007220216606498195, "grad_norm": 2.68963885307312, "learning_rate": 0.0, "loss": 1.880106806755066, "step": 1 }, { "epoch": 0.01444043321299639, "grad_norm": 2.3691370487213135, "learning_rate": 8e-05, "loss": 1.882678508758545, "step": 2 }, { "epoch": 0.021660649819494584, "grad_norm": 3.7024030685424805, "learning_rate": 0.00016, "loss": 1.9426246881484985, "step": 3 }, { "epoch": 0.02888086642599278, "grad_norm": 0.7646510601043701, "learning_rate": 0.00024, "loss": 1.6261038780212402, "step": 4 }, { "epoch": 0.036101083032490974, "grad_norm": 0.6628475189208984, "learning_rate": 0.00032, "loss": 1.478842854499817, "step": 5 }, { "epoch": 0.04332129963898917, "grad_norm": 0.7141556143760681, "learning_rate": 0.0004, "loss": 1.1376413106918335, "step": 6 }, { "epoch": 0.05054151624548736, "grad_norm": 0.7212001085281372, "learning_rate": 0.0003999983135391464, "loss": 1.0410869121551514, "step": 7 }, { "epoch": 0.05776173285198556, "grad_norm": 0.7515780925750732, "learning_rate": 0.0003999932541850271, "loss": 0.8246882557868958, "step": 8 }, { "epoch": 0.06498194945848375, "grad_norm": 2.474442481994629, "learning_rate": 0.0003999848220229661, "loss": 0.7787771821022034, "step": 9 }, { "epoch": 0.07220216606498195, "grad_norm": 1.2009592056274414, "learning_rate": 0.0003999730171951686, "loss": 0.6830324530601501, "step": 10 }, { "epoch": 0.07942238267148015, "grad_norm": 0.7545379996299744, "learning_rate": 0.00039995783990071825, "loss": 0.7438161969184875, "step": 11 }, { "epoch": 0.08664259927797834, "grad_norm": 0.9728625416755676, "learning_rate": 0.0003999392903955744, "loss": 0.6851149797439575, "step": 12 }, { "epoch": 0.09386281588447654, "grad_norm": 0.5509404540061951, "learning_rate": 0.00039991736899256693, "loss": 0.5853980183601379, "step": 13 }, { "epoch": 0.10108303249097472, "grad_norm": 0.715829074382782, "learning_rate": 0.0003998920760613919, "loss": 0.7855823040008545, "step": 14 }, { "epoch": 0.10830324909747292, "grad_norm": 0.6777167320251465, "learning_rate": 0.0003998634120286046, "loss": 0.5202292799949646, "step": 15 }, { "epoch": 0.11552346570397112, "grad_norm": 0.5348321199417114, "learning_rate": 0.0003998313773776128, "loss": 0.6527559757232666, "step": 16 }, { "epoch": 0.12274368231046931, "grad_norm": 0.4939132332801819, "learning_rate": 0.0003997959726486683, "loss": 0.584251880645752, "step": 17 }, { "epoch": 0.1299638989169675, "grad_norm": 0.47802454233169556, "learning_rate": 0.00039975719843885805, "loss": 0.5235528349876404, "step": 18 }, { "epoch": 0.1371841155234657, "grad_norm": 0.44072362780570984, "learning_rate": 0.0003997150554020938, "loss": 0.5184933543205261, "step": 19 }, { "epoch": 0.1444043321299639, "grad_norm": 0.4251253306865692, "learning_rate": 0.00039966954424910147, "loss": 0.6167617440223694, "step": 20 }, { "epoch": 0.15162454873646208, "grad_norm": 0.5198166370391846, "learning_rate": 0.00039962066574740886, "loss": 0.7412333488464355, "step": 21 }, { "epoch": 0.1588447653429603, "grad_norm": 0.44667476415634155, "learning_rate": 0.00039956842072133266, "loss": 0.6361726522445679, "step": 22 }, { "epoch": 0.16606498194945848, "grad_norm": 0.40787550806999207, "learning_rate": 0.0003995128100519649, "loss": 0.49531102180480957, "step": 23 }, { "epoch": 0.17328519855595667, "grad_norm": 0.4681910276412964, "learning_rate": 0.00039945383467715766, "loss": 0.584551990032196, "step": 24 }, { "epoch": 0.18050541516245489, "grad_norm": 0.41188451647758484, "learning_rate": 0.00039939149559150753, "loss": 0.6003429889678955, "step": 25 }, { "epoch": 0.18772563176895307, "grad_norm": 0.43262341618537903, "learning_rate": 0.0003993257938463389, "loss": 0.5725842118263245, "step": 26 }, { "epoch": 0.19494584837545126, "grad_norm": 0.45091742277145386, "learning_rate": 0.0003992567305496859, "loss": 0.6373569965362549, "step": 27 }, { "epoch": 0.20216606498194944, "grad_norm": 0.3687536120414734, "learning_rate": 0.00039918430686627406, "loss": 0.4958152174949646, "step": 28 }, { "epoch": 0.20938628158844766, "grad_norm": 0.3994904160499573, "learning_rate": 0.00039910852401750036, "loss": 0.6146813631057739, "step": 29 }, { "epoch": 0.21660649819494585, "grad_norm": 0.41312268376350403, "learning_rate": 0.0003990293832814129, "loss": 0.5313689708709717, "step": 30 }, { "epoch": 0.22382671480144403, "grad_norm": 0.37877994775772095, "learning_rate": 0.0003989468859926893, "loss": 0.5694329738616943, "step": 31 }, { "epoch": 0.23104693140794225, "grad_norm": 0.32251644134521484, "learning_rate": 0.0003988610335426139, "loss": 0.4476504921913147, "step": 32 }, { "epoch": 0.23826714801444043, "grad_norm": 0.40205422043800354, "learning_rate": 0.0003987718273790548, "loss": 0.4696302115917206, "step": 33 }, { "epoch": 0.24548736462093862, "grad_norm": 0.3664834201335907, "learning_rate": 0.0003986792690064389, "loss": 0.5304725170135498, "step": 34 }, { "epoch": 0.2527075812274368, "grad_norm": 0.5150560736656189, "learning_rate": 0.000398583359985727, "loss": 0.5886795520782471, "step": 35 }, { "epoch": 0.259927797833935, "grad_norm": 0.37544649839401245, "learning_rate": 0.0003984841019343872, "loss": 0.6374161839485168, "step": 36 }, { "epoch": 0.26714801444043323, "grad_norm": 0.3426123857498169, "learning_rate": 0.0003983814965263676, "loss": 0.507249116897583, "step": 37 }, { "epoch": 0.2743682310469314, "grad_norm": 0.36909183859825134, "learning_rate": 0.00039827554549206836, "loss": 0.5216564536094666, "step": 38 }, { "epoch": 0.2815884476534296, "grad_norm": 0.3864560127258301, "learning_rate": 0.00039816625061831206, "loss": 0.4607892632484436, "step": 39 }, { "epoch": 0.2888086642599278, "grad_norm": 0.3553301692008972, "learning_rate": 0.0003980536137483141, "loss": 0.6021184921264648, "step": 40 }, { "epoch": 0.296028880866426, "grad_norm": 0.3300604224205017, "learning_rate": 0.0003979376367816511, "loss": 0.49985307455062866, "step": 41 }, { "epoch": 0.30324909747292417, "grad_norm": 0.3894491493701935, "learning_rate": 0.0003978183216742292, "loss": 0.5403987169265747, "step": 42 }, { "epoch": 0.3104693140794224, "grad_norm": 0.32339245080947876, "learning_rate": 0.0003976956704382511, "loss": 0.5262795090675354, "step": 43 }, { "epoch": 0.3176895306859206, "grad_norm": 0.3121689260005951, "learning_rate": 0.00039756968514218166, "loss": 0.4988410174846649, "step": 44 }, { "epoch": 0.3249097472924188, "grad_norm": 0.37562957406044006, "learning_rate": 0.0003974403679107138, "loss": 0.5539376735687256, "step": 45 }, { "epoch": 0.33212996389891697, "grad_norm": 0.3560684323310852, "learning_rate": 0.0003973077209247319, "loss": 0.4829791784286499, "step": 46 }, { "epoch": 0.33935018050541516, "grad_norm": 0.3543923497200012, "learning_rate": 0.0003971717464212755, "loss": 0.4733068346977234, "step": 47 }, { "epoch": 0.34657039711191334, "grad_norm": 0.3253995180130005, "learning_rate": 0.0003970324466935013, "loss": 0.4318773150444031, "step": 48 }, { "epoch": 0.35379061371841153, "grad_norm": 0.3587929904460907, "learning_rate": 0.0003968898240906447, "loss": 0.6348429322242737, "step": 49 }, { "epoch": 0.36101083032490977, "grad_norm": 0.3772837519645691, "learning_rate": 0.0003967438810179802, "loss": 0.5192855596542358, "step": 50 }, { "epoch": 0.36823104693140796, "grad_norm": 0.3878256380558014, "learning_rate": 0.0003965946199367804, "loss": 0.4653000831604004, "step": 51 }, { "epoch": 0.37545126353790614, "grad_norm": 0.306704044342041, "learning_rate": 0.00039644204336427506, "loss": 0.46054646372795105, "step": 52 }, { "epoch": 0.38267148014440433, "grad_norm": 0.3290722072124481, "learning_rate": 0.0003962861538736084, "loss": 0.4637136459350586, "step": 53 }, { "epoch": 0.3898916967509025, "grad_norm": 0.3877376914024353, "learning_rate": 0.00039612695409379557, "loss": 0.6008591651916504, "step": 54 }, { "epoch": 0.3971119133574007, "grad_norm": 0.3738860785961151, "learning_rate": 0.00039596444670967864, "loss": 0.49255216121673584, "step": 55 }, { "epoch": 0.4043321299638989, "grad_norm": 0.43176719546318054, "learning_rate": 0.0003957986344618809, "loss": 0.6328399777412415, "step": 56 }, { "epoch": 0.41155234657039713, "grad_norm": 0.34965041279792786, "learning_rate": 0.00039562952014676116, "loss": 0.5274545550346375, "step": 57 }, { "epoch": 0.4187725631768953, "grad_norm": 0.3525114059448242, "learning_rate": 0.000395457106616366, "loss": 0.5174035429954529, "step": 58 }, { "epoch": 0.4259927797833935, "grad_norm": 0.36805179715156555, "learning_rate": 0.0003952813967783822, "loss": 0.5132086277008057, "step": 59 }, { "epoch": 0.4332129963898917, "grad_norm": 0.3320341110229492, "learning_rate": 0.00039510239359608735, "loss": 0.47466331720352173, "step": 60 }, { "epoch": 0.4404332129963899, "grad_norm": 0.36727970838546753, "learning_rate": 0.0003949201000883001, "loss": 0.5845056772232056, "step": 61 }, { "epoch": 0.44765342960288806, "grad_norm": 0.31494706869125366, "learning_rate": 0.000394734519329329, "loss": 0.5123739242553711, "step": 62 }, { "epoch": 0.4548736462093863, "grad_norm": 0.31959858536720276, "learning_rate": 0.00039454565444892094, "loss": 0.3888605237007141, "step": 63 }, { "epoch": 0.4620938628158845, "grad_norm": 0.4212755262851715, "learning_rate": 0.0003943535086322083, "loss": 0.6065781712532043, "step": 64 }, { "epoch": 0.4693140794223827, "grad_norm": 0.2938922643661499, "learning_rate": 0.0003941580851196549, "loss": 0.47488927841186523, "step": 65 }, { "epoch": 0.47653429602888087, "grad_norm": 0.27896228432655334, "learning_rate": 0.00039395938720700196, "loss": 0.36559349298477173, "step": 66 }, { "epoch": 0.48375451263537905, "grad_norm": 0.34284013509750366, "learning_rate": 0.0003937574182452118, "loss": 0.49944937229156494, "step": 67 }, { "epoch": 0.49097472924187724, "grad_norm": 0.3411049544811249, "learning_rate": 0.00039355218164041194, "loss": 0.5018436312675476, "step": 68 }, { "epoch": 0.4981949458483754, "grad_norm": 0.35431182384490967, "learning_rate": 0.0003933436808538375, "loss": 0.42643067240715027, "step": 69 }, { "epoch": 0.5054151624548736, "grad_norm": 0.3351145386695862, "learning_rate": 0.00039313191940177256, "loss": 0.5355833768844604, "step": 70 }, { "epoch": 0.5126353790613718, "grad_norm": 0.3065779507160187, "learning_rate": 0.00039291690085549105, "loss": 0.4329970180988312, "step": 71 }, { "epoch": 0.51985559566787, "grad_norm": 0.5208595395088196, "learning_rate": 0.00039269862884119664, "loss": 0.5850666165351868, "step": 72 }, { "epoch": 0.5270758122743683, "grad_norm": 0.3183553218841553, "learning_rate": 0.00039247710703996144, "loss": 0.5500473380088806, "step": 73 }, { "epoch": 0.5342960288808665, "grad_norm": 0.32788676023483276, "learning_rate": 0.0003922523391876638, "loss": 0.5200982093811035, "step": 74 }, { "epoch": 0.5415162454873647, "grad_norm": 0.2959986925125122, "learning_rate": 0.00039202432907492566, "loss": 0.4983074963092804, "step": 75 }, { "epoch": 0.5487364620938628, "grad_norm": 0.32182440161705017, "learning_rate": 0.0003917930805470483, "loss": 0.44362854957580566, "step": 76 }, { "epoch": 0.555956678700361, "grad_norm": 0.2816471755504608, "learning_rate": 0.00039155859750394754, "loss": 0.44515836238861084, "step": 77 }, { "epoch": 0.5631768953068592, "grad_norm": 0.290783554315567, "learning_rate": 0.0003913208839000882, "loss": 0.46542418003082275, "step": 78 }, { "epoch": 0.5703971119133574, "grad_norm": 0.366773396730423, "learning_rate": 0.0003910799437444171, "loss": 0.5370551943778992, "step": 79 }, { "epoch": 0.5776173285198556, "grad_norm": 0.32066911458969116, "learning_rate": 0.0003908357811002957, "loss": 0.5146384239196777, "step": 80 }, { "epoch": 0.5848375451263538, "grad_norm": 0.3420974016189575, "learning_rate": 0.00039058840008543135, "loss": 0.6218519806861877, "step": 81 }, { "epoch": 0.592057761732852, "grad_norm": 0.3156997263431549, "learning_rate": 0.000390337804871808, "loss": 0.549630343914032, "step": 82 }, { "epoch": 0.5992779783393501, "grad_norm": 0.3494798541069031, "learning_rate": 0.00039008399968561597, "loss": 0.5499352216720581, "step": 83 }, { "epoch": 0.6064981949458483, "grad_norm": 0.3652164936065674, "learning_rate": 0.0003898269888071803, "loss": 0.5562577843666077, "step": 84 }, { "epoch": 0.6137184115523465, "grad_norm": 0.34751319885253906, "learning_rate": 0.0003895667765708887, "loss": 0.5954532027244568, "step": 85 }, { "epoch": 0.6209386281588448, "grad_norm": 0.28338274359703064, "learning_rate": 0.0003893033673651189, "loss": 0.4894581735134125, "step": 86 }, { "epoch": 0.628158844765343, "grad_norm": 0.29019951820373535, "learning_rate": 0.00038903676563216394, "loss": 0.4635826349258423, "step": 87 }, { "epoch": 0.6353790613718412, "grad_norm": 0.2782176434993744, "learning_rate": 0.00038876697586815763, "loss": 0.4121658205986023, "step": 88 }, { "epoch": 0.6425992779783394, "grad_norm": 0.27983078360557556, "learning_rate": 0.0003884940026229988, "loss": 0.40188780426979065, "step": 89 }, { "epoch": 0.6498194945848376, "grad_norm": 0.2786664366722107, "learning_rate": 0.0003882178505002744, "loss": 0.44132208824157715, "step": 90 }, { "epoch": 0.6570397111913358, "grad_norm": 0.3355483412742615, "learning_rate": 0.0003879385241571817, "loss": 0.512315034866333, "step": 91 }, { "epoch": 0.6642599277978339, "grad_norm": 0.29554808139801025, "learning_rate": 0.0003876560283044503, "loss": 0.4540112018585205, "step": 92 }, { "epoch": 0.6714801444043321, "grad_norm": 0.3431720435619354, "learning_rate": 0.00038737036770626214, "loss": 0.45164141058921814, "step": 93 }, { "epoch": 0.6787003610108303, "grad_norm": 0.37090009450912476, "learning_rate": 0.0003870815471801714, "loss": 0.4878736734390259, "step": 94 }, { "epoch": 0.6859205776173285, "grad_norm": 0.35553082823753357, "learning_rate": 0.0003867895715970231, "loss": 0.5573399662971497, "step": 95 }, { "epoch": 0.6931407942238267, "grad_norm": 0.30914071202278137, "learning_rate": 0.00038649444588087115, "loss": 0.5417919754981995, "step": 96 }, { "epoch": 0.7003610108303249, "grad_norm": 0.27384090423583984, "learning_rate": 0.0003861961750088954, "loss": 0.4475242793560028, "step": 97 }, { "epoch": 0.7075812274368231, "grad_norm": 0.2919006049633026, "learning_rate": 0.0003858947640113171, "loss": 0.44341611862182617, "step": 98 }, { "epoch": 0.7148014440433214, "grad_norm": 0.33408623933792114, "learning_rate": 0.000385590217971315, "loss": 0.5056685209274292, "step": 99 }, { "epoch": 0.7220216606498195, "grad_norm": 0.3498390018939972, "learning_rate": 0.0003852825420249386, "loss": 0.45975226163864136, "step": 100 }, { "epoch": 0.7292418772563177, "grad_norm": 0.32381942868232727, "learning_rate": 0.0003849717413610224, "loss": 0.43437302112579346, "step": 101 }, { "epoch": 0.7364620938628159, "grad_norm": 0.30022308230400085, "learning_rate": 0.0003846578212210979, "loss": 0.41799411177635193, "step": 102 }, { "epoch": 0.7436823104693141, "grad_norm": 0.31061094999313354, "learning_rate": 0.0003843407868993054, "loss": 0.5682204365730286, "step": 103 }, { "epoch": 0.7509025270758123, "grad_norm": 0.31062883138656616, "learning_rate": 0.0003840206437423047, "loss": 0.3520132303237915, "step": 104 }, { "epoch": 0.7581227436823105, "grad_norm": 0.36826908588409424, "learning_rate": 0.00038369739714918464, "loss": 0.546819269657135, "step": 105 }, { "epoch": 0.7653429602888087, "grad_norm": 0.35304203629493713, "learning_rate": 0.00038337105257137263, "loss": 0.5633726716041565, "step": 106 }, { "epoch": 0.7725631768953068, "grad_norm": 0.2879215478897095, "learning_rate": 0.00038304161551254207, "loss": 0.42212507128715515, "step": 107 }, { "epoch": 0.779783393501805, "grad_norm": 0.32802048325538635, "learning_rate": 0.0003827090915285202, "loss": 0.4973101019859314, "step": 108 }, { "epoch": 0.7870036101083032, "grad_norm": 0.29462599754333496, "learning_rate": 0.0003823734862271937, "loss": 0.4543662667274475, "step": 109 }, { "epoch": 0.7942238267148014, "grad_norm": 0.34288832545280457, "learning_rate": 0.0003820348052684146, "loss": 0.5599743127822876, "step": 110 }, { "epoch": 0.8014440433212996, "grad_norm": 0.2871699333190918, "learning_rate": 0.00038169305436390477, "loss": 0.5132560133934021, "step": 111 }, { "epoch": 0.8086642599277978, "grad_norm": 0.3307580053806305, "learning_rate": 0.00038134823927715933, "loss": 0.5319223403930664, "step": 112 }, { "epoch": 0.8158844765342961, "grad_norm": 0.3292873799800873, "learning_rate": 0.0003810003658233498, "loss": 0.5933513045310974, "step": 113 }, { "epoch": 0.8231046931407943, "grad_norm": 0.34675127267837524, "learning_rate": 0.0003806494398692258, "loss": 0.6370081901550293, "step": 114 }, { "epoch": 0.8303249097472925, "grad_norm": 0.31837624311447144, "learning_rate": 0.00038029546733301615, "loss": 0.44530773162841797, "step": 115 }, { "epoch": 0.8375451263537906, "grad_norm": 0.3014152944087982, "learning_rate": 0.0003799384541843291, "loss": 0.5310161113739014, "step": 116 }, { "epoch": 0.8447653429602888, "grad_norm": 0.30693644285202026, "learning_rate": 0.00037957840644405164, "loss": 0.5011177659034729, "step": 117 }, { "epoch": 0.851985559566787, "grad_norm": 0.2840467095375061, "learning_rate": 0.000379215330184248, "loss": 0.37618446350097656, "step": 118 }, { "epoch": 0.8592057761732852, "grad_norm": 0.3380180597305298, "learning_rate": 0.00037884923152805715, "loss": 0.5536396503448486, "step": 119 }, { "epoch": 0.8664259927797834, "grad_norm": 0.30026137828826904, "learning_rate": 0.00037848011664958956, "loss": 0.5487097501754761, "step": 120 }, { "epoch": 0.8736462093862816, "grad_norm": 0.3990669250488281, "learning_rate": 0.00037810799177382325, "loss": 0.34712734818458557, "step": 121 }, { "epoch": 0.8808664259927798, "grad_norm": 0.3168427646160126, "learning_rate": 0.0003777328631764986, "loss": 0.44405245780944824, "step": 122 }, { "epoch": 0.8880866425992779, "grad_norm": 0.2761262059211731, "learning_rate": 0.0003773547371840124, "loss": 0.4112606942653656, "step": 123 }, { "epoch": 0.8953068592057761, "grad_norm": 0.27056002616882324, "learning_rate": 0.00037697362017331164, "loss": 0.45218032598495483, "step": 124 }, { "epoch": 0.9025270758122743, "grad_norm": 0.2833767831325531, "learning_rate": 0.00037658951857178543, "loss": 0.4833817481994629, "step": 125 }, { "epoch": 0.9097472924187726, "grad_norm": 0.30818742513656616, "learning_rate": 0.0003762024388571569, "loss": 0.4847618341445923, "step": 126 }, { "epoch": 0.9169675090252708, "grad_norm": 0.32447490096092224, "learning_rate": 0.00037581238755737405, "loss": 0.5068320035934448, "step": 127 }, { "epoch": 0.924187725631769, "grad_norm": 0.307781845331192, "learning_rate": 0.0003754193712504992, "loss": 0.455619752407074, "step": 128 }, { "epoch": 0.9314079422382672, "grad_norm": 0.29909971356391907, "learning_rate": 0.00037502339656459853, "loss": 0.5618020296096802, "step": 129 }, { "epoch": 0.9386281588447654, "grad_norm": 0.33699020743370056, "learning_rate": 0.0003746244701776302, "loss": 0.50435870885849, "step": 130 }, { "epoch": 0.9458483754512635, "grad_norm": 0.2932925820350647, "learning_rate": 0.0003742225988173315, "loss": 0.503471314907074, "step": 131 }, { "epoch": 0.9530685920577617, "grad_norm": 0.27647075057029724, "learning_rate": 0.00037381778926110567, "loss": 0.4979417324066162, "step": 132 }, { "epoch": 0.9602888086642599, "grad_norm": 0.23902934789657593, "learning_rate": 0.00037341004833590734, "loss": 0.4127359390258789, "step": 133 }, { "epoch": 0.9675090252707581, "grad_norm": 0.3308359980583191, "learning_rate": 0.0003729993829181276, "loss": 0.6199836730957031, "step": 134 }, { "epoch": 0.9747292418772563, "grad_norm": 0.30997487902641296, "learning_rate": 0.00037258579993347795, "loss": 0.5043972134590149, "step": 135 }, { "epoch": 0.9819494584837545, "grad_norm": 0.5225651264190674, "learning_rate": 0.00037216930635687357, "loss": 0.5078832507133484, "step": 136 }, { "epoch": 0.9891696750902527, "grad_norm": 0.2910468280315399, "learning_rate": 0.00037174990921231557, "loss": 0.5478032827377319, "step": 137 }, { "epoch": 0.9963898916967509, "grad_norm": 0.3448226749897003, "learning_rate": 0.00037132761557277257, "loss": 0.5461761355400085, "step": 138 }, { "epoch": 1.0, "grad_norm": 0.5731388926506042, "learning_rate": 0.00037090243256006155, "loss": 0.39478886127471924, "step": 139 }, { "epoch": 1.0072202166064983, "grad_norm": 0.3017331659793854, "learning_rate": 0.0003704743673447276, "loss": 0.3729703426361084, "step": 140 }, { "epoch": 1.0144404332129964, "grad_norm": 0.29272574186325073, "learning_rate": 0.0003700434271459229, "loss": 0.45498228073120117, "step": 141 }, { "epoch": 1.0216606498194947, "grad_norm": 0.2818097770214081, "learning_rate": 0.0003696096192312852, "loss": 0.37411680817604065, "step": 142 }, { "epoch": 1.0288808664259927, "grad_norm": 0.31325557827949524, "learning_rate": 0.00036917295091681526, "loss": 0.4299224019050598, "step": 143 }, { "epoch": 1.036101083032491, "grad_norm": 0.2971343994140625, "learning_rate": 0.0003687334295667533, "loss": 0.4103966951370239, "step": 144 }, { "epoch": 1.0433212996389891, "grad_norm": 0.32004889845848083, "learning_rate": 0.00036829106259345465, "loss": 0.49494248628616333, "step": 145 }, { "epoch": 1.0505415162454874, "grad_norm": 0.29688122868537903, "learning_rate": 0.00036784585745726535, "loss": 0.4550933241844177, "step": 146 }, { "epoch": 1.0577617328519855, "grad_norm": 0.3322184681892395, "learning_rate": 0.00036739782166639556, "loss": 0.46367114782333374, "step": 147 }, { "epoch": 1.0649819494584838, "grad_norm": 0.267844021320343, "learning_rate": 0.00036694696277679367, "loss": 0.33760106563568115, "step": 148 }, { "epoch": 1.0722021660649819, "grad_norm": 0.3240787386894226, "learning_rate": 0.0003664932883920182, "loss": 0.43709027767181396, "step": 149 }, { "epoch": 1.0794223826714802, "grad_norm": 0.29441356658935547, "learning_rate": 0.00036603680616311015, "loss": 0.4205693304538727, "step": 150 }, { "epoch": 1.0866425992779782, "grad_norm": 0.33195292949676514, "learning_rate": 0.00036557752378846357, "loss": 0.44472751021385193, "step": 151 }, { "epoch": 1.0938628158844765, "grad_norm": 0.3113086223602295, "learning_rate": 0.00036511544901369597, "loss": 0.3894999921321869, "step": 152 }, { "epoch": 1.1010830324909748, "grad_norm": 0.30926311016082764, "learning_rate": 0.00036465058963151744, "loss": 0.33893927931785583, "step": 153 }, { "epoch": 1.108303249097473, "grad_norm": 0.31725767254829407, "learning_rate": 0.00036418295348159967, "loss": 0.4384872019290924, "step": 154 }, { "epoch": 1.1155234657039712, "grad_norm": 0.3411625027656555, "learning_rate": 0.00036371254845044297, "loss": 0.4130849838256836, "step": 155 }, { "epoch": 1.1227436823104693, "grad_norm": 0.3470022976398468, "learning_rate": 0.0003632393824712444, "loss": 0.4436883330345154, "step": 156 }, { "epoch": 1.1299638989169676, "grad_norm": 0.31315702199935913, "learning_rate": 0.00036276346352376266, "loss": 0.39922523498535156, "step": 157 }, { "epoch": 1.1371841155234657, "grad_norm": 0.36140701174736023, "learning_rate": 0.00036228479963418465, "loss": 0.44266432523727417, "step": 158 }, { "epoch": 1.144404332129964, "grad_norm": 0.37621912360191345, "learning_rate": 0.0003618033988749895, "loss": 0.3703776001930237, "step": 159 }, { "epoch": 1.151624548736462, "grad_norm": 0.40457242727279663, "learning_rate": 0.0003613192693648125, "loss": 0.48410341143608093, "step": 160 }, { "epoch": 1.1588447653429603, "grad_norm": 0.35907864570617676, "learning_rate": 0.00036083241926830833, "loss": 0.3249608874320984, "step": 161 }, { "epoch": 1.1660649819494584, "grad_norm": 0.3328985869884491, "learning_rate": 0.00036034285679601336, "loss": 0.3642440140247345, "step": 162 }, { "epoch": 1.1732851985559567, "grad_norm": 0.3639651834964752, "learning_rate": 0.00035985059020420695, "loss": 0.4125223457813263, "step": 163 }, { "epoch": 1.1805054151624548, "grad_norm": 0.3122009038925171, "learning_rate": 0.0003593556277947725, "loss": 0.26333218812942505, "step": 164 }, { "epoch": 1.187725631768953, "grad_norm": 0.3917737901210785, "learning_rate": 0.0003588579779150572, "loss": 0.41683149337768555, "step": 165 }, { "epoch": 1.1949458483754514, "grad_norm": 0.3892008066177368, "learning_rate": 0.00035835764895773164, "loss": 0.34965166449546814, "step": 166 }, { "epoch": 1.2021660649819494, "grad_norm": 0.32206663489341736, "learning_rate": 0.0003578546493606477, "loss": 0.3289134204387665, "step": 167 }, { "epoch": 1.2093862815884477, "grad_norm": 0.373098760843277, "learning_rate": 0.0003573489876066967, "loss": 0.37044239044189453, "step": 168 }, { "epoch": 1.2166064981949458, "grad_norm": 0.4301343262195587, "learning_rate": 0.00035684067222366614, "loss": 0.46339157223701477, "step": 169 }, { "epoch": 1.2238267148014441, "grad_norm": 0.3458724915981293, "learning_rate": 0.00035632971178409603, "loss": 0.41736456751823425, "step": 170 }, { "epoch": 1.2310469314079422, "grad_norm": 0.45238223671913147, "learning_rate": 0.0003558161149051341, "loss": 0.4182761311531067, "step": 171 }, { "epoch": 1.2382671480144405, "grad_norm": 0.32411810755729675, "learning_rate": 0.0003552998902483907, "loss": 0.30536049604415894, "step": 172 }, { "epoch": 1.2454873646209386, "grad_norm": 0.40648043155670166, "learning_rate": 0.0003547810465197926, "loss": 0.45940107107162476, "step": 173 }, { "epoch": 1.2527075812274369, "grad_norm": 0.34474775195121765, "learning_rate": 0.00035425959246943614, "loss": 0.4267578423023224, "step": 174 }, { "epoch": 1.259927797833935, "grad_norm": 0.33473461866378784, "learning_rate": 0.00035373553689143977, "loss": 0.31750980019569397, "step": 175 }, { "epoch": 1.2671480144404332, "grad_norm": 0.29968079924583435, "learning_rate": 0.0003532088886237956, "loss": 0.3412264883518219, "step": 176 }, { "epoch": 1.2743682310469313, "grad_norm": 0.3362952172756195, "learning_rate": 0.00035267965654822063, "loss": 0.347223162651062, "step": 177 }, { "epoch": 1.2815884476534296, "grad_norm": 0.38441580533981323, "learning_rate": 0.0003521478495900065, "loss": 0.4132072925567627, "step": 178 }, { "epoch": 1.288808664259928, "grad_norm": 0.3259715437889099, "learning_rate": 0.00035161347671786947, "loss": 0.3259287476539612, "step": 179 }, { "epoch": 1.296028880866426, "grad_norm": 0.30564218759536743, "learning_rate": 0.00035107654694379876, "loss": 0.33284130692481995, "step": 180 }, { "epoch": 1.303249097472924, "grad_norm": 0.3464779853820801, "learning_rate": 0.0003505370693229049, "loss": 0.3613452613353729, "step": 181 }, { "epoch": 1.3104693140794224, "grad_norm": 0.4113422632217407, "learning_rate": 0.0003499950529532668, "loss": 0.49952858686447144, "step": 182 }, { "epoch": 1.3176895306859207, "grad_norm": 0.3503414988517761, "learning_rate": 0.0003494505069757782, "loss": 0.43928611278533936, "step": 183 }, { "epoch": 1.3249097472924187, "grad_norm": 0.3231402337551117, "learning_rate": 0.000348903440573994, "loss": 0.35302072763442993, "step": 184 }, { "epoch": 1.332129963898917, "grad_norm": 0.3849830627441406, "learning_rate": 0.00034835386297397486, "loss": 0.42382097244262695, "step": 185 }, { "epoch": 1.339350180505415, "grad_norm": 0.35116714239120483, "learning_rate": 0.00034780178344413185, "loss": 0.3274853229522705, "step": 186 }, { "epoch": 1.3465703971119134, "grad_norm": 0.3941578269004822, "learning_rate": 0.0003472472112950701, "loss": 0.34617680311203003, "step": 187 }, { "epoch": 1.3537906137184115, "grad_norm": 0.32104700803756714, "learning_rate": 0.00034669015587943184, "loss": 0.39410364627838135, "step": 188 }, { "epoch": 1.3610108303249098, "grad_norm": 0.42431211471557617, "learning_rate": 0.00034613062659173867, "loss": 0.49896761775016785, "step": 189 }, { "epoch": 1.3682310469314078, "grad_norm": 0.4387979209423065, "learning_rate": 0.00034556863286823284, "loss": 0.4435005784034729, "step": 190 }, { "epoch": 1.3754512635379061, "grad_norm": 0.3983216881752014, "learning_rate": 0.00034500418418671855, "loss": 0.44415876269340515, "step": 191 }, { "epoch": 1.3826714801444044, "grad_norm": 0.30740320682525635, "learning_rate": 0.00034443729006640184, "loss": 0.36664414405822754, "step": 192 }, { "epoch": 1.3898916967509025, "grad_norm": 0.35397860407829285, "learning_rate": 0.0003438679600677302, "loss": 0.36680418252944946, "step": 193 }, { "epoch": 1.3971119133574006, "grad_norm": 0.3522591292858124, "learning_rate": 0.0003432962037922312, "loss": 0.3908054232597351, "step": 194 }, { "epoch": 1.404332129963899, "grad_norm": 0.36350661516189575, "learning_rate": 0.0003427220308823505, "loss": 0.34363672137260437, "step": 195 }, { "epoch": 1.4115523465703972, "grad_norm": 0.3392064869403839, "learning_rate": 0.00034214545102128946, "loss": 0.40172988176345825, "step": 196 }, { "epoch": 1.4187725631768953, "grad_norm": 0.3826808035373688, "learning_rate": 0.0003415664739328418, "loss": 0.4292600452899933, "step": 197 }, { "epoch": 1.4259927797833936, "grad_norm": 0.4102179706096649, "learning_rate": 0.00034098510938122947, "loss": 0.40238481760025024, "step": 198 }, { "epoch": 1.4332129963898916, "grad_norm": 0.38712596893310547, "learning_rate": 0.00034040136717093796, "loss": 0.3980088233947754, "step": 199 }, { "epoch": 1.44043321299639, "grad_norm": 0.35182803869247437, "learning_rate": 0.0003398152571465512, "loss": 0.3569096028804779, "step": 200 }, { "epoch": 1.447653429602888, "grad_norm": 0.4129750728607178, "learning_rate": 0.0003392267891925854, "loss": 0.4729324281215668, "step": 201 }, { "epoch": 1.4548736462093863, "grad_norm": 0.3427615463733673, "learning_rate": 0.000338635973233322, "loss": 0.38466876745224, "step": 202 }, { "epoch": 1.4620938628158844, "grad_norm": 0.29937732219696045, "learning_rate": 0.0003380428192326411, "loss": 0.24431845545768738, "step": 203 }, { "epoch": 1.4693140794223827, "grad_norm": 0.35563215613365173, "learning_rate": 0.00033744733719385253, "loss": 0.34825772047042847, "step": 204 }, { "epoch": 1.476534296028881, "grad_norm": 0.3395077586174011, "learning_rate": 0.0003368495371595279, "loss": 0.37704023718833923, "step": 205 }, { "epoch": 1.483754512635379, "grad_norm": 0.32358789443969727, "learning_rate": 0.0003362494292113308, "loss": 0.29866379499435425, "step": 206 }, { "epoch": 1.4909747292418771, "grad_norm": 0.44219517707824707, "learning_rate": 0.00033564702346984684, "loss": 0.5036117434501648, "step": 207 }, { "epoch": 1.4981949458483754, "grad_norm": 0.358549565076828, "learning_rate": 0.00033504233009441287, "loss": 0.42227745056152344, "step": 208 }, { "epoch": 1.5054151624548737, "grad_norm": 0.3784799575805664, "learning_rate": 0.0003344353592829461, "loss": 0.4267178773880005, "step": 209 }, { "epoch": 1.5126353790613718, "grad_norm": 0.3619599938392639, "learning_rate": 0.00033382612127177166, "loss": 0.4075395166873932, "step": 210 }, { "epoch": 1.5198555956678699, "grad_norm": 0.44484809041023254, "learning_rate": 0.0003332146263354501, "loss": 0.5016494989395142, "step": 211 }, { "epoch": 1.5270758122743682, "grad_norm": 0.34379443526268005, "learning_rate": 0.00033260088478660407, "loss": 0.317508339881897, "step": 212 }, { "epoch": 1.5342960288808665, "grad_norm": 0.4008065164089203, "learning_rate": 0.0003319849069757446, "loss": 0.3974913954734802, "step": 213 }, { "epoch": 1.5415162454873648, "grad_norm": 0.4006946384906769, "learning_rate": 0.00033136670329109624, "loss": 0.5162532925605774, "step": 214 }, { "epoch": 1.5487364620938628, "grad_norm": 0.32255908846855164, "learning_rate": 0.0003307462841584223, "loss": 0.37669965624809265, "step": 215 }, { "epoch": 1.555956678700361, "grad_norm": 0.7169090509414673, "learning_rate": 0.0003301236600408484, "loss": 0.4477896988391876, "step": 216 }, { "epoch": 1.5631768953068592, "grad_norm": 0.3694959580898285, "learning_rate": 0.00032949884143868675, "loss": 0.45744967460632324, "step": 217 }, { "epoch": 1.5703971119133575, "grad_norm": 0.3709769546985626, "learning_rate": 0.0003288718388892583, "loss": 0.4952532947063446, "step": 218 }, { "epoch": 1.5776173285198556, "grad_norm": 0.29159069061279297, "learning_rate": 0.00032824266296671567, "loss": 0.28994375467300415, "step": 219 }, { "epoch": 1.5848375451263537, "grad_norm": 0.3600374460220337, "learning_rate": 0.0003276113242818645, "loss": 0.43825557827949524, "step": 220 }, { "epoch": 1.592057761732852, "grad_norm": 0.2843645215034485, "learning_rate": 0.0003269778334819846, "loss": 0.32124435901641846, "step": 221 }, { "epoch": 1.5992779783393503, "grad_norm": 0.3688318431377411, "learning_rate": 0.0003263422012506502, "loss": 0.4742942750453949, "step": 222 }, { "epoch": 1.6064981949458483, "grad_norm": 0.3213002383708954, "learning_rate": 0.00032570443830755015, "loss": 0.42100387811660767, "step": 223 }, { "epoch": 1.6137184115523464, "grad_norm": 0.3240319490432739, "learning_rate": 0.0003250645554083068, "loss": 0.29403799772262573, "step": 224 }, { "epoch": 1.6209386281588447, "grad_norm": 0.3569975197315216, "learning_rate": 0.0003244225633442948, "loss": 0.4555712938308716, "step": 225 }, { "epoch": 1.628158844765343, "grad_norm": 0.43396008014678955, "learning_rate": 0.00032377847294245895, "loss": 0.4877527356147766, "step": 226 }, { "epoch": 1.6353790613718413, "grad_norm": 0.3694853186607361, "learning_rate": 0.0003231322950651316, "loss": 0.41299888491630554, "step": 227 }, { "epoch": 1.6425992779783394, "grad_norm": 0.3358525335788727, "learning_rate": 0.00032248404060985, "loss": 0.4551587998867035, "step": 228 }, { "epoch": 1.6498194945848375, "grad_norm": 0.38492903113365173, "learning_rate": 0.0003218337205091715, "loss": 0.4316423535346985, "step": 229 }, { "epoch": 1.6570397111913358, "grad_norm": 0.34664952754974365, "learning_rate": 0.0003211813457304902, "loss": 0.39314523339271545, "step": 230 }, { "epoch": 1.664259927797834, "grad_norm": 0.3254722058773041, "learning_rate": 0.00032052692727585133, "loss": 0.3327226936817169, "step": 231 }, { "epoch": 1.6714801444043321, "grad_norm": 0.4386709928512573, "learning_rate": 0.0003198704761817658, "loss": 0.4558185935020447, "step": 232 }, { "epoch": 1.6787003610108302, "grad_norm": 0.5447010397911072, "learning_rate": 0.00031921200351902446, "loss": 0.48867422342300415, "step": 233 }, { "epoch": 1.6859205776173285, "grad_norm": 0.2976516783237457, "learning_rate": 0.000318551520392511, "loss": 0.3376544117927551, "step": 234 }, { "epoch": 1.6931407942238268, "grad_norm": 0.3854595124721527, "learning_rate": 0.00031788903794101477, "loss": 0.45083147287368774, "step": 235 }, { "epoch": 1.7003610108303249, "grad_norm": 0.37051668763160706, "learning_rate": 0.00031722456733704297, "loss": 0.4483460485935211, "step": 236 }, { "epoch": 1.707581227436823, "grad_norm": 0.42628687620162964, "learning_rate": 0.0003165581197866322, "loss": 0.5173429250717163, "step": 237 }, { "epoch": 1.7148014440433212, "grad_norm": 0.44184717535972595, "learning_rate": 0.0003158897065291596, "loss": 0.45995786786079407, "step": 238 }, { "epoch": 1.7220216606498195, "grad_norm": 0.3687286376953125, "learning_rate": 0.00031521933883715293, "loss": 0.33321040868759155, "step": 239 }, { "epoch": 1.7292418772563178, "grad_norm": 0.3514050841331482, "learning_rate": 0.00031454702801610103, "loss": 0.3461155891418457, "step": 240 }, { "epoch": 1.736462093862816, "grad_norm": 0.4148181676864624, "learning_rate": 0.0003138727854042627, "loss": 0.4323638677597046, "step": 241 }, { "epoch": 1.743682310469314, "grad_norm": 0.44285446405410767, "learning_rate": 0.0003131966223724756, "loss": 0.4780181646347046, "step": 242 }, { "epoch": 1.7509025270758123, "grad_norm": 0.3196166157722473, "learning_rate": 0.00031251855032396464, "loss": 0.3380797505378723, "step": 243 }, { "epoch": 1.7581227436823106, "grad_norm": 0.48847395181655884, "learning_rate": 0.0003118385806941494, "loss": 0.4878506064414978, "step": 244 }, { "epoch": 1.7653429602888087, "grad_norm": 0.398876816034317, "learning_rate": 0.0003111567249504515, "loss": 0.4532281458377838, "step": 245 }, { "epoch": 1.7725631768953067, "grad_norm": 0.34218019247055054, "learning_rate": 0.0003104729945921012, "loss": 0.3738666772842407, "step": 246 }, { "epoch": 1.779783393501805, "grad_norm": 0.4071042835712433, "learning_rate": 0.0003097874011499433, "loss": 0.4437788724899292, "step": 247 }, { "epoch": 1.7870036101083033, "grad_norm": 0.35893604159355164, "learning_rate": 0.0003090999561862428, "loss": 0.35174691677093506, "step": 248 }, { "epoch": 1.7942238267148014, "grad_norm": 0.34495967626571655, "learning_rate": 0.0003084106712944899, "loss": 0.34072697162628174, "step": 249 }, { "epoch": 1.8014440433212995, "grad_norm": 0.4653945565223694, "learning_rate": 0.0003077195580992045, "loss": 0.45023709535598755, "step": 250 }, { "epoch": 1.8086642599277978, "grad_norm": 0.42124322056770325, "learning_rate": 0.0003070266282557401, "loss": 0.4846741557121277, "step": 251 }, { "epoch": 1.815884476534296, "grad_norm": 0.38208404183387756, "learning_rate": 0.0003063318934500872, "loss": 0.3900390565395355, "step": 252 }, { "epoch": 1.8231046931407944, "grad_norm": 0.3911641538143158, "learning_rate": 0.0003056353653986764, "loss": 0.3795917332172394, "step": 253 }, { "epoch": 1.8303249097472925, "grad_norm": 0.34185168147087097, "learning_rate": 0.00030493705584818065, "loss": 0.39876672625541687, "step": 254 }, { "epoch": 1.8375451263537905, "grad_norm": 0.4229717552661896, "learning_rate": 0.000304236976575317, "loss": 0.357430100440979, "step": 255 }, { "epoch": 1.8447653429602888, "grad_norm": 0.35156992077827454, "learning_rate": 0.0003035351393866485, "loss": 0.37869274616241455, "step": 256 }, { "epoch": 1.8519855595667871, "grad_norm": 0.3945099711418152, "learning_rate": 0.0003028315561183845, "loss": 0.45211949944496155, "step": 257 }, { "epoch": 1.8592057761732852, "grad_norm": 0.31473544239997864, "learning_rate": 0.0003021262386361814, "loss": 0.32624444365501404, "step": 258 }, { "epoch": 1.8664259927797833, "grad_norm": 0.4487520158290863, "learning_rate": 0.00030141919883494247, "loss": 0.48553428053855896, "step": 259 }, { "epoch": 1.8736462093862816, "grad_norm": 0.3847423791885376, "learning_rate": 0.00030071044863861713, "loss": 0.457511842250824, "step": 260 }, { "epoch": 1.8808664259927799, "grad_norm": 0.3822494447231293, "learning_rate": 0.00030000000000000003, "loss": 0.38700124621391296, "step": 261 }, { "epoch": 1.888086642599278, "grad_norm": 0.37934961915016174, "learning_rate": 0.0002992878649005293, "loss": 0.4150720238685608, "step": 262 }, { "epoch": 1.895306859205776, "grad_norm": 0.37126392126083374, "learning_rate": 0.00029857405535008467, "loss": 0.3389851450920105, "step": 263 }, { "epoch": 1.9025270758122743, "grad_norm": 0.3264472484588623, "learning_rate": 0.0002978585833867847, "loss": 0.327554315328598, "step": 264 }, { "epoch": 1.9097472924187726, "grad_norm": 0.3889877200126648, "learning_rate": 0.00029714146107678413, "loss": 0.4393918812274933, "step": 265 }, { "epoch": 1.916967509025271, "grad_norm": 0.4926806092262268, "learning_rate": 0.0002964227005140698, "loss": 0.5468560457229614, "step": 266 }, { "epoch": 1.924187725631769, "grad_norm": 0.4259951114654541, "learning_rate": 0.0002957023138202573, "loss": 0.39216533303260803, "step": 267 }, { "epoch": 1.931407942238267, "grad_norm": 0.39791861176490784, "learning_rate": 0.00029498031314438626, "loss": 0.37104716897010803, "step": 268 }, { "epoch": 1.9386281588447654, "grad_norm": 0.3682475686073303, "learning_rate": 0.0002942567106627155, "loss": 0.41986221075057983, "step": 269 }, { "epoch": 1.9458483754512637, "grad_norm": 0.3230769634246826, "learning_rate": 0.00029353151857851735, "loss": 0.35548532009124756, "step": 270 }, { "epoch": 1.9530685920577617, "grad_norm": 0.4607864320278168, "learning_rate": 0.0002928047491218727, "loss": 0.4246275722980499, "step": 271 }, { "epoch": 1.9602888086642598, "grad_norm": 0.4302678108215332, "learning_rate": 0.0002920764145494638, "loss": 0.42147913575172424, "step": 272 }, { "epoch": 1.967509025270758, "grad_norm": 0.36212071776390076, "learning_rate": 0.00029134652714436814, "loss": 0.4659682512283325, "step": 273 }, { "epoch": 1.9747292418772564, "grad_norm": 0.31535375118255615, "learning_rate": 0.00029061509921585107, "loss": 0.38229790329933167, "step": 274 }, { "epoch": 1.9819494584837545, "grad_norm": 0.3685518801212311, "learning_rate": 0.0002898821430991582, "loss": 0.46467846632003784, "step": 275 }, { "epoch": 1.9891696750902526, "grad_norm": 0.3998563885688782, "learning_rate": 0.0002891476711553077, "loss": 0.42970407009124756, "step": 276 }, { "epoch": 1.9963898916967509, "grad_norm": 0.35800614953041077, "learning_rate": 0.00028841169577088115, "loss": 0.4408795237541199, "step": 277 }, { "epoch": 2.0, "grad_norm": 0.5006654858589172, "learning_rate": 0.0002876742293578155, "loss": 0.38447684049606323, "step": 278 }, { "epoch": 2.0072202166064983, "grad_norm": 0.31330448389053345, "learning_rate": 0.00028693528435319305, "loss": 0.30373793840408325, "step": 279 }, { "epoch": 2.0144404332129966, "grad_norm": 0.29946643114089966, "learning_rate": 0.0002861948732190319, "loss": 0.26699960231781006, "step": 280 }, { "epoch": 2.0216606498194944, "grad_norm": 0.26591944694519043, "learning_rate": 0.0002854530084420762, "loss": 0.23807109892368317, "step": 281 }, { "epoch": 2.0288808664259927, "grad_norm": 0.34681931138038635, "learning_rate": 0.00028470970253358487, "loss": 0.32200539112091064, "step": 282 }, { "epoch": 2.036101083032491, "grad_norm": 0.3542478382587433, "learning_rate": 0.0002839649680291211, "loss": 0.3349723815917969, "step": 283 }, { "epoch": 2.0433212996389893, "grad_norm": 0.3007175624370575, "learning_rate": 0.0002832188174883408, "loss": 0.2172810435295105, "step": 284 }, { "epoch": 2.050541516245487, "grad_norm": 0.381719708442688, "learning_rate": 0.00028247126349478075, "loss": 0.3058876693248749, "step": 285 }, { "epoch": 2.0577617328519855, "grad_norm": 0.4056151211261749, "learning_rate": 0.0002817223186556463, "loss": 0.25402823090553284, "step": 286 }, { "epoch": 2.064981949458484, "grad_norm": 0.5069922804832458, "learning_rate": 0.00028097199560159913, "loss": 0.2375117838382721, "step": 287 }, { "epoch": 2.072202166064982, "grad_norm": 0.574663519859314, "learning_rate": 0.00028022030698654374, "loss": 0.3134710490703583, "step": 288 }, { "epoch": 2.07942238267148, "grad_norm": 0.5239531993865967, "learning_rate": 0.00027946726548741443, "loss": 0.20423674583435059, "step": 289 }, { "epoch": 2.0866425992779782, "grad_norm": 0.5845417976379395, "learning_rate": 0.0002787128838039612, "loss": 0.3297584354877472, "step": 290 }, { "epoch": 2.0938628158844765, "grad_norm": 0.47120949625968933, "learning_rate": 0.00027795717465853585, "loss": 0.32672739028930664, "step": 291 }, { "epoch": 2.101083032490975, "grad_norm": 0.40136632323265076, "learning_rate": 0.00027720015079587743, "loss": 0.23871661722660065, "step": 292 }, { "epoch": 2.108303249097473, "grad_norm": 0.5156406164169312, "learning_rate": 0.00027644182498289677, "loss": 0.23360338807106018, "step": 293 }, { "epoch": 2.115523465703971, "grad_norm": 0.4302510917186737, "learning_rate": 0.0002756822100084621, "loss": 0.31100887060165405, "step": 294 }, { "epoch": 2.1227436823104693, "grad_norm": 0.3673456609249115, "learning_rate": 0.00027492131868318246, "loss": 0.21802300214767456, "step": 295 }, { "epoch": 2.1299638989169676, "grad_norm": 0.32138535380363464, "learning_rate": 0.00027415916383919216, "loss": 0.19008119404315948, "step": 296 }, { "epoch": 2.137184115523466, "grad_norm": 0.3203730285167694, "learning_rate": 0.00027339575832993444, "loss": 0.1889929622411728, "step": 297 }, { "epoch": 2.1444043321299637, "grad_norm": 0.4612288773059845, "learning_rate": 0.0002726311150299443, "loss": 0.22323200106620789, "step": 298 }, { "epoch": 2.151624548736462, "grad_norm": 0.36709311604499817, "learning_rate": 0.00027186524683463155, "loss": 0.20466932654380798, "step": 299 }, { "epoch": 2.1588447653429603, "grad_norm": 0.4253864586353302, "learning_rate": 0.0002710981666600636, "loss": 0.2431459128856659, "step": 300 }, { "epoch": 2.1660649819494586, "grad_norm": 0.4018993675708771, "learning_rate": 0.00027032988744274736, "loss": 0.2751820385456085, "step": 301 }, { "epoch": 2.1732851985559565, "grad_norm": 0.4428861737251282, "learning_rate": 0.000269560422139411, "loss": 0.269814133644104, "step": 302 }, { "epoch": 2.1805054151624548, "grad_norm": 0.47122061252593994, "learning_rate": 0.00026878978372678564, "loss": 0.22829411923885345, "step": 303 }, { "epoch": 2.187725631768953, "grad_norm": 0.4426339268684387, "learning_rate": 0.00026801798520138646, "loss": 0.2637779414653778, "step": 304 }, { "epoch": 2.1949458483754514, "grad_norm": 0.4073690176010132, "learning_rate": 0.0002672450395792935, "loss": 0.1904142051935196, "step": 305 }, { "epoch": 2.2021660649819497, "grad_norm": 0.4896828234195709, "learning_rate": 0.00026647095989593193, "loss": 0.34650635719299316, "step": 306 }, { "epoch": 2.2093862815884475, "grad_norm": 0.48900970816612244, "learning_rate": 0.0002656957592058528, "loss": 0.2542106807231903, "step": 307 }, { "epoch": 2.216606498194946, "grad_norm": 0.4600171446800232, "learning_rate": 0.0002649194505825121, "loss": 0.21008341014385223, "step": 308 }, { "epoch": 2.223826714801444, "grad_norm": 0.43380653858184814, "learning_rate": 0.00026414204711805103, "loss": 0.2683791518211365, "step": 309 }, { "epoch": 2.2310469314079424, "grad_norm": 0.5792773365974426, "learning_rate": 0.0002633635619230746, "loss": 0.33160915970802307, "step": 310 }, { "epoch": 2.2382671480144403, "grad_norm": 0.5626965165138245, "learning_rate": 0.0002625840081264309, "loss": 0.26700398325920105, "step": 311 }, { "epoch": 2.2454873646209386, "grad_norm": 0.4952887296676636, "learning_rate": 0.00026180339887498953, "loss": 0.2689879536628723, "step": 312 }, { "epoch": 2.252707581227437, "grad_norm": 0.4501405656337738, "learning_rate": 0.0002610217473334199, "loss": 0.26021096110343933, "step": 313 }, { "epoch": 2.259927797833935, "grad_norm": 0.36552777886390686, "learning_rate": 0.00026023906668396933, "loss": 0.26156651973724365, "step": 314 }, { "epoch": 2.2671480144404335, "grad_norm": 0.3733731806278229, "learning_rate": 0.00025945537012624054, "loss": 0.21406950056552887, "step": 315 }, { "epoch": 2.2743682310469313, "grad_norm": 0.4240867495536804, "learning_rate": 0.00025867067087696923, "loss": 0.32105109095573425, "step": 316 }, { "epoch": 2.2815884476534296, "grad_norm": 0.4198870360851288, "learning_rate": 0.000257884982169801, "loss": 0.3047807216644287, "step": 317 }, { "epoch": 2.288808664259928, "grad_norm": 0.4444144666194916, "learning_rate": 0.00025709831725506845, "loss": 0.23268437385559082, "step": 318 }, { "epoch": 2.2960288808664258, "grad_norm": 0.41258376836776733, "learning_rate": 0.00025631068939956726, "loss": 0.26838576793670654, "step": 319 }, { "epoch": 2.303249097472924, "grad_norm": 0.5179259777069092, "learning_rate": 0.00025552211188633293, "loss": 0.2882261574268341, "step": 320 }, { "epoch": 2.3104693140794224, "grad_norm": 0.4838998019695282, "learning_rate": 0.0002547325980144166, "loss": 0.30366814136505127, "step": 321 }, { "epoch": 2.3176895306859207, "grad_norm": 0.4862940311431885, "learning_rate": 0.0002539421610986605, "loss": 0.29283270239830017, "step": 322 }, { "epoch": 2.324909747292419, "grad_norm": 0.441685289144516, "learning_rate": 0.000253150814469474, "loss": 0.22393643856048584, "step": 323 }, { "epoch": 2.332129963898917, "grad_norm": 0.39844486117362976, "learning_rate": 0.0002523585714726081, "loss": 0.17744283378124237, "step": 324 }, { "epoch": 2.339350180505415, "grad_norm": 0.4977899491786957, "learning_rate": 0.0002515654454689307, "loss": 0.24069495499134064, "step": 325 }, { "epoch": 2.3465703971119134, "grad_norm": 0.4101859927177429, "learning_rate": 0.0002507714498342016, "loss": 0.25044453144073486, "step": 326 }, { "epoch": 2.3537906137184117, "grad_norm": 0.3521929085254669, "learning_rate": 0.0002499765979588462, "loss": 0.20651675760746002, "step": 327 }, { "epoch": 2.3610108303249095, "grad_norm": 0.5269883871078491, "learning_rate": 0.00024918090324773024, "loss": 0.301647424697876, "step": 328 }, { "epoch": 2.368231046931408, "grad_norm": 0.47672051191329956, "learning_rate": 0.0002483843791199335, "loss": 0.23506474494934082, "step": 329 }, { "epoch": 2.375451263537906, "grad_norm": 0.5546545386314392, "learning_rate": 0.00024758703900852375, "loss": 0.3066970705986023, "step": 330 }, { "epoch": 2.3826714801444044, "grad_norm": 0.3792511820793152, "learning_rate": 0.0002467888963603298, "loss": 0.20687991380691528, "step": 331 }, { "epoch": 2.3898916967509027, "grad_norm": 0.4949963092803955, "learning_rate": 0.00024598996463571474, "loss": 0.2865826189517975, "step": 332 }, { "epoch": 2.3971119133574006, "grad_norm": 0.4250241219997406, "learning_rate": 0.00024519025730834954, "loss": 0.25499409437179565, "step": 333 }, { "epoch": 2.404332129963899, "grad_norm": 0.30932316184043884, "learning_rate": 0.00024438978786498524, "loss": 0.18888786435127258, "step": 334 }, { "epoch": 2.411552346570397, "grad_norm": 0.5183025002479553, "learning_rate": 0.00024358856980522556, "loss": 0.2950358986854553, "step": 335 }, { "epoch": 2.4187725631768955, "grad_norm": 0.46398693323135376, "learning_rate": 0.00024278661664129948, "loss": 0.28177422285079956, "step": 336 }, { "epoch": 2.4259927797833933, "grad_norm": 0.4938473701477051, "learning_rate": 0.00024198394189783317, "loss": 0.31616371870040894, "step": 337 }, { "epoch": 2.4332129963898916, "grad_norm": 0.4439496099948883, "learning_rate": 0.00024118055911162198, "loss": 0.2702808380126953, "step": 338 }, { "epoch": 2.44043321299639, "grad_norm": 0.45705899596214294, "learning_rate": 0.00024037648183140205, "loss": 0.33506399393081665, "step": 339 }, { "epoch": 2.4476534296028882, "grad_norm": 0.4123782217502594, "learning_rate": 0.00023957172361762199, "loss": 0.2569669783115387, "step": 340 }, { "epoch": 2.4548736462093865, "grad_norm": 0.5178057551383972, "learning_rate": 0.00023876629804221402, "loss": 0.2357858568429947, "step": 341 }, { "epoch": 2.4620938628158844, "grad_norm": 0.5206875205039978, "learning_rate": 0.0002379602186883652, "loss": 0.30847087502479553, "step": 342 }, { "epoch": 2.4693140794223827, "grad_norm": 0.4264805018901825, "learning_rate": 0.00023715349915028823, "loss": 0.27529531717300415, "step": 343 }, { "epoch": 2.476534296028881, "grad_norm": 0.5162002444267273, "learning_rate": 0.00023634615303299233, "loss": 0.3068305253982544, "step": 344 }, { "epoch": 2.483754512635379, "grad_norm": 0.45099541544914246, "learning_rate": 0.00023553819395205378, "loss": 0.2681335508823395, "step": 345 }, { "epoch": 2.490974729241877, "grad_norm": 0.49566006660461426, "learning_rate": 0.00023472963553338613, "loss": 0.2626950144767761, "step": 346 }, { "epoch": 2.4981949458483754, "grad_norm": 0.4181882441043854, "learning_rate": 0.00023392049141301055, "loss": 0.2935839891433716, "step": 347 }, { "epoch": 2.5054151624548737, "grad_norm": 0.4298352897167206, "learning_rate": 0.000233110775236826, "loss": 0.2565048038959503, "step": 348 }, { "epoch": 2.512635379061372, "grad_norm": 0.46044519543647766, "learning_rate": 0.00023230050066037872, "loss": 0.2740374207496643, "step": 349 }, { "epoch": 2.51985559566787, "grad_norm": 0.3774242401123047, "learning_rate": 0.00023148968134863233, "loss": 0.23688335716724396, "step": 350 }, { "epoch": 2.527075812274368, "grad_norm": 0.4179275631904602, "learning_rate": 0.00023067833097573714, "loss": 0.2116960883140564, "step": 351 }, { "epoch": 2.5342960288808665, "grad_norm": 0.5179049968719482, "learning_rate": 0.0002298664632247994, "loss": 0.2446313500404358, "step": 352 }, { "epoch": 2.5415162454873648, "grad_norm": 0.44997915625572205, "learning_rate": 0.000229054091787651, "loss": 0.3158394694328308, "step": 353 }, { "epoch": 2.5487364620938626, "grad_norm": 0.41355574131011963, "learning_rate": 0.0002282412303646183, "loss": 0.2720823884010315, "step": 354 }, { "epoch": 2.555956678700361, "grad_norm": 0.36182165145874023, "learning_rate": 0.00022742789266429095, "loss": 0.22511911392211914, "step": 355 }, { "epoch": 2.563176895306859, "grad_norm": 0.43841102719306946, "learning_rate": 0.00022661409240329076, "loss": 0.21212677657604218, "step": 356 }, { "epoch": 2.5703971119133575, "grad_norm": 0.4933500289916992, "learning_rate": 0.0002257998433060407, "loss": 0.3581770956516266, "step": 357 }, { "epoch": 2.577617328519856, "grad_norm": 0.4325965344905853, "learning_rate": 0.00022498515910453296, "loss": 0.3252311646938324, "step": 358 }, { "epoch": 2.5848375451263537, "grad_norm": 0.45968931913375854, "learning_rate": 0.00022417005353809772, "loss": 0.3097744286060333, "step": 359 }, { "epoch": 2.592057761732852, "grad_norm": 0.43083667755126953, "learning_rate": 0.00022335454035317124, "loss": 0.25429341197013855, "step": 360 }, { "epoch": 2.5992779783393503, "grad_norm": 0.4504421353340149, "learning_rate": 0.00022253863330306425, "loss": 0.315784215927124, "step": 361 }, { "epoch": 2.606498194945848, "grad_norm": 0.44943565130233765, "learning_rate": 0.0002217223461477296, "loss": 0.22878926992416382, "step": 362 }, { "epoch": 2.6137184115523464, "grad_norm": 0.4279639720916748, "learning_rate": 0.00022090569265353072, "loss": 0.2623154819011688, "step": 363 }, { "epoch": 2.6209386281588447, "grad_norm": 0.4907088279724121, "learning_rate": 0.00022008868659300905, "loss": 0.317205548286438, "step": 364 }, { "epoch": 2.628158844765343, "grad_norm": 0.49302950501441956, "learning_rate": 0.000219271341744652, "loss": 0.30752891302108765, "step": 365 }, { "epoch": 2.6353790613718413, "grad_norm": 0.38831761479377747, "learning_rate": 0.00021845367189266042, "loss": 0.27115532755851746, "step": 366 }, { "epoch": 2.6425992779783396, "grad_norm": 0.47725188732147217, "learning_rate": 0.00021763569082671635, "loss": 0.20296484231948853, "step": 367 }, { "epoch": 2.6498194945848375, "grad_norm": 0.4198381006717682, "learning_rate": 0.00021681741234175027, "loss": 0.22836127877235413, "step": 368 }, { "epoch": 2.6570397111913358, "grad_norm": 0.5218875408172607, "learning_rate": 0.00021599885023770835, "loss": 0.29816269874572754, "step": 369 }, { "epoch": 2.664259927797834, "grad_norm": 0.4321244955062866, "learning_rate": 0.00021518001831932022, "loss": 0.21631377935409546, "step": 370 }, { "epoch": 2.671480144404332, "grad_norm": 0.4036301374435425, "learning_rate": 0.00021436093039586552, "loss": 0.19903570413589478, "step": 371 }, { "epoch": 2.67870036101083, "grad_norm": 0.543828010559082, "learning_rate": 0.00021354160028094153, "loss": 0.2922933101654053, "step": 372 }, { "epoch": 2.6859205776173285, "grad_norm": 0.5146738290786743, "learning_rate": 0.00021272204179222982, "loss": 0.2749359607696533, "step": 373 }, { "epoch": 2.693140794223827, "grad_norm": 0.47690504789352417, "learning_rate": 0.00021190226875126352, "loss": 0.296896755695343, "step": 374 }, { "epoch": 2.700361010830325, "grad_norm": 0.545520007610321, "learning_rate": 0.000211082294983194, "loss": 0.27578210830688477, "step": 375 }, { "epoch": 2.707581227436823, "grad_norm": 0.42757925391197205, "learning_rate": 0.00021026213431655792, "loss": 0.2315436154603958, "step": 376 }, { "epoch": 2.7148014440433212, "grad_norm": 0.5044121146202087, "learning_rate": 0.0002094418005830439, "loss": 0.34600791335105896, "step": 377 }, { "epoch": 2.7220216606498195, "grad_norm": 0.5155430436134338, "learning_rate": 0.00020862130761725917, "loss": 0.3247469663619995, "step": 378 }, { "epoch": 2.729241877256318, "grad_norm": 0.5196051001548767, "learning_rate": 0.00020780066925649634, "loss": 0.3277815878391266, "step": 379 }, { "epoch": 2.7364620938628157, "grad_norm": 0.44720444083213806, "learning_rate": 0.00020697989934050025, "loss": 0.25156158208847046, "step": 380 }, { "epoch": 2.743682310469314, "grad_norm": 0.43984708189964294, "learning_rate": 0.00020615901171123412, "loss": 0.28022634983062744, "step": 381 }, { "epoch": 2.7509025270758123, "grad_norm": 0.47703853249549866, "learning_rate": 0.00020533802021264648, "loss": 0.2780344486236572, "step": 382 }, { "epoch": 2.7581227436823106, "grad_norm": 0.43746253848075867, "learning_rate": 0.00020451693869043763, "loss": 0.23613542318344116, "step": 383 }, { "epoch": 2.765342960288809, "grad_norm": 0.41468992829322815, "learning_rate": 0.00020369578099182597, "loss": 0.3185104429721832, "step": 384 }, { "epoch": 2.7725631768953067, "grad_norm": 0.5149108171463013, "learning_rate": 0.0002028745609653147, "loss": 0.2768135070800781, "step": 385 }, { "epoch": 2.779783393501805, "grad_norm": 0.47348129749298096, "learning_rate": 0.00020205329246045798, "loss": 0.31962212920188904, "step": 386 }, { "epoch": 2.7870036101083033, "grad_norm": 0.4738944172859192, "learning_rate": 0.00020123198932762778, "loss": 0.23498407006263733, "step": 387 }, { "epoch": 2.794223826714801, "grad_norm": 0.48899075388908386, "learning_rate": 0.0002004106654177798, "loss": 0.23857903480529785, "step": 388 }, { "epoch": 2.8014440433212995, "grad_norm": 0.45867517590522766, "learning_rate": 0.00019958933458222026, "loss": 0.3246588706970215, "step": 389 }, { "epoch": 2.808664259927798, "grad_norm": 0.5840234160423279, "learning_rate": 0.0001987680106723723, "loss": 0.2935522198677063, "step": 390 }, { "epoch": 2.815884476534296, "grad_norm": 0.3764658272266388, "learning_rate": 0.00019794670753954204, "loss": 0.2096547782421112, "step": 391 }, { "epoch": 2.8231046931407944, "grad_norm": 0.44675493240356445, "learning_rate": 0.00019712543903468535, "loss": 0.2988222539424896, "step": 392 }, { "epoch": 2.8303249097472927, "grad_norm": 0.43063509464263916, "learning_rate": 0.00019630421900817407, "loss": 0.2222481369972229, "step": 393 }, { "epoch": 2.8375451263537905, "grad_norm": 0.47969475388526917, "learning_rate": 0.00019548306130956242, "loss": 0.26167353987693787, "step": 394 }, { "epoch": 2.844765342960289, "grad_norm": 0.47235366702079773, "learning_rate": 0.00019466197978735354, "loss": 0.30541300773620605, "step": 395 }, { "epoch": 2.851985559566787, "grad_norm": 0.47744160890579224, "learning_rate": 0.00019384098828876598, "loss": 0.23645427823066711, "step": 396 }, { "epoch": 2.859205776173285, "grad_norm": 0.5037774443626404, "learning_rate": 0.00019302010065949988, "loss": 0.26774662733078003, "step": 397 }, { "epoch": 2.8664259927797833, "grad_norm": 0.4689895510673523, "learning_rate": 0.0001921993307435037, "loss": 0.3405228853225708, "step": 398 }, { "epoch": 2.8736462093862816, "grad_norm": 0.44282686710357666, "learning_rate": 0.00019137869238274093, "loss": 0.26063668727874756, "step": 399 }, { "epoch": 2.88086642599278, "grad_norm": 0.4379267990589142, "learning_rate": 0.0001905581994169562, "loss": 0.26478758454322815, "step": 400 }, { "epoch": 2.888086642599278, "grad_norm": 0.519749641418457, "learning_rate": 0.00018973786568344205, "loss": 0.31215131282806396, "step": 401 }, { "epoch": 2.895306859205776, "grad_norm": 0.4515770375728607, "learning_rate": 0.000188917705016806, "loss": 0.2881377339363098, "step": 402 }, { "epoch": 2.9025270758122743, "grad_norm": 0.4650344252586365, "learning_rate": 0.0001880977312487365, "loss": 0.2557958960533142, "step": 403 }, { "epoch": 2.9097472924187726, "grad_norm": 0.4621173143386841, "learning_rate": 0.0001872779582077702, "loss": 0.23401126265525818, "step": 404 }, { "epoch": 2.916967509025271, "grad_norm": 0.4094926118850708, "learning_rate": 0.00018645839971905852, "loss": 0.24556361138820648, "step": 405 }, { "epoch": 2.9241877256317688, "grad_norm": 0.48592710494995117, "learning_rate": 0.0001856390696041345, "loss": 0.27918654680252075, "step": 406 }, { "epoch": 2.931407942238267, "grad_norm": 0.4039171040058136, "learning_rate": 0.00018481998168067983, "loss": 0.22918914258480072, "step": 407 }, { "epoch": 2.9386281588447654, "grad_norm": 0.5521522760391235, "learning_rate": 0.0001840011497622917, "loss": 0.35596632957458496, "step": 408 }, { "epoch": 2.9458483754512637, "grad_norm": 0.5527123212814331, "learning_rate": 0.0001831825876582498, "loss": 0.25526192784309387, "step": 409 }, { "epoch": 2.953068592057762, "grad_norm": 0.4386238157749176, "learning_rate": 0.0001823643091732837, "loss": 0.2627878189086914, "step": 410 }, { "epoch": 2.96028880866426, "grad_norm": 0.48318973183631897, "learning_rate": 0.00018154632810733962, "loss": 0.29408901929855347, "step": 411 }, { "epoch": 2.967509025270758, "grad_norm": 0.5193997025489807, "learning_rate": 0.00018072865825534805, "loss": 0.2762555480003357, "step": 412 }, { "epoch": 2.9747292418772564, "grad_norm": 0.488490492105484, "learning_rate": 0.000179911313406991, "loss": 0.3119064271450043, "step": 413 }, { "epoch": 2.9819494584837543, "grad_norm": 0.4461340010166168, "learning_rate": 0.00017909430734646935, "loss": 0.2738839089870453, "step": 414 }, { "epoch": 2.9891696750902526, "grad_norm": 0.4488433003425598, "learning_rate": 0.0001782776538522704, "loss": 0.25788843631744385, "step": 415 }, { "epoch": 2.996389891696751, "grad_norm": 0.5179741382598877, "learning_rate": 0.0001774613666969358, "loss": 0.2556946575641632, "step": 416 }, { "epoch": 3.0, "grad_norm": 0.6138741374015808, "learning_rate": 0.00017664545964682878, "loss": 0.23274531960487366, "step": 417 }, { "epoch": 3.0072202166064983, "grad_norm": 0.27745357155799866, "learning_rate": 0.00017582994646190233, "loss": 0.14754480123519897, "step": 418 }, { "epoch": 3.0144404332129966, "grad_norm": 0.40759217739105225, "learning_rate": 0.00017501484089546708, "loss": 0.21522817015647888, "step": 419 }, { "epoch": 3.0216606498194944, "grad_norm": 0.3123788833618164, "learning_rate": 0.00017420015669395938, "loss": 0.15608534216880798, "step": 420 }, { "epoch": 3.0288808664259927, "grad_norm": 0.4599035978317261, "learning_rate": 0.00017338590759670934, "loss": 0.1309528797864914, "step": 421 }, { "epoch": 3.036101083032491, "grad_norm": 0.38352325558662415, "learning_rate": 0.00017257210733570916, "loss": 0.1724625676870346, "step": 422 }, { "epoch": 3.0433212996389893, "grad_norm": 0.44771838188171387, "learning_rate": 0.00017175876963538178, "loss": 0.1661359965801239, "step": 423 }, { "epoch": 3.050541516245487, "grad_norm": 0.5998123288154602, "learning_rate": 0.00017094590821234906, "loss": 0.1318468153476715, "step": 424 }, { "epoch": 3.0577617328519855, "grad_norm": 0.47632187604904175, "learning_rate": 0.00017013353677520064, "loss": 0.10464094579219818, "step": 425 }, { "epoch": 3.064981949458484, "grad_norm": 0.3745962083339691, "learning_rate": 0.00016932166902426288, "loss": 0.10677687078714371, "step": 426 }, { "epoch": 3.072202166064982, "grad_norm": 0.6340925693511963, "learning_rate": 0.00016851031865136763, "loss": 0.1600916087627411, "step": 427 }, { "epoch": 3.07942238267148, "grad_norm": 0.8183390498161316, "learning_rate": 0.00016769949933962127, "loss": 0.13513822853565216, "step": 428 }, { "epoch": 3.0866425992779782, "grad_norm": 0.8068543076515198, "learning_rate": 0.00016688922476317398, "loss": 0.14319300651550293, "step": 429 }, { "epoch": 3.0938628158844765, "grad_norm": 0.6381204128265381, "learning_rate": 0.00016607950858698946, "loss": 0.1376371830701828, "step": 430 }, { "epoch": 3.101083032490975, "grad_norm": 0.5576755404472351, "learning_rate": 0.00016527036446661395, "loss": 0.14407199621200562, "step": 431 }, { "epoch": 3.108303249097473, "grad_norm": 0.6089445948600769, "learning_rate": 0.00016446180604794627, "loss": 0.1575869917869568, "step": 432 }, { "epoch": 3.115523465703971, "grad_norm": 0.6250684857368469, "learning_rate": 0.0001636538469670077, "loss": 0.17905279994010925, "step": 433 }, { "epoch": 3.1227436823104693, "grad_norm": 0.4243159592151642, "learning_rate": 0.00016284650084971185, "loss": 0.12401723116636276, "step": 434 }, { "epoch": 3.1299638989169676, "grad_norm": 0.4369412362575531, "learning_rate": 0.00016203978131163485, "loss": 0.11108125746250153, "step": 435 }, { "epoch": 3.137184115523466, "grad_norm": 0.45113661885261536, "learning_rate": 0.000161233701957786, "loss": 0.1364113986492157, "step": 436 }, { "epoch": 3.1444043321299637, "grad_norm": 0.33899277448654175, "learning_rate": 0.00016042827638237806, "loss": 0.1253184676170349, "step": 437 }, { "epoch": 3.151624548736462, "grad_norm": 0.48023247718811035, "learning_rate": 0.000159623518168598, "loss": 0.1729699671268463, "step": 438 }, { "epoch": 3.1588447653429603, "grad_norm": 0.4260300397872925, "learning_rate": 0.00015881944088837804, "loss": 0.1630483865737915, "step": 439 }, { "epoch": 3.1660649819494586, "grad_norm": 0.34896689653396606, "learning_rate": 0.00015801605810216685, "loss": 0.14477582275867462, "step": 440 }, { "epoch": 3.1732851985559565, "grad_norm": 0.46189841628074646, "learning_rate": 0.00015721338335870057, "loss": 0.16415920853614807, "step": 441 }, { "epoch": 3.1805054151624548, "grad_norm": 0.40856343507766724, "learning_rate": 0.00015641143019477446, "loss": 0.12028312683105469, "step": 442 }, { "epoch": 3.187725631768953, "grad_norm": 0.36561018228530884, "learning_rate": 0.00015561021213501483, "loss": 0.12599362432956696, "step": 443 }, { "epoch": 3.1949458483754514, "grad_norm": 0.4341471493244171, "learning_rate": 0.0001548097426916505, "loss": 0.14446969330310822, "step": 444 }, { "epoch": 3.2021660649819497, "grad_norm": 0.3864421844482422, "learning_rate": 0.00015401003536428534, "loss": 0.11642799526453018, "step": 445 }, { "epoch": 3.2093862815884475, "grad_norm": 0.5154263973236084, "learning_rate": 0.00015321110363967035, "loss": 0.17028403282165527, "step": 446 }, { "epoch": 3.216606498194946, "grad_norm": 0.4077145755290985, "learning_rate": 0.0001524129609914763, "loss": 0.11544163525104523, "step": 447 }, { "epoch": 3.223826714801444, "grad_norm": 0.41609570384025574, "learning_rate": 0.00015161562088006648, "loss": 0.1203891783952713, "step": 448 }, { "epoch": 3.2310469314079424, "grad_norm": 0.430624783039093, "learning_rate": 0.00015081909675226975, "loss": 0.09192588925361633, "step": 449 }, { "epoch": 3.2382671480144403, "grad_norm": 0.5538221001625061, "learning_rate": 0.0001500234020411538, "loss": 0.16532525420188904, "step": 450 }, { "epoch": 3.2454873646209386, "grad_norm": 0.6360041499137878, "learning_rate": 0.00014922855016579845, "loss": 0.13534685969352722, "step": 451 }, { "epoch": 3.252707581227437, "grad_norm": 0.556538462638855, "learning_rate": 0.00014843455453106925, "loss": 0.16164641082286835, "step": 452 }, { "epoch": 3.259927797833935, "grad_norm": 0.8166404366493225, "learning_rate": 0.00014764142852739193, "loss": 0.17865729331970215, "step": 453 }, { "epoch": 3.2671480144404335, "grad_norm": 0.6231580972671509, "learning_rate": 0.00014684918553052603, "loss": 0.11281149089336395, "step": 454 }, { "epoch": 3.2743682310469313, "grad_norm": 0.5884938836097717, "learning_rate": 0.00014605783890133948, "loss": 0.15221013128757477, "step": 455 }, { "epoch": 3.2815884476534296, "grad_norm": 0.40336933732032776, "learning_rate": 0.00014526740198558344, "loss": 0.08725619316101074, "step": 456 }, { "epoch": 3.288808664259928, "grad_norm": 0.7082973718643188, "learning_rate": 0.0001444778881136671, "loss": 0.1795981079339981, "step": 457 }, { "epoch": 3.2960288808664258, "grad_norm": 0.44312670826911926, "learning_rate": 0.0001436893106004328, "loss": 0.11532506346702576, "step": 458 }, { "epoch": 3.303249097472924, "grad_norm": 0.4892312288284302, "learning_rate": 0.0001429016827449316, "loss": 0.15218153595924377, "step": 459 }, { "epoch": 3.3104693140794224, "grad_norm": 0.36000511050224304, "learning_rate": 0.000142115017830199, "loss": 0.1174839586019516, "step": 460 }, { "epoch": 3.3176895306859207, "grad_norm": 0.42502933740615845, "learning_rate": 0.00014132932912303082, "loss": 0.1354883909225464, "step": 461 }, { "epoch": 3.324909747292419, "grad_norm": 0.5004318356513977, "learning_rate": 0.00014054462987375947, "loss": 0.14879979193210602, "step": 462 }, { "epoch": 3.332129963898917, "grad_norm": 0.5428838729858398, "learning_rate": 0.00013976093331603072, "loss": 0.1601594090461731, "step": 463 }, { "epoch": 3.339350180505415, "grad_norm": 0.5523360967636108, "learning_rate": 0.00013897825266658015, "loss": 0.1349392980337143, "step": 464 }, { "epoch": 3.3465703971119134, "grad_norm": 0.4672771394252777, "learning_rate": 0.00013819660112501054, "loss": 0.1563996523618698, "step": 465 }, { "epoch": 3.3537906137184117, "grad_norm": 0.4747507870197296, "learning_rate": 0.00013741599187356917, "loss": 0.12792187929153442, "step": 466 }, { "epoch": 3.3610108303249095, "grad_norm": 0.4293171167373657, "learning_rate": 0.00013663643807692547, "loss": 0.13704359531402588, "step": 467 }, { "epoch": 3.368231046931408, "grad_norm": 0.406646728515625, "learning_rate": 0.00013585795288194902, "loss": 0.14688344299793243, "step": 468 }, { "epoch": 3.375451263537906, "grad_norm": 0.44870486855506897, "learning_rate": 0.00013508054941748792, "loss": 0.11994664371013641, "step": 469 }, { "epoch": 3.3826714801444044, "grad_norm": 0.46383994817733765, "learning_rate": 0.00013430424079414728, "loss": 0.13228918612003326, "step": 470 }, { "epoch": 3.3898916967509027, "grad_norm": 0.873016893863678, "learning_rate": 0.0001335290401040681, "loss": 0.14495904743671417, "step": 471 }, { "epoch": 3.3971119133574006, "grad_norm": 0.5575613975524902, "learning_rate": 0.0001327549604207066, "loss": 0.17171433568000793, "step": 472 }, { "epoch": 3.404332129963899, "grad_norm": 0.541860818862915, "learning_rate": 0.00013198201479861353, "loss": 0.14465925097465515, "step": 473 }, { "epoch": 3.411552346570397, "grad_norm": 0.5210021734237671, "learning_rate": 0.00013121021627321438, "loss": 0.15102122724056244, "step": 474 }, { "epoch": 3.4187725631768955, "grad_norm": 0.46373456716537476, "learning_rate": 0.00013043957786058904, "loss": 0.15518702566623688, "step": 475 }, { "epoch": 3.4259927797833933, "grad_norm": 0.48513785004615784, "learning_rate": 0.00012967011255725263, "loss": 0.14110815525054932, "step": 476 }, { "epoch": 3.4332129963898916, "grad_norm": 0.5956123471260071, "learning_rate": 0.0001289018333399364, "loss": 0.14913466572761536, "step": 477 }, { "epoch": 3.44043321299639, "grad_norm": 0.5387746095657349, "learning_rate": 0.0001281347531653685, "loss": 0.14882197976112366, "step": 478 }, { "epoch": 3.4476534296028882, "grad_norm": 0.43160125613212585, "learning_rate": 0.00012736888497005578, "loss": 0.13204768300056458, "step": 479 }, { "epoch": 3.4548736462093865, "grad_norm": 0.41130170226097107, "learning_rate": 0.0001266042416700656, "loss": 0.13427188992500305, "step": 480 }, { "epoch": 3.4620938628158844, "grad_norm": 0.5126029253005981, "learning_rate": 0.00012584083616080786, "loss": 0.12719954550266266, "step": 481 }, { "epoch": 3.4693140794223827, "grad_norm": 0.4994872212409973, "learning_rate": 0.0001250786813168176, "loss": 0.15062202513217926, "step": 482 }, { "epoch": 3.476534296028881, "grad_norm": 0.4697980284690857, "learning_rate": 0.00012431778999153796, "loss": 0.14715459942817688, "step": 483 }, { "epoch": 3.483754512635379, "grad_norm": 0.5356115102767944, "learning_rate": 0.00012355817501710328, "loss": 0.12667298316955566, "step": 484 }, { "epoch": 3.490974729241877, "grad_norm": 0.48629310727119446, "learning_rate": 0.00012279984920412264, "loss": 0.12396648526191711, "step": 485 }, { "epoch": 3.4981949458483754, "grad_norm": 0.6295841336250305, "learning_rate": 0.00012204282534146414, "loss": 0.1579296588897705, "step": 486 }, { "epoch": 3.5054151624548737, "grad_norm": 0.524736225605011, "learning_rate": 0.00012128711619603885, "loss": 0.14176616072654724, "step": 487 }, { "epoch": 3.512635379061372, "grad_norm": 0.4429115355014801, "learning_rate": 0.0001205327345125856, "loss": 0.16075468063354492, "step": 488 }, { "epoch": 3.51985559566787, "grad_norm": 0.5202926993370056, "learning_rate": 0.00011977969301345627, "loss": 0.155876025557518, "step": 489 }, { "epoch": 3.527075812274368, "grad_norm": 0.5090999603271484, "learning_rate": 0.00011902800439840091, "loss": 0.1420796513557434, "step": 490 }, { "epoch": 3.5342960288808665, "grad_norm": 0.5266745686531067, "learning_rate": 0.0001182776813443537, "loss": 0.1911281943321228, "step": 491 }, { "epoch": 3.5415162454873648, "grad_norm": 0.4100717306137085, "learning_rate": 0.00011752873650521934, "loss": 0.12646690011024475, "step": 492 }, { "epoch": 3.5487364620938626, "grad_norm": 0.4897909164428711, "learning_rate": 0.00011678118251165929, "loss": 0.14893396198749542, "step": 493 }, { "epoch": 3.555956678700361, "grad_norm": 0.48633724451065063, "learning_rate": 0.00011603503197087893, "loss": 0.11430861800909042, "step": 494 }, { "epoch": 3.563176895306859, "grad_norm": 0.3797968327999115, "learning_rate": 0.00011529029746641522, "loss": 0.10336866229772568, "step": 495 }, { "epoch": 3.5703971119133575, "grad_norm": 0.4714502692222595, "learning_rate": 0.00011454699155792387, "loss": 0.12609894573688507, "step": 496 }, { "epoch": 3.577617328519856, "grad_norm": 0.5311459898948669, "learning_rate": 0.00011380512678096805, "loss": 0.14762037992477417, "step": 497 }, { "epoch": 3.5848375451263537, "grad_norm": 0.44842103123664856, "learning_rate": 0.00011306471564680703, "loss": 0.1137317568063736, "step": 498 }, { "epoch": 3.592057761732852, "grad_norm": 0.6083047986030579, "learning_rate": 0.0001123257706421845, "loss": 0.1538015455007553, "step": 499 }, { "epoch": 3.5992779783393503, "grad_norm": 0.454277902841568, "learning_rate": 0.00011158830422911882, "loss": 0.13900214433670044, "step": 500 }, { "epoch": 3.606498194945848, "grad_norm": 0.46477752923965454, "learning_rate": 0.00011085232884469236, "loss": 0.1317400485277176, "step": 501 }, { "epoch": 3.6137184115523464, "grad_norm": 0.5491266250610352, "learning_rate": 0.00011011785690084178, "loss": 0.20672887563705444, "step": 502 }, { "epoch": 3.6209386281588447, "grad_norm": 0.48666200041770935, "learning_rate": 0.00010938490078414902, "loss": 0.15307360887527466, "step": 503 }, { "epoch": 3.628158844765343, "grad_norm": 0.557000458240509, "learning_rate": 0.00010865347285563189, "loss": 0.18239088356494904, "step": 504 }, { "epoch": 3.6353790613718413, "grad_norm": 0.45608651638031006, "learning_rate": 0.0001079235854505362, "loss": 0.15936806797981262, "step": 505 }, { "epoch": 3.6425992779783396, "grad_norm": 0.455575555562973, "learning_rate": 0.00010719525087812736, "loss": 0.12316213548183441, "step": 506 }, { "epoch": 3.6498194945848375, "grad_norm": 0.48413556814193726, "learning_rate": 0.00010646848142148267, "loss": 0.13552163541316986, "step": 507 }, { "epoch": 3.6570397111913358, "grad_norm": 0.31646862626075745, "learning_rate": 0.0001057432893372846, "loss": 0.09000004827976227, "step": 508 }, { "epoch": 3.664259927797834, "grad_norm": 0.43280190229415894, "learning_rate": 0.00010501968685561379, "loss": 0.13823340833187103, "step": 509 }, { "epoch": 3.671480144404332, "grad_norm": 0.48566341400146484, "learning_rate": 0.00010429768617974271, "loss": 0.1494002640247345, "step": 510 }, { "epoch": 3.67870036101083, "grad_norm": 0.5894930362701416, "learning_rate": 0.00010357729948593022, "loss": 0.16500279307365417, "step": 511 }, { "epoch": 3.6859205776173285, "grad_norm": 0.4345530867576599, "learning_rate": 0.00010285853892321597, "loss": 0.15392449498176575, "step": 512 }, { "epoch": 3.693140794223827, "grad_norm": 0.38582298159599304, "learning_rate": 0.0001021414166132153, "loss": 0.11830861121416092, "step": 513 }, { "epoch": 3.700361010830325, "grad_norm": 0.5571300983428955, "learning_rate": 0.00010142594464991538, "loss": 0.17028063535690308, "step": 514 }, { "epoch": 3.707581227436823, "grad_norm": 0.442558616399765, "learning_rate": 0.00010071213509947078, "loss": 0.13233450055122375, "step": 515 }, { "epoch": 3.7148014440433212, "grad_norm": 0.38991135358810425, "learning_rate": 0.00010000000000000005, "loss": 0.11185239255428314, "step": 516 }, { "epoch": 3.7220216606498195, "grad_norm": 0.46868595480918884, "learning_rate": 9.928955136138292e-05, "loss": 0.13171209394931793, "step": 517 }, { "epoch": 3.729241877256318, "grad_norm": 0.3938763439655304, "learning_rate": 9.858080116505763e-05, "loss": 0.12150520086288452, "step": 518 }, { "epoch": 3.7364620938628157, "grad_norm": 0.5428814888000488, "learning_rate": 9.787376136381865e-05, "loss": 0.17324000597000122, "step": 519 }, { "epoch": 3.743682310469314, "grad_norm": 0.5528351664543152, "learning_rate": 9.716844388161555e-05, "loss": 0.13830417394638062, "step": 520 }, { "epoch": 3.7509025270758123, "grad_norm": 0.43436095118522644, "learning_rate": 9.646486061335151e-05, "loss": 0.12715619802474976, "step": 521 }, { "epoch": 3.7581227436823106, "grad_norm": 0.5142949223518372, "learning_rate": 9.576302342468297e-05, "loss": 0.13084562122821808, "step": 522 }, { "epoch": 3.765342960288809, "grad_norm": 0.6056551933288574, "learning_rate": 9.506294415181942e-05, "loss": 0.14757800102233887, "step": 523 }, { "epoch": 3.7725631768953067, "grad_norm": 0.7759309411048889, "learning_rate": 9.436463460132359e-05, "loss": 0.14157943427562714, "step": 524 }, { "epoch": 3.779783393501805, "grad_norm": 0.5732694268226624, "learning_rate": 9.366810654991276e-05, "loss": 0.12228554487228394, "step": 525 }, { "epoch": 3.7870036101083033, "grad_norm": 0.5100772976875305, "learning_rate": 9.297337174425995e-05, "loss": 0.13228444755077362, "step": 526 }, { "epoch": 3.794223826714801, "grad_norm": 0.3461199104785919, "learning_rate": 9.228044190079553e-05, "loss": 0.11099657416343689, "step": 527 }, { "epoch": 3.8014440433212995, "grad_norm": 0.5344772338867188, "learning_rate": 9.158932870551011e-05, "loss": 0.1298564374446869, "step": 528 }, { "epoch": 3.808664259927798, "grad_norm": 0.5982236266136169, "learning_rate": 9.09000438137573e-05, "loss": 0.13814345002174377, "step": 529 }, { "epoch": 3.815884476534296, "grad_norm": 0.540401816368103, "learning_rate": 9.021259885005678e-05, "loss": 0.15752826631069183, "step": 530 }, { "epoch": 3.8231046931407944, "grad_norm": 0.5778837203979492, "learning_rate": 8.952700540789885e-05, "loss": 0.11452727019786835, "step": 531 }, { "epoch": 3.8303249097472927, "grad_norm": 0.5168759226799011, "learning_rate": 8.884327504954857e-05, "loss": 0.18627649545669556, "step": 532 }, { "epoch": 3.8375451263537905, "grad_norm": 0.5951539278030396, "learning_rate": 8.816141930585067e-05, "loss": 0.17812731862068176, "step": 533 }, { "epoch": 3.844765342960289, "grad_norm": 0.46762675046920776, "learning_rate": 8.748144967603538e-05, "loss": 0.1432511955499649, "step": 534 }, { "epoch": 3.851985559566787, "grad_norm": 0.4484635889530182, "learning_rate": 8.680337762752444e-05, "loss": 0.13658685982227325, "step": 535 }, { "epoch": 3.859205776173285, "grad_norm": 0.518476128578186, "learning_rate": 8.612721459573734e-05, "loss": 0.14582251012325287, "step": 536 }, { "epoch": 3.8664259927797833, "grad_norm": 0.5416932106018066, "learning_rate": 8.545297198389896e-05, "loss": 0.15608924627304077, "step": 537 }, { "epoch": 3.8736462093862816, "grad_norm": 0.502341628074646, "learning_rate": 8.478066116284713e-05, "loss": 0.14149951934814453, "step": 538 }, { "epoch": 3.88086642599278, "grad_norm": 0.3645715117454529, "learning_rate": 8.411029347084049e-05, "loss": 0.1299845278263092, "step": 539 }, { "epoch": 3.888086642599278, "grad_norm": 0.4532802700996399, "learning_rate": 8.344188021336783e-05, "loss": 0.11957640945911407, "step": 540 }, { "epoch": 3.895306859205776, "grad_norm": 0.39594075083732605, "learning_rate": 8.277543266295713e-05, "loss": 0.09765089303255081, "step": 541 }, { "epoch": 3.9025270758122743, "grad_norm": 0.4911992847919464, "learning_rate": 8.211096205898529e-05, "loss": 0.13661810755729675, "step": 542 }, { "epoch": 3.9097472924187726, "grad_norm": 0.41524603962898254, "learning_rate": 8.144847960748904e-05, "loss": 0.11461715400218964, "step": 543 }, { "epoch": 3.916967509025271, "grad_norm": 0.6111493110656738, "learning_rate": 8.078799648097555e-05, "loss": 0.14272424578666687, "step": 544 }, { "epoch": 3.9241877256317688, "grad_norm": 0.6367743611335754, "learning_rate": 8.01295238182342e-05, "loss": 0.15256257355213165, "step": 545 }, { "epoch": 3.931407942238267, "grad_norm": 0.5055350065231323, "learning_rate": 7.947307272414874e-05, "loss": 0.12612725794315338, "step": 546 }, { "epoch": 3.9386281588447654, "grad_norm": 0.6354690790176392, "learning_rate": 7.881865426950979e-05, "loss": 0.18444940447807312, "step": 547 }, { "epoch": 3.9458483754512637, "grad_norm": 0.48749735951423645, "learning_rate": 7.816627949082844e-05, "loss": 0.14657062292099, "step": 548 }, { "epoch": 3.953068592057762, "grad_norm": 0.5144006013870239, "learning_rate": 7.751595939015004e-05, "loss": 0.13369342684745789, "step": 549 }, { "epoch": 3.96028880866426, "grad_norm": 0.6086586117744446, "learning_rate": 7.686770493486834e-05, "loss": 0.12459664046764374, "step": 550 }, { "epoch": 3.967509025270758, "grad_norm": 0.3899252712726593, "learning_rate": 7.622152705754109e-05, "loss": 0.09838660806417465, "step": 551 }, { "epoch": 3.9747292418772564, "grad_norm": 0.5426426529884338, "learning_rate": 7.557743665570524e-05, "loss": 0.12174687534570694, "step": 552 }, { "epoch": 3.9819494584837543, "grad_norm": 0.4566539227962494, "learning_rate": 7.49354445916932e-05, "loss": 0.12463852018117905, "step": 553 }, { "epoch": 3.9891696750902526, "grad_norm": 0.4546673595905304, "learning_rate": 7.429556169244984e-05, "loss": 0.145570307970047, "step": 554 }, { "epoch": 3.996389891696751, "grad_norm": 0.4223948121070862, "learning_rate": 7.365779874934985e-05, "loss": 0.1327395737171173, "step": 555 }, { "epoch": 4.0, "grad_norm": 0.9028692245483398, "learning_rate": 7.302216651801546e-05, "loss": 0.1679631471633911, "step": 556 }, { "epoch": 4.007220216606498, "grad_norm": 0.2356356829404831, "learning_rate": 7.23886757181355e-05, "loss": 0.07719021290540695, "step": 557 }, { "epoch": 4.014440433212997, "grad_norm": 0.2547387480735779, "learning_rate": 7.175733703328438e-05, "loss": 0.08970507979393005, "step": 558 }, { "epoch": 4.021660649819495, "grad_norm": 0.2796124815940857, "learning_rate": 7.112816111074175e-05, "loss": 0.08122967183589935, "step": 559 }, { "epoch": 4.028880866425993, "grad_norm": 0.3060125410556793, "learning_rate": 7.050115856131327e-05, "loss": 0.10071869194507599, "step": 560 }, { "epoch": 4.036101083032491, "grad_norm": 0.23872718214988708, "learning_rate": 6.987633995915164e-05, "loss": 0.07773898541927338, "step": 561 }, { "epoch": 4.043321299638989, "grad_norm": 0.2474278211593628, "learning_rate": 6.925371584157776e-05, "loss": 0.08390785753726959, "step": 562 }, { "epoch": 4.050541516245487, "grad_norm": 0.2561244070529938, "learning_rate": 6.863329670890379e-05, "loss": 0.07564981281757355, "step": 563 }, { "epoch": 4.0577617328519855, "grad_norm": 0.23667839169502258, "learning_rate": 6.801509302425553e-05, "loss": 0.06735668331384659, "step": 564 }, { "epoch": 4.064981949458484, "grad_norm": 0.30244675278663635, "learning_rate": 6.739911521339603e-05, "loss": 0.08246147632598877, "step": 565 }, { "epoch": 4.072202166064982, "grad_norm": 0.322549968957901, "learning_rate": 6.678537366454998e-05, "loss": 0.0768466591835022, "step": 566 }, { "epoch": 4.07942238267148, "grad_norm": 0.3082004487514496, "learning_rate": 6.617387872822842e-05, "loss": 0.07058844715356827, "step": 567 }, { "epoch": 4.086642599277979, "grad_norm": 0.4027414917945862, "learning_rate": 6.556464071705386e-05, "loss": 0.06483431905508041, "step": 568 }, { "epoch": 4.093862815884476, "grad_norm": 0.32241883873939514, "learning_rate": 6.495766990558716e-05, "loss": 0.08380893617868423, "step": 569 }, { "epoch": 4.101083032490974, "grad_norm": 0.6206040978431702, "learning_rate": 6.435297653015319e-05, "loss": 0.08805538713932037, "step": 570 }, { "epoch": 4.108303249097473, "grad_norm": 0.37588536739349365, "learning_rate": 6.375057078866916e-05, "loss": 0.06972259283065796, "step": 571 }, { "epoch": 4.115523465703971, "grad_norm": 0.7320848703384399, "learning_rate": 6.315046284047207e-05, "loss": 0.07265684008598328, "step": 572 }, { "epoch": 4.122743682310469, "grad_norm": 0.6254473328590393, "learning_rate": 6.255266280614747e-05, "loss": 0.06212320178747177, "step": 573 }, { "epoch": 4.129963898916968, "grad_norm": 0.5760471820831299, "learning_rate": 6.195718076735894e-05, "loss": 0.08673934638500214, "step": 574 }, { "epoch": 4.137184115523466, "grad_norm": 0.49967148900032043, "learning_rate": 6.136402676667806e-05, "loss": 0.06496398150920868, "step": 575 }, { "epoch": 4.144404332129964, "grad_norm": 0.43019023537635803, "learning_rate": 6.077321080741469e-05, "loss": 0.06604783982038498, "step": 576 }, { "epoch": 4.1516245487364625, "grad_norm": 0.4466577172279358, "learning_rate": 6.0184742853448796e-05, "loss": 0.06710691004991531, "step": 577 }, { "epoch": 4.15884476534296, "grad_norm": 0.5732907056808472, "learning_rate": 5.959863282906208e-05, "loss": 0.09353934228420258, "step": 578 }, { "epoch": 4.166064981949458, "grad_norm": 0.3326954245567322, "learning_rate": 5.901489061877059e-05, "loss": 0.06768821179866791, "step": 579 }, { "epoch": 4.1732851985559565, "grad_norm": 0.5838833451271057, "learning_rate": 5.843352606715819e-05, "loss": 0.09657922387123108, "step": 580 }, { "epoch": 4.180505415162455, "grad_norm": 0.6080812215805054, "learning_rate": 5.7854548978710586e-05, "loss": 0.07862652093172073, "step": 581 }, { "epoch": 4.187725631768953, "grad_norm": 0.35978272557258606, "learning_rate": 5.7277969117649554e-05, "loss": 0.06750591099262238, "step": 582 }, { "epoch": 4.194945848375451, "grad_norm": 0.5055731534957886, "learning_rate": 5.670379620776882e-05, "loss": 0.07750953733921051, "step": 583 }, { "epoch": 4.20216606498195, "grad_norm": 0.3300122618675232, "learning_rate": 5.613203993226981e-05, "loss": 0.05905657261610031, "step": 584 }, { "epoch": 4.209386281588448, "grad_norm": 0.39663630723953247, "learning_rate": 5.5562709933598156e-05, "loss": 0.06651993840932846, "step": 585 }, { "epoch": 4.216606498194946, "grad_norm": 0.2632470726966858, "learning_rate": 5.4995815813281483e-05, "loss": 0.04662226885557175, "step": 586 }, { "epoch": 4.223826714801444, "grad_norm": 0.3943081796169281, "learning_rate": 5.443136713176724e-05, "loss": 0.055592283606529236, "step": 587 }, { "epoch": 4.231046931407942, "grad_norm": 0.3067891001701355, "learning_rate": 5.386937340826139e-05, "loss": 0.06188460439443588, "step": 588 }, { "epoch": 4.23826714801444, "grad_norm": 0.5765032768249512, "learning_rate": 5.330984412056814e-05, "loss": 0.07393093407154083, "step": 589 }, { "epoch": 4.245487364620939, "grad_norm": 0.3043580949306488, "learning_rate": 5.2752788704929944e-05, "loss": 0.04691813141107559, "step": 590 }, { "epoch": 4.252707581227437, "grad_norm": 0.41853103041648865, "learning_rate": 5.2198216555868206e-05, "loss": 0.08564123511314392, "step": 591 }, { "epoch": 4.259927797833935, "grad_norm": 0.366338849067688, "learning_rate": 5.164613702602519e-05, "loss": 0.06465509533882141, "step": 592 }, { "epoch": 4.2671480144404335, "grad_norm": 0.4903438985347748, "learning_rate": 5.1096559426005994e-05, "loss": 0.0653722956776619, "step": 593 }, { "epoch": 4.274368231046932, "grad_norm": 0.3799297511577606, "learning_rate": 5.054949302422178e-05, "loss": 0.06573019176721573, "step": 594 }, { "epoch": 4.28158844765343, "grad_norm": 0.3962422013282776, "learning_rate": 5.000494704673328e-05, "loss": 0.06330323964357376, "step": 595 }, { "epoch": 4.2888086642599275, "grad_norm": 0.40771204233169556, "learning_rate": 4.946293067709511e-05, "loss": 0.0744054913520813, "step": 596 }, { "epoch": 4.296028880866426, "grad_norm": 0.358303040266037, "learning_rate": 4.892345305620123e-05, "loss": 0.06824557483196259, "step": 597 }, { "epoch": 4.303249097472924, "grad_norm": 0.5681015253067017, "learning_rate": 4.83865232821306e-05, "loss": 0.07922443747520447, "step": 598 }, { "epoch": 4.310469314079422, "grad_norm": 0.40545016527175903, "learning_rate": 4.785215040999354e-05, "loss": 0.06584658473730087, "step": 599 }, { "epoch": 4.317689530685921, "grad_norm": 0.4532371759414673, "learning_rate": 4.73203434517794e-05, "loss": 0.06798739731311798, "step": 600 }, { "epoch": 4.324909747292419, "grad_norm": 0.34823986887931824, "learning_rate": 4.679111137620442e-05, "loss": 0.060036975890398026, "step": 601 }, { "epoch": 4.332129963898917, "grad_norm": 0.7183476686477661, "learning_rate": 4.626446310856027e-05, "loss": 0.09737586230039597, "step": 602 }, { "epoch": 4.3393501805054155, "grad_norm": 0.3646744191646576, "learning_rate": 4.574040753056385e-05, "loss": 0.07068562507629395, "step": 603 }, { "epoch": 4.346570397111913, "grad_norm": 0.7281261682510376, "learning_rate": 4.521895348020744e-05, "loss": 0.0830075815320015, "step": 604 }, { "epoch": 4.353790613718411, "grad_norm": 0.3349117338657379, "learning_rate": 4.470010975160932e-05, "loss": 0.057703591883182526, "step": 605 }, { "epoch": 4.3610108303249095, "grad_norm": 0.4553733766078949, "learning_rate": 4.418388509486591e-05, "loss": 0.07166286557912827, "step": 606 }, { "epoch": 4.368231046931408, "grad_norm": 0.44004717469215393, "learning_rate": 4.3670288215904044e-05, "loss": 0.06402762979269028, "step": 607 }, { "epoch": 4.375451263537906, "grad_norm": 0.4734828472137451, "learning_rate": 4.315932777633391e-05, "loss": 0.06359273195266724, "step": 608 }, { "epoch": 4.382671480144404, "grad_norm": 0.42865970730781555, "learning_rate": 4.265101239330336e-05, "loss": 0.05663272365927696, "step": 609 }, { "epoch": 4.389891696750903, "grad_norm": 0.3049170970916748, "learning_rate": 4.214535063935236e-05, "loss": 0.05514144152402878, "step": 610 }, { "epoch": 4.397111913357401, "grad_norm": 0.5843676328659058, "learning_rate": 4.1642351042268413e-05, "loss": 0.0814075767993927, "step": 611 }, { "epoch": 4.404332129963899, "grad_norm": 0.45971769094467163, "learning_rate": 4.11420220849428e-05, "loss": 0.08407925814390182, "step": 612 }, { "epoch": 4.411552346570397, "grad_norm": 0.3730124235153198, "learning_rate": 4.064437220522759e-05, "loss": 0.08016012609004974, "step": 613 }, { "epoch": 4.418772563176895, "grad_norm": 0.47459709644317627, "learning_rate": 4.0149409795793116e-05, "loss": 0.06777369976043701, "step": 614 }, { "epoch": 4.425992779783393, "grad_norm": 0.43131333589553833, "learning_rate": 3.965714320398668e-05, "loss": 0.06275366246700287, "step": 615 }, { "epoch": 4.433212996389892, "grad_norm": 0.5383766293525696, "learning_rate": 3.9167580731691644e-05, "loss": 0.07567012310028076, "step": 616 }, { "epoch": 4.44043321299639, "grad_norm": 0.5049149990081787, "learning_rate": 3.868073063518749e-05, "loss": 0.07030000537633896, "step": 617 }, { "epoch": 4.447653429602888, "grad_norm": 0.4588138461112976, "learning_rate": 3.819660112501053e-05, "loss": 0.06413187086582184, "step": 618 }, { "epoch": 4.4548736462093865, "grad_norm": 0.5270079374313354, "learning_rate": 3.771520036581535e-05, "loss": 0.08776699006557465, "step": 619 }, { "epoch": 4.462093862815885, "grad_norm": 0.4434908330440521, "learning_rate": 3.723653647623735e-05, "loss": 0.06680673360824585, "step": 620 }, { "epoch": 4.469314079422382, "grad_norm": 0.40275996923446655, "learning_rate": 3.6760617528755684e-05, "loss": 0.06380727887153625, "step": 621 }, { "epoch": 4.4765342960288805, "grad_norm": 0.33545786142349243, "learning_rate": 3.6287451549557016e-05, "loss": 0.06220482289791107, "step": 622 }, { "epoch": 4.483754512635379, "grad_norm": 0.358530193567276, "learning_rate": 3.5817046518400385e-05, "loss": 0.07663938403129578, "step": 623 }, { "epoch": 4.490974729241877, "grad_norm": 0.4760512411594391, "learning_rate": 3.534941036848258e-05, "loss": 0.07023464143276215, "step": 624 }, { "epoch": 4.498194945848375, "grad_norm": 0.5976642966270447, "learning_rate": 3.4884550986304074e-05, "loss": 0.07387922704219818, "step": 625 }, { "epoch": 4.505415162454874, "grad_norm": 0.4269956052303314, "learning_rate": 3.442247621153643e-05, "loss": 0.06992574781179428, "step": 626 }, { "epoch": 4.512635379061372, "grad_norm": 0.5576417446136475, "learning_rate": 3.3963193836889905e-05, "loss": 0.06586994230747223, "step": 627 }, { "epoch": 4.51985559566787, "grad_norm": 0.27365460991859436, "learning_rate": 3.3506711607981824e-05, "loss": 0.05236521363258362, "step": 628 }, { "epoch": 4.527075812274369, "grad_norm": 0.32343900203704834, "learning_rate": 3.305303722320636e-05, "loss": 0.0833154246211052, "step": 629 }, { "epoch": 4.534296028880867, "grad_norm": 0.6516956686973572, "learning_rate": 3.260217833360446e-05, "loss": 0.08919483423233032, "step": 630 }, { "epoch": 4.541516245487364, "grad_norm": 0.3491802513599396, "learning_rate": 3.215414254273468e-05, "loss": 0.06530842185020447, "step": 631 }, { "epoch": 4.548736462093863, "grad_norm": 0.4860411286354065, "learning_rate": 3.170893740654533e-05, "loss": 0.08303852379322052, "step": 632 }, { "epoch": 4.555956678700361, "grad_norm": 0.4222336709499359, "learning_rate": 3.126657043324677e-05, "loss": 0.0689733549952507, "step": 633 }, { "epoch": 4.563176895306859, "grad_norm": 0.6741357445716858, "learning_rate": 3.082704908318474e-05, "loss": 0.07350336015224457, "step": 634 }, { "epoch": 4.5703971119133575, "grad_norm": 0.3562377691268921, "learning_rate": 3.039038076871481e-05, "loss": 0.06916918605566025, "step": 635 }, { "epoch": 4.577617328519856, "grad_norm": 0.5075894594192505, "learning_rate": 2.9956572854077202e-05, "loss": 0.05884088948369026, "step": 636 }, { "epoch": 4.584837545126354, "grad_norm": 0.49467241764068604, "learning_rate": 2.9525632655272482e-05, "loss": 0.05398339778184891, "step": 637 }, { "epoch": 4.5920577617328515, "grad_norm": 0.30880579352378845, "learning_rate": 2.9097567439938478e-05, "loss": 0.06401652097702026, "step": 638 }, { "epoch": 4.59927797833935, "grad_norm": 0.3337753713130951, "learning_rate": 2.8672384427227484e-05, "loss": 0.06975904852151871, "step": 639 }, { "epoch": 4.606498194945848, "grad_norm": 1.321722149848938, "learning_rate": 2.8250090787684437e-05, "loss": 0.07469654083251953, "step": 640 }, { "epoch": 4.613718411552346, "grad_norm": 0.3456924855709076, "learning_rate": 2.783069364312647e-05, "loss": 0.061028316617012024, "step": 641 }, { "epoch": 4.620938628158845, "grad_norm": 0.9587940573692322, "learning_rate": 2.7414200066522067e-05, "loss": 0.08222727477550507, "step": 642 }, { "epoch": 4.628158844765343, "grad_norm": 0.34773939847946167, "learning_rate": 2.7000617081872402e-05, "loss": 0.06201966106891632, "step": 643 }, { "epoch": 4.635379061371841, "grad_norm": 0.4925590455532074, "learning_rate": 2.6589951664092706e-05, "loss": 0.06247459352016449, "step": 644 }, { "epoch": 4.64259927797834, "grad_norm": 0.34186214208602905, "learning_rate": 2.618221073889433e-05, "loss": 0.054603368043899536, "step": 645 }, { "epoch": 4.649819494584838, "grad_norm": 0.3584718406200409, "learning_rate": 2.5777401182668446e-05, "loss": 0.05636471137404442, "step": 646 }, { "epoch": 4.657039711191336, "grad_norm": 0.33370015025138855, "learning_rate": 2.5375529822369815e-05, "loss": 0.06488262116909027, "step": 647 }, { "epoch": 4.664259927797834, "grad_norm": 0.428996205329895, "learning_rate": 2.497660343540147e-05, "loss": 0.08126433193683624, "step": 648 }, { "epoch": 4.671480144404332, "grad_norm": 0.306087464094162, "learning_rate": 2.4580628749500844e-05, "loss": 0.056748297065496445, "step": 649 }, { "epoch": 4.67870036101083, "grad_norm": 0.45556285977363586, "learning_rate": 2.4187612442626016e-05, "loss": 0.08052687346935272, "step": 650 }, { "epoch": 4.6859205776173285, "grad_norm": 0.24860779941082, "learning_rate": 2.3797561142843105e-05, "loss": 0.06610046327114105, "step": 651 }, { "epoch": 4.693140794223827, "grad_norm": 0.452534556388855, "learning_rate": 2.3410481428214603e-05, "loss": 0.056214869022369385, "step": 652 }, { "epoch": 4.700361010830325, "grad_norm": 0.38668516278266907, "learning_rate": 2.302637982668843e-05, "loss": 0.06206169351935387, "step": 653 }, { "epoch": 4.707581227436823, "grad_norm": 0.2503337562084198, "learning_rate": 2.2645262815987622e-05, "loss": 0.06331133097410202, "step": 654 }, { "epoch": 4.714801444043322, "grad_norm": 0.40086954832077026, "learning_rate": 2.2267136823501455e-05, "loss": 0.06614114344120026, "step": 655 }, { "epoch": 4.722021660649819, "grad_norm": 0.3914971351623535, "learning_rate": 2.1892008226176762e-05, "loss": 0.06879283487796783, "step": 656 }, { "epoch": 4.729241877256317, "grad_norm": 0.3335961699485779, "learning_rate": 2.1519883350410443e-05, "loss": 0.07069812715053558, "step": 657 }, { "epoch": 4.736462093862816, "grad_norm": 0.45657894015312195, "learning_rate": 2.115076847194293e-05, "loss": 0.07669930905103683, "step": 658 }, { "epoch": 4.743682310469314, "grad_norm": 0.28149735927581787, "learning_rate": 2.0784669815752044e-05, "loss": 0.049429114907979965, "step": 659 }, { "epoch": 4.750902527075812, "grad_norm": 0.3293699026107788, "learning_rate": 2.0421593555948393e-05, "loss": 0.05612622946500778, "step": 660 }, { "epoch": 4.758122743682311, "grad_norm": 0.4966195821762085, "learning_rate": 2.0061545815670967e-05, "loss": 0.0798168033361435, "step": 661 }, { "epoch": 4.765342960288809, "grad_norm": 0.40849727392196655, "learning_rate": 1.970453266698391e-05, "loss": 0.06668490171432495, "step": 662 }, { "epoch": 4.772563176895307, "grad_norm": 0.37348175048828125, "learning_rate": 1.935056013077423e-05, "loss": 0.06494098156690598, "step": 663 }, { "epoch": 4.7797833935018055, "grad_norm": 0.29984328150749207, "learning_rate": 1.8999634176650205e-05, "loss": 0.051809899508953094, "step": 664 }, { "epoch": 4.787003610108303, "grad_norm": 0.366937518119812, "learning_rate": 1.8651760722840672e-05, "loss": 0.06859136372804642, "step": 665 }, { "epoch": 4.794223826714801, "grad_norm": 0.46067899465560913, "learning_rate": 1.830694563609525e-05, "loss": 0.06466175615787506, "step": 666 }, { "epoch": 4.8014440433212995, "grad_norm": 0.45101597905158997, "learning_rate": 1.7965194731585422e-05, "loss": 0.062113627791404724, "step": 667 }, { "epoch": 4.808664259927798, "grad_norm": 0.39610734581947327, "learning_rate": 1.762651377280633e-05, "loss": 0.06937667727470398, "step": 668 }, { "epoch": 4.815884476534296, "grad_norm": 0.406488835811615, "learning_rate": 1.7290908471479805e-05, "loss": 0.06851080060005188, "step": 669 }, { "epoch": 4.823104693140794, "grad_norm": 0.4525218904018402, "learning_rate": 1.6958384487457923e-05, "loss": 0.05941899120807648, "step": 670 }, { "epoch": 4.830324909747293, "grad_norm": 0.3088747560977936, "learning_rate": 1.6628947428627438e-05, "loss": 0.05612848699092865, "step": 671 }, { "epoch": 4.837545126353791, "grad_norm": 0.330507755279541, "learning_rate": 1.6302602850815397e-05, "loss": 0.06643390655517578, "step": 672 }, { "epoch": 4.844765342960288, "grad_norm": 0.3919360041618347, "learning_rate": 1.59793562576954e-05, "loss": 0.06949331611394882, "step": 673 }, { "epoch": 4.851985559566787, "grad_norm": 0.43942415714263916, "learning_rate": 1.5659213100694626e-05, "loss": 0.06495010107755661, "step": 674 }, { "epoch": 4.859205776173285, "grad_norm": 0.3314034640789032, "learning_rate": 1.5342178778902116e-05, "loss": 0.0641111359000206, "step": 675 }, { "epoch": 4.866425992779783, "grad_norm": 0.6920318007469177, "learning_rate": 1.5028258638977677e-05, "loss": 0.08164434880018234, "step": 676 }, { "epoch": 4.873646209386282, "grad_norm": 0.4740530550479889, "learning_rate": 1.4717457975061455e-05, "loss": 0.07530295848846436, "step": 677 }, { "epoch": 4.88086642599278, "grad_norm": 0.4190375506877899, "learning_rate": 1.4409782028685104e-05, "loss": 0.07053963840007782, "step": 678 }, { "epoch": 4.888086642599278, "grad_norm": 0.3520694375038147, "learning_rate": 1.4105235988682896e-05, "loss": 0.06393534690141678, "step": 679 }, { "epoch": 4.8953068592057765, "grad_norm": 0.40102067589759827, "learning_rate": 1.3803824991104663e-05, "loss": 0.07699807733297348, "step": 680 }, { "epoch": 4.902527075812275, "grad_norm": 0.4686407148838043, "learning_rate": 1.350555411912886e-05, "loss": 0.06640764325857162, "step": 681 }, { "epoch": 4.909747292418773, "grad_norm": 0.31581881642341614, "learning_rate": 1.3210428402976971e-05, "loss": 0.06866060942411423, "step": 682 }, { "epoch": 4.9169675090252705, "grad_norm": 0.39863917231559753, "learning_rate": 1.2918452819828663e-05, "loss": 0.05927126109600067, "step": 683 }, { "epoch": 4.924187725631769, "grad_norm": 0.7777624130249023, "learning_rate": 1.2629632293737903e-05, "loss": 0.07142484188079834, "step": 684 }, { "epoch": 4.931407942238267, "grad_norm": 0.3229425251483917, "learning_rate": 1.2343971695549727e-05, "loss": 0.06009211018681526, "step": 685 }, { "epoch": 4.938628158844765, "grad_norm": 0.5577942132949829, "learning_rate": 1.2061475842818338e-05, "loss": 0.08745917677879333, "step": 686 }, { "epoch": 4.945848375451264, "grad_norm": 0.4703651964664459, "learning_rate": 1.1782149499725714e-05, "loss": 0.059495482593774796, "step": 687 }, { "epoch": 4.953068592057762, "grad_norm": 0.4242652356624603, "learning_rate": 1.150599737700122e-05, "loss": 0.05950907990336418, "step": 688 }, { "epoch": 4.96028880866426, "grad_norm": 0.454228013753891, "learning_rate": 1.1233024131842374e-05, "loss": 0.08122330904006958, "step": 689 }, { "epoch": 4.967509025270758, "grad_norm": 0.28975188732147217, "learning_rate": 1.0963234367836106e-05, "loss": 0.05605383217334747, "step": 690 }, { "epoch": 4.974729241877256, "grad_norm": 0.38855794072151184, "learning_rate": 1.0696632634881099e-05, "loss": 0.059195615351200104, "step": 691 }, { "epoch": 4.981949458483754, "grad_norm": 0.46655526757240295, "learning_rate": 1.043322342911126e-05, "loss": 0.07140585780143738, "step": 692 }, { "epoch": 4.9891696750902526, "grad_norm": 0.4152084290981293, "learning_rate": 1.017301119281977e-05, "loss": 0.07906543463468552, "step": 693 }, { "epoch": 4.996389891696751, "grad_norm": 0.39406299591064453, "learning_rate": 9.916000314384044e-06, "loss": 0.0740804374217987, "step": 694 }, { "epoch": 5.0, "grad_norm": 0.4732819199562073, "learning_rate": 9.662195128191997e-06, "loss": 0.06949488818645477, "step": 695 }, { "epoch": 5.007220216606498, "grad_norm": 0.2503197491168976, "learning_rate": 9.411599914568702e-06, "loss": 0.048096612095832825, "step": 696 }, { "epoch": 5.014440433212997, "grad_norm": 0.20594151318073273, "learning_rate": 9.164218899704336e-06, "loss": 0.044573940336704254, "step": 697 }, { "epoch": 5.021660649819495, "grad_norm": 0.3241271674633026, "learning_rate": 8.920056255582943e-06, "loss": 0.05117207020521164, "step": 698 }, { "epoch": 5.028880866425993, "grad_norm": 0.18174150586128235, "learning_rate": 8.679116099911854e-06, "loss": 0.04179506003856659, "step": 699 }, { "epoch": 5.036101083032491, "grad_norm": 0.1984250396490097, "learning_rate": 8.441402496052496e-06, "loss": 0.053596131503582, "step": 700 }, { "epoch": 5.043321299638989, "grad_norm": 0.21945425868034363, "learning_rate": 8.206919452951778e-06, "loss": 0.042779602110385895, "step": 701 }, { "epoch": 5.050541516245487, "grad_norm": 0.260366827249527, "learning_rate": 7.975670925074386e-06, "loss": 0.06215063855051994, "step": 702 }, { "epoch": 5.0577617328519855, "grad_norm": 0.14613628387451172, "learning_rate": 7.747660812336222e-06, "loss": 0.04405611380934715, "step": 703 }, { "epoch": 5.064981949458484, "grad_norm": 0.14823344349861145, "learning_rate": 7.52289296003863e-06, "loss": 0.040258534252643585, "step": 704 }, { "epoch": 5.072202166064982, "grad_norm": 0.11375842988491058, "learning_rate": 7.301371158803383e-06, "loss": 0.037275541573762894, "step": 705 }, { "epoch": 5.07942238267148, "grad_norm": 0.18186606466770172, "learning_rate": 7.083099144508976e-06, "loss": 0.04733014106750488, "step": 706 }, { "epoch": 5.086642599277979, "grad_norm": 0.21977707743644714, "learning_rate": 6.8680805982275e-06, "loss": 0.04784165322780609, "step": 707 }, { "epoch": 5.093862815884476, "grad_norm": 0.2574014663696289, "learning_rate": 6.656319146162515e-06, "loss": 0.05287627875804901, "step": 708 }, { "epoch": 5.101083032490974, "grad_norm": 0.24309682846069336, "learning_rate": 6.447818359588054e-06, "loss": 0.05101453512907028, "step": 709 }, { "epoch": 5.108303249097473, "grad_norm": 0.18539798259735107, "learning_rate": 6.242581754788268e-06, "loss": 0.04852500185370445, "step": 710 }, { "epoch": 5.115523465703971, "grad_norm": 0.2783971130847931, "learning_rate": 6.040612792998123e-06, "loss": 0.04606611281633377, "step": 711 }, { "epoch": 5.122743682310469, "grad_norm": 0.2432968020439148, "learning_rate": 5.841914880345111e-06, "loss": 0.05015704780817032, "step": 712 }, { "epoch": 5.129963898916968, "grad_norm": 0.21053971350193024, "learning_rate": 5.64649136779174e-06, "loss": 0.04552324116230011, "step": 713 }, { "epoch": 5.137184115523466, "grad_norm": 0.2822020947933197, "learning_rate": 5.454345551079043e-06, "loss": 0.04811366647481918, "step": 714 }, { "epoch": 5.144404332129964, "grad_norm": 0.3009110689163208, "learning_rate": 5.265480670671053e-06, "loss": 0.057499293237924576, "step": 715 }, { "epoch": 5.1516245487364625, "grad_norm": 0.2482829988002777, "learning_rate": 5.07989991169997e-06, "loss": 0.04851806163787842, "step": 716 }, { "epoch": 5.15884476534296, "grad_norm": 0.4322669506072998, "learning_rate": 4.8976064039126805e-06, "loss": 0.040396105498075485, "step": 717 }, { "epoch": 5.166064981949458, "grad_norm": 0.22342568635940552, "learning_rate": 4.718603221617834e-06, "loss": 0.047426290810108185, "step": 718 }, { "epoch": 5.1732851985559565, "grad_norm": 0.3421519994735718, "learning_rate": 4.542893383634028e-06, "loss": 0.050219837576150894, "step": 719 }, { "epoch": 5.180505415162455, "grad_norm": 0.2810814678668976, "learning_rate": 4.370479853238863e-06, "loss": 0.039352793246507645, "step": 720 }, { "epoch": 5.187725631768953, "grad_norm": 0.16700661182403564, "learning_rate": 4.201365538119096e-06, "loss": 0.03363039717078209, "step": 721 }, { "epoch": 5.194945848375451, "grad_norm": 0.21759076416492462, "learning_rate": 4.035553290321414e-06, "loss": 0.04147165268659592, "step": 722 }, { "epoch": 5.20216606498195, "grad_norm": 0.13985103368759155, "learning_rate": 3.873045906204453e-06, "loss": 0.043792638927698135, "step": 723 }, { "epoch": 5.209386281588448, "grad_norm": 0.2625335156917572, "learning_rate": 3.7138461263916513e-06, "loss": 0.05631346255540848, "step": 724 }, { "epoch": 5.216606498194946, "grad_norm": 0.16157718002796173, "learning_rate": 3.5579566357249793e-06, "loss": 0.03788129985332489, "step": 725 }, { "epoch": 5.223826714801444, "grad_norm": 0.19516326487064362, "learning_rate": 3.405380063219643e-06, "loss": 0.044141985476017, "step": 726 }, { "epoch": 5.231046931407942, "grad_norm": 0.15679222345352173, "learning_rate": 3.256118982019851e-06, "loss": 0.04181442782282829, "step": 727 }, { "epoch": 5.23826714801444, "grad_norm": 0.17116223275661469, "learning_rate": 3.1101759093552728e-06, "loss": 0.036379870027303696, "step": 728 }, { "epoch": 5.245487364620939, "grad_norm": 0.13156861066818237, "learning_rate": 2.9675533064986936e-06, "loss": 0.035947442054748535, "step": 729 }, { "epoch": 5.252707581227437, "grad_norm": 0.31538212299346924, "learning_rate": 2.828253578724538e-06, "loss": 0.04313802346587181, "step": 730 }, { "epoch": 5.259927797833935, "grad_norm": 0.16643384099006653, "learning_rate": 2.6922790752681004e-06, "loss": 0.03961348533630371, "step": 731 }, { "epoch": 5.2671480144404335, "grad_norm": 0.2200719565153122, "learning_rate": 2.5596320892862013e-06, "loss": 0.051561444997787476, "step": 732 }, { "epoch": 5.274368231046932, "grad_norm": 0.3769620954990387, "learning_rate": 2.4303148578183497e-06, "loss": 0.07501974701881409, "step": 733 }, { "epoch": 5.28158844765343, "grad_norm": 0.20830948650836945, "learning_rate": 2.3043295617489746e-06, "loss": 0.04220421612262726, "step": 734 }, { "epoch": 5.2888086642599275, "grad_norm": 0.19785495102405548, "learning_rate": 2.1816783257708083e-06, "loss": 0.03773656487464905, "step": 735 }, { "epoch": 5.296028880866426, "grad_norm": 0.2532854974269867, "learning_rate": 2.0623632183489396e-06, "loss": 0.04975757375359535, "step": 736 }, { "epoch": 5.303249097472924, "grad_norm": 0.21113696694374084, "learning_rate": 1.9463862516859276e-06, "loss": 0.04151391610503197, "step": 737 }, { "epoch": 5.310469314079422, "grad_norm": 0.41521739959716797, "learning_rate": 1.8337493816879436e-06, "loss": 0.04442136362195015, "step": 738 }, { "epoch": 5.317689530685921, "grad_norm": 0.21844181418418884, "learning_rate": 1.724454507931683e-06, "loss": 0.04176602512598038, "step": 739 }, { "epoch": 5.324909747292419, "grad_norm": 0.17051415145397186, "learning_rate": 1.6185034736324156e-06, "loss": 0.03697787970304489, "step": 740 }, { "epoch": 5.332129963898917, "grad_norm": 0.18966346979141235, "learning_rate": 1.515898065612853e-06, "loss": 0.038028594106435776, "step": 741 }, { "epoch": 5.3393501805054155, "grad_norm": 0.17065943777561188, "learning_rate": 1.4166400142730407e-06, "loss": 0.035803914070129395, "step": 742 }, { "epoch": 5.346570397111913, "grad_norm": 0.31374311447143555, "learning_rate": 1.3207309935611367e-06, "loss": 0.056527022272348404, "step": 743 }, { "epoch": 5.353790613718411, "grad_norm": 0.1854705810546875, "learning_rate": 1.2281726209452781e-06, "loss": 0.043832339346408844, "step": 744 }, { "epoch": 5.3610108303249095, "grad_norm": 0.14780643582344055, "learning_rate": 1.138966457386137e-06, "loss": 0.03484298288822174, "step": 745 }, { "epoch": 5.368231046931408, "grad_norm": 0.20580680668354034, "learning_rate": 1.0531140073107626e-06, "loss": 0.042070530354976654, "step": 746 }, { "epoch": 5.375451263537906, "grad_norm": 0.18943031132221222, "learning_rate": 9.70616718587114e-07, "loss": 0.04158391058444977, "step": 747 }, { "epoch": 5.382671480144404, "grad_norm": 0.1829371601343155, "learning_rate": 8.914759824996788e-07, "loss": 0.03915674239397049, "step": 748 }, { "epoch": 5.389891696750903, "grad_norm": 0.23177319765090942, "learning_rate": 8.15693133725981e-07, "loss": 0.03685947507619858, "step": 749 }, { "epoch": 5.397111913357401, "grad_norm": 0.2238905131816864, "learning_rate": 7.432694503141102e-07, "loss": 0.04124440625309944, "step": 750 }, { "epoch": 5.404332129963899, "grad_norm": 0.19328288733959198, "learning_rate": 6.742061536611166e-07, "loss": 0.04230578616261482, "step": 751 }, { "epoch": 5.411552346570397, "grad_norm": 0.2329723834991455, "learning_rate": 6.085044084924718e-07, "loss": 0.04548173397779465, "step": 752 }, { "epoch": 5.418772563176895, "grad_norm": 0.17935019731521606, "learning_rate": 5.461653228423957e-07, "loss": 0.04122517257928848, "step": 753 }, { "epoch": 5.425992779783393, "grad_norm": 0.19964119791984558, "learning_rate": 4.871899480351605e-07, "loss": 0.03701728954911232, "step": 754 }, { "epoch": 5.433212996389892, "grad_norm": 0.17978627979755402, "learning_rate": 4.315792786673489e-07, "loss": 0.04114942252635956, "step": 755 }, { "epoch": 5.44043321299639, "grad_norm": 0.3241960406303406, "learning_rate": 3.7933425259117914e-07, "loss": 0.0526733361184597, "step": 756 }, { "epoch": 5.447653429602888, "grad_norm": 0.22060273587703705, "learning_rate": 3.3045575089853954e-07, "loss": 0.04859931021928787, "step": 757 }, { "epoch": 5.4548736462093865, "grad_norm": 0.22234350442886353, "learning_rate": 2.849445979062226e-07, "loss": 0.0446445494890213, "step": 758 }, { "epoch": 5.462093862815885, "grad_norm": 0.3016712963581085, "learning_rate": 2.428015611420253e-07, "loss": 0.0628553107380867, "step": 759 }, { "epoch": 5.469314079422382, "grad_norm": 0.21905440092086792, "learning_rate": 2.040273513317148e-07, "loss": 0.04386765509843826, "step": 760 }, { "epoch": 5.4765342960288805, "grad_norm": 0.24936382472515106, "learning_rate": 1.686226223872378e-07, "loss": 0.0364193394780159, "step": 761 }, { "epoch": 5.483754512635379, "grad_norm": 0.21109315752983093, "learning_rate": 1.3658797139541878e-07, "loss": 0.036623530089855194, "step": 762 }, { "epoch": 5.490974729241877, "grad_norm": 0.3846798241138458, "learning_rate": 1.0792393860814543e-07, "loss": 0.039832159876823425, "step": 763 }, { "epoch": 5.498194945848375, "grad_norm": 0.3015185296535492, "learning_rate": 8.263100743310937e-08, "loss": 0.054986074566841125, "step": 764 }, { "epoch": 5.505415162454874, "grad_norm": 0.22440600395202637, "learning_rate": 6.070960442567941e-08, "loss": 0.04008050635457039, "step": 765 }, { "epoch": 5.512635379061372, "grad_norm": 0.2297099381685257, "learning_rate": 4.216009928172948e-08, "loss": 0.05160742253065109, "step": 766 }, { "epoch": 5.51985559566787, "grad_norm": 0.2735305726528168, "learning_rate": 2.698280483142135e-08, "loss": 0.047364577651023865, "step": 767 }, { "epoch": 5.527075812274369, "grad_norm": 0.18135972321033478, "learning_rate": 1.517797703387558e-08, "loss": 0.03480882942676544, "step": 768 }, { "epoch": 5.534296028880867, "grad_norm": 0.28884103894233704, "learning_rate": 6.745814972908271e-09, "loss": 0.049272023141384125, "step": 769 }, { "epoch": 5.541516245487364, "grad_norm": 0.22770103812217712, "learning_rate": 1.6864608536115711e-09, "loss": 0.04704172536730766, "step": 770 }, { "epoch": 5.541516245487364, "eval_loss": 0.7697137594223022, "eval_runtime": 48.9364, "eval_samples_per_second": 2.003, "eval_steps_per_second": 0.511, "step": 770 } ], "logging_steps": 1, "max_steps": 770, "num_input_tokens_seen": 0, "num_train_epochs": 6, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.51023270933007e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }