diff --git "a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/trainer_state.json" "b/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/trainer_state.json" deleted file mode 100644--- "a/HVU_QA/t5-viet-qg-finetuned/checkpoint-10000/trainer_state.json" +++ /dev/null @@ -1,7034 +0,0 @@ -{ - "best_global_step": null, - "best_metric": null, - "best_model_checkpoint": null, - "epoch": 0.15598677232170713, - "eval_steps": 500, - "global_step": 10000, - "is_hyper_param_search": false, - "is_local_process_zero": true, - "is_world_process_zero": true, - "log_history": [ - { - "epoch": 0.00015598677232170712, - "grad_norm": 3.216968059539795, - "learning_rate": 0.0001999906407936607, - "loss": 4.7744, - "step": 10 - }, - { - "epoch": 0.00031197354464341425, - "grad_norm": 9.366360664367676, - "learning_rate": 0.00019998024167550594, - "loss": 0.9639, - "step": 20 - }, - { - "epoch": 0.00046796031696512137, - "grad_norm": 4.382473945617676, - "learning_rate": 0.00019996984255735113, - "loss": 0.6338, - "step": 30 - }, - { - "epoch": 0.0006239470892868285, - "grad_norm": 14.415914535522461, - "learning_rate": 0.00019995944343919637, - "loss": 0.4962, - "step": 40 - }, - { - "epoch": 0.0007799338616085356, - "grad_norm": 7.440260410308838, - "learning_rate": 0.00019994904432104157, - "loss": 0.4619, - "step": 50 - }, - { - "epoch": 0.0009359206339302427, - "grad_norm": 27.18022918701172, - "learning_rate": 0.00019993864520288681, - "loss": 0.5335, - "step": 60 - }, - { - "epoch": 0.0010919074062519499, - "grad_norm": 7.954453468322754, - "learning_rate": 0.000199928246084732, - "loss": 0.8085, - "step": 70 - }, - { - "epoch": 0.001247894178573657, - "grad_norm": 4.762608528137207, - "learning_rate": 0.00019991784696657725, - "loss": 0.6333, - "step": 80 - }, - { - "epoch": 0.0014038809508953641, - "grad_norm": 9.577717781066895, - "learning_rate": 0.00019990744784842245, - "loss": 0.4558, - "step": 90 - }, - { - "epoch": 0.0015598677232170712, - "grad_norm": 7.586806297302246, - "learning_rate": 0.0001998970487302677, - "loss": 0.3227, - "step": 100 - }, - { - "epoch": 0.0017158544955387784, - "grad_norm": 4.262249946594238, - "learning_rate": 0.00019988664961211288, - "loss": 0.3261, - "step": 110 - }, - { - "epoch": 0.0018718412678604855, - "grad_norm": 6.639514923095703, - "learning_rate": 0.00019987625049395813, - "loss": 0.6788, - "step": 120 - }, - { - "epoch": 0.0020278280401821924, - "grad_norm": 3.7448127269744873, - "learning_rate": 0.00019986585137580332, - "loss": 0.4224, - "step": 130 - }, - { - "epoch": 0.0021838148125038997, - "grad_norm": 13.205582618713379, - "learning_rate": 0.00019985545225764857, - "loss": 0.4074, - "step": 140 - }, - { - "epoch": 0.0023398015848256066, - "grad_norm": 5.55623722076416, - "learning_rate": 0.00019984505313949376, - "loss": 0.3408, - "step": 150 - }, - { - "epoch": 0.002495788357147314, - "grad_norm": 10.253294944763184, - "learning_rate": 0.000199834654021339, - "loss": 0.4513, - "step": 160 - }, - { - "epoch": 0.002651775129469021, - "grad_norm": 12.778948783874512, - "learning_rate": 0.0001998242549031842, - "loss": 0.4287, - "step": 170 - }, - { - "epoch": 0.0028077619017907282, - "grad_norm": 8.961956977844238, - "learning_rate": 0.00019981385578502945, - "loss": 0.6746, - "step": 180 - }, - { - "epoch": 0.002963748674112435, - "grad_norm": 6.611206531524658, - "learning_rate": 0.00019980345666687467, - "loss": 0.4106, - "step": 190 - }, - { - "epoch": 0.0031197354464341425, - "grad_norm": 4.179671764373779, - "learning_rate": 0.0001997930575487199, - "loss": 0.4234, - "step": 200 - }, - { - "epoch": 0.0032757222187558494, - "grad_norm": 7.957104206085205, - "learning_rate": 0.0001997826584305651, - "loss": 0.4062, - "step": 210 - }, - { - "epoch": 0.0034317089910775567, - "grad_norm": 10.04617977142334, - "learning_rate": 0.00019977225931241033, - "loss": 0.4792, - "step": 220 - }, - { - "epoch": 0.0035876957633992636, - "grad_norm": 5.119971752166748, - "learning_rate": 0.00019976186019425555, - "loss": 0.3385, - "step": 230 - }, - { - "epoch": 0.003743682535720971, - "grad_norm": 6.039366722106934, - "learning_rate": 0.00019975146107610077, - "loss": 0.4406, - "step": 240 - }, - { - "epoch": 0.003899669308042678, - "grad_norm": 9.345914840698242, - "learning_rate": 0.000199741061957946, - "loss": 0.5163, - "step": 250 - }, - { - "epoch": 0.004055656080364385, - "grad_norm": 3.707460641860962, - "learning_rate": 0.0001997306628397912, - "loss": 0.2971, - "step": 260 - }, - { - "epoch": 0.0042116428526860926, - "grad_norm": 11.500152587890625, - "learning_rate": 0.00019972026372163643, - "loss": 0.3659, - "step": 270 - }, - { - "epoch": 0.0043676296250077995, - "grad_norm": 0.7893226742744446, - "learning_rate": 0.00019970986460348165, - "loss": 0.5052, - "step": 280 - }, - { - "epoch": 0.004523616397329506, - "grad_norm": 3.2410888671875, - "learning_rate": 0.00019969946548532687, - "loss": 0.4038, - "step": 290 - }, - { - "epoch": 0.004679603169651213, - "grad_norm": 4.468445777893066, - "learning_rate": 0.00019968906636717208, - "loss": 0.6076, - "step": 300 - }, - { - "epoch": 0.004835589941972921, - "grad_norm": 8.308616638183594, - "learning_rate": 0.0001996786672490173, - "loss": 0.4465, - "step": 310 - }, - { - "epoch": 0.004991576714294628, - "grad_norm": 3.736629009246826, - "learning_rate": 0.0001996682681308625, - "loss": 0.4133, - "step": 320 - }, - { - "epoch": 0.005147563486616335, - "grad_norm": 2.157092809677124, - "learning_rate": 0.00019965786901270774, - "loss": 0.3663, - "step": 330 - }, - { - "epoch": 0.005303550258938042, - "grad_norm": 5.357859134674072, - "learning_rate": 0.00019964746989455294, - "loss": 0.5599, - "step": 340 - }, - { - "epoch": 0.0054595370312597495, - "grad_norm": 5.352982044219971, - "learning_rate": 0.00019963707077639818, - "loss": 0.6796, - "step": 350 - }, - { - "epoch": 0.0056155238035814565, - "grad_norm": 3.7645204067230225, - "learning_rate": 0.00019962667165824338, - "loss": 0.5774, - "step": 360 - }, - { - "epoch": 0.005771510575903163, - "grad_norm": 6.222738265991211, - "learning_rate": 0.00019961627254008862, - "loss": 0.641, - "step": 370 - }, - { - "epoch": 0.00592749734822487, - "grad_norm": 6.460178852081299, - "learning_rate": 0.00019960587342193381, - "loss": 0.3702, - "step": 380 - }, - { - "epoch": 0.006083484120546578, - "grad_norm": 3.1196858882904053, - "learning_rate": 0.00019959547430377906, - "loss": 0.5036, - "step": 390 - }, - { - "epoch": 0.006239470892868285, - "grad_norm": 7.03672981262207, - "learning_rate": 0.00019958507518562425, - "loss": 0.41, - "step": 400 - }, - { - "epoch": 0.006395457665189992, - "grad_norm": 15.546775817871094, - "learning_rate": 0.0001995746760674695, - "loss": 0.5623, - "step": 410 - }, - { - "epoch": 0.006551444437511699, - "grad_norm": 4.868813514709473, - "learning_rate": 0.0001995642769493147, - "loss": 0.4, - "step": 420 - }, - { - "epoch": 0.0067074312098334065, - "grad_norm": 3.5767409801483154, - "learning_rate": 0.00019955387783115994, - "loss": 0.5193, - "step": 430 - }, - { - "epoch": 0.0068634179821551134, - "grad_norm": 10.07199764251709, - "learning_rate": 0.00019954347871300513, - "loss": 0.5692, - "step": 440 - }, - { - "epoch": 0.00701940475447682, - "grad_norm": 2.809025764465332, - "learning_rate": 0.00019953307959485038, - "loss": 0.4361, - "step": 450 - }, - { - "epoch": 0.007175391526798527, - "grad_norm": 14.244531631469727, - "learning_rate": 0.00019952268047669557, - "loss": 0.3646, - "step": 460 - }, - { - "epoch": 0.007331378299120235, - "grad_norm": 5.827071666717529, - "learning_rate": 0.00019951228135854082, - "loss": 0.5302, - "step": 470 - }, - { - "epoch": 0.007487365071441942, - "grad_norm": 9.067925453186035, - "learning_rate": 0.000199501882240386, - "loss": 0.5045, - "step": 480 - }, - { - "epoch": 0.007643351843763649, - "grad_norm": 8.674382209777832, - "learning_rate": 0.00019949148312223126, - "loss": 0.5778, - "step": 490 - }, - { - "epoch": 0.007799338616085356, - "grad_norm": 8.94734001159668, - "learning_rate": 0.00019948108400407645, - "loss": 0.4141, - "step": 500 - }, - { - "epoch": 0.007955325388407063, - "grad_norm": 10.810057640075684, - "learning_rate": 0.0001994706848859217, - "loss": 0.4069, - "step": 510 - }, - { - "epoch": 0.00811131216072877, - "grad_norm": 9.531766891479492, - "learning_rate": 0.0001994602857677669, - "loss": 0.5935, - "step": 520 - }, - { - "epoch": 0.008267298933050476, - "grad_norm": 13.168464660644531, - "learning_rate": 0.00019944988664961214, - "loss": 0.4725, - "step": 530 - }, - { - "epoch": 0.008423285705372185, - "grad_norm": 4.697941780090332, - "learning_rate": 0.00019943948753145733, - "loss": 0.4516, - "step": 540 - }, - { - "epoch": 0.008579272477693892, - "grad_norm": 2.92124342918396, - "learning_rate": 0.00019942908841330258, - "loss": 0.4562, - "step": 550 - }, - { - "epoch": 0.008735259250015599, - "grad_norm": 9.23039722442627, - "learning_rate": 0.00019941868929514777, - "loss": 0.4051, - "step": 560 - }, - { - "epoch": 0.008891246022337306, - "grad_norm": 10.528753280639648, - "learning_rate": 0.00019940829017699302, - "loss": 0.3475, - "step": 570 - }, - { - "epoch": 0.009047232794659013, - "grad_norm": 6.963693618774414, - "learning_rate": 0.0001993978910588382, - "loss": 0.5338, - "step": 580 - }, - { - "epoch": 0.00920321956698072, - "grad_norm": 6.09390926361084, - "learning_rate": 0.00019938749194068345, - "loss": 0.6501, - "step": 590 - }, - { - "epoch": 0.009359206339302427, - "grad_norm": 4.319295406341553, - "learning_rate": 0.00019937709282252865, - "loss": 0.4163, - "step": 600 - }, - { - "epoch": 0.009515193111624133, - "grad_norm": 7.514628887176514, - "learning_rate": 0.0001993666937043739, - "loss": 0.5193, - "step": 610 - }, - { - "epoch": 0.009671179883945842, - "grad_norm": 5.011287212371826, - "learning_rate": 0.00019935629458621909, - "loss": 0.5743, - "step": 620 - }, - { - "epoch": 0.009827166656267549, - "grad_norm": 6.090816974639893, - "learning_rate": 0.00019934589546806433, - "loss": 0.3902, - "step": 630 - }, - { - "epoch": 0.009983153428589256, - "grad_norm": 10.161015510559082, - "learning_rate": 0.00019933549634990953, - "loss": 0.4463, - "step": 640 - }, - { - "epoch": 0.010139140200910963, - "grad_norm": 6.14881706237793, - "learning_rate": 0.00019932509723175477, - "loss": 0.5009, - "step": 650 - }, - { - "epoch": 0.01029512697323267, - "grad_norm": 3.2960498332977295, - "learning_rate": 0.00019931469811359996, - "loss": 0.4597, - "step": 660 - }, - { - "epoch": 0.010451113745554377, - "grad_norm": 7.595890998840332, - "learning_rate": 0.0001993042989954452, - "loss": 0.5948, - "step": 670 - }, - { - "epoch": 0.010607100517876084, - "grad_norm": 7.178264617919922, - "learning_rate": 0.0001992938998772904, - "loss": 0.7185, - "step": 680 - }, - { - "epoch": 0.01076308729019779, - "grad_norm": 4.092066764831543, - "learning_rate": 0.00019928350075913565, - "loss": 0.3544, - "step": 690 - }, - { - "epoch": 0.010919074062519499, - "grad_norm": 13.203527450561523, - "learning_rate": 0.00019927310164098084, - "loss": 0.5719, - "step": 700 - }, - { - "epoch": 0.011075060834841206, - "grad_norm": 5.39072847366333, - "learning_rate": 0.0001992627025228261, - "loss": 0.5527, - "step": 710 - }, - { - "epoch": 0.011231047607162913, - "grad_norm": 5.769185543060303, - "learning_rate": 0.00019925230340467128, - "loss": 0.4555, - "step": 720 - }, - { - "epoch": 0.01138703437948462, - "grad_norm": 3.0080807209014893, - "learning_rate": 0.00019924190428651653, - "loss": 0.277, - "step": 730 - }, - { - "epoch": 0.011543021151806327, - "grad_norm": 1.0591176748275757, - "learning_rate": 0.00019923150516836172, - "loss": 0.513, - "step": 740 - }, - { - "epoch": 0.011699007924128034, - "grad_norm": 8.458100318908691, - "learning_rate": 0.00019922110605020697, - "loss": 0.5038, - "step": 750 - }, - { - "epoch": 0.01185499469644974, - "grad_norm": 3.121551990509033, - "learning_rate": 0.00019921070693205216, - "loss": 0.2639, - "step": 760 - }, - { - "epoch": 0.012010981468771447, - "grad_norm": 6.6408610343933105, - "learning_rate": 0.00019920030781389738, - "loss": 0.6632, - "step": 770 - }, - { - "epoch": 0.012166968241093156, - "grad_norm": 7.180611610412598, - "learning_rate": 0.0001991899086957426, - "loss": 0.5173, - "step": 780 - }, - { - "epoch": 0.012322955013414863, - "grad_norm": 4.404491901397705, - "learning_rate": 0.00019917950957758782, - "loss": 0.3628, - "step": 790 - }, - { - "epoch": 0.01247894178573657, - "grad_norm": 5.945431232452393, - "learning_rate": 0.00019916911045943304, - "loss": 0.3574, - "step": 800 - }, - { - "epoch": 0.012634928558058277, - "grad_norm": 5.105412483215332, - "learning_rate": 0.00019915871134127826, - "loss": 0.7038, - "step": 810 - }, - { - "epoch": 0.012790915330379984, - "grad_norm": 3.7237210273742676, - "learning_rate": 0.00019914831222312348, - "loss": 0.4426, - "step": 820 - }, - { - "epoch": 0.01294690210270169, - "grad_norm": 10.387605667114258, - "learning_rate": 0.0001991379131049687, - "loss": 0.3152, - "step": 830 - }, - { - "epoch": 0.013102888875023398, - "grad_norm": 2.875211477279663, - "learning_rate": 0.00019912751398681392, - "loss": 0.424, - "step": 840 - }, - { - "epoch": 0.013258875647345104, - "grad_norm": 1.0501248836517334, - "learning_rate": 0.00019911711486865914, - "loss": 0.5571, - "step": 850 - }, - { - "epoch": 0.013414862419666813, - "grad_norm": 6.748145580291748, - "learning_rate": 0.00019910671575050436, - "loss": 0.5226, - "step": 860 - }, - { - "epoch": 0.01357084919198852, - "grad_norm": 6.29374885559082, - "learning_rate": 0.00019909631663234958, - "loss": 0.3036, - "step": 870 - }, - { - "epoch": 0.013726835964310227, - "grad_norm": 5.8310699462890625, - "learning_rate": 0.0001990859175141948, - "loss": 0.4704, - "step": 880 - }, - { - "epoch": 0.013882822736631934, - "grad_norm": 1.4619086980819702, - "learning_rate": 0.00019907551839604002, - "loss": 0.3639, - "step": 890 - }, - { - "epoch": 0.01403880950895364, - "grad_norm": 8.177449226379395, - "learning_rate": 0.00019906511927788524, - "loss": 0.3853, - "step": 900 - }, - { - "epoch": 0.014194796281275348, - "grad_norm": 2.045549154281616, - "learning_rate": 0.00019905472015973046, - "loss": 0.3728, - "step": 910 - }, - { - "epoch": 0.014350783053597055, - "grad_norm": 6.4138875007629395, - "learning_rate": 0.00019904432104157568, - "loss": 0.4615, - "step": 920 - }, - { - "epoch": 0.014506769825918761, - "grad_norm": 2.442739486694336, - "learning_rate": 0.0001990339219234209, - "loss": 0.3592, - "step": 930 - }, - { - "epoch": 0.01466275659824047, - "grad_norm": 3.0741333961486816, - "learning_rate": 0.00019902352280526611, - "loss": 0.4118, - "step": 940 - }, - { - "epoch": 0.014818743370562177, - "grad_norm": 5.163224220275879, - "learning_rate": 0.00019901312368711133, - "loss": 0.7433, - "step": 950 - }, - { - "epoch": 0.014974730142883884, - "grad_norm": 5.6979193687438965, - "learning_rate": 0.00019900272456895655, - "loss": 0.3094, - "step": 960 - }, - { - "epoch": 0.01513071691520559, - "grad_norm": 3.2027347087860107, - "learning_rate": 0.00019899232545080177, - "loss": 0.3749, - "step": 970 - }, - { - "epoch": 0.015286703687527298, - "grad_norm": 9.097466468811035, - "learning_rate": 0.000198981926332647, - "loss": 0.5344, - "step": 980 - }, - { - "epoch": 0.015442690459849005, - "grad_norm": 3.730250835418701, - "learning_rate": 0.0001989715272144922, - "loss": 0.4085, - "step": 990 - }, - { - "epoch": 0.015598677232170712, - "grad_norm": 4.03507661819458, - "learning_rate": 0.00019896112809633743, - "loss": 0.7935, - "step": 1000 - }, - { - "epoch": 0.01575466400449242, - "grad_norm": 6.676806449890137, - "learning_rate": 0.00019895072897818265, - "loss": 0.4276, - "step": 1010 - }, - { - "epoch": 0.015910650776814125, - "grad_norm": 2.294931173324585, - "learning_rate": 0.00019894032986002787, - "loss": 0.2544, - "step": 1020 - }, - { - "epoch": 0.016066637549135832, - "grad_norm": 15.660760879516602, - "learning_rate": 0.0001989299307418731, - "loss": 0.2441, - "step": 1030 - }, - { - "epoch": 0.01622262432145754, - "grad_norm": 3.4223055839538574, - "learning_rate": 0.00019891953162371834, - "loss": 0.4861, - "step": 1040 - }, - { - "epoch": 0.016378611093779246, - "grad_norm": 8.601675987243652, - "learning_rate": 0.00019890913250556353, - "loss": 0.7212, - "step": 1050 - }, - { - "epoch": 0.016534597866100953, - "grad_norm": 3.814286708831787, - "learning_rate": 0.00019889873338740878, - "loss": 0.4785, - "step": 1060 - }, - { - "epoch": 0.016690584638422663, - "grad_norm": 10.6279296875, - "learning_rate": 0.00019888833426925397, - "loss": 0.278, - "step": 1070 - }, - { - "epoch": 0.01684657141074437, - "grad_norm": 2.3790910243988037, - "learning_rate": 0.00019887793515109922, - "loss": 0.3589, - "step": 1080 - }, - { - "epoch": 0.017002558183066077, - "grad_norm": 8.939804077148438, - "learning_rate": 0.0001988675360329444, - "loss": 0.5794, - "step": 1090 - }, - { - "epoch": 0.017158544955387784, - "grad_norm": 7.628971576690674, - "learning_rate": 0.00019885713691478966, - "loss": 0.5317, - "step": 1100 - }, - { - "epoch": 0.01731453172770949, - "grad_norm": 1.8617459535598755, - "learning_rate": 0.00019884673779663485, - "loss": 0.433, - "step": 1110 - }, - { - "epoch": 0.017470518500031198, - "grad_norm": 6.928651809692383, - "learning_rate": 0.0001988363386784801, - "loss": 0.216, - "step": 1120 - }, - { - "epoch": 0.017626505272352905, - "grad_norm": 3.6206703186035156, - "learning_rate": 0.0001988259395603253, - "loss": 0.4963, - "step": 1130 - }, - { - "epoch": 0.01778249204467461, - "grad_norm": 4.615293025970459, - "learning_rate": 0.00019881554044217053, - "loss": 0.3579, - "step": 1140 - }, - { - "epoch": 0.01793847881699632, - "grad_norm": 3.3206489086151123, - "learning_rate": 0.00019880514132401573, - "loss": 0.3717, - "step": 1150 - }, - { - "epoch": 0.018094465589318025, - "grad_norm": 7.5789408683776855, - "learning_rate": 0.00019879474220586097, - "loss": 0.2646, - "step": 1160 - }, - { - "epoch": 0.018250452361639732, - "grad_norm": 4.467035293579102, - "learning_rate": 0.00019878434308770617, - "loss": 0.1804, - "step": 1170 - }, - { - "epoch": 0.01840643913396144, - "grad_norm": 9.318495750427246, - "learning_rate": 0.0001987739439695514, - "loss": 0.3257, - "step": 1180 - }, - { - "epoch": 0.018562425906283146, - "grad_norm": 3.1046931743621826, - "learning_rate": 0.0001987635448513966, - "loss": 0.4267, - "step": 1190 - }, - { - "epoch": 0.018718412678604853, - "grad_norm": 4.199711799621582, - "learning_rate": 0.00019875314573324185, - "loss": 0.6676, - "step": 1200 - }, - { - "epoch": 0.01887439945092656, - "grad_norm": 12.733593940734863, - "learning_rate": 0.00019874274661508705, - "loss": 0.5991, - "step": 1210 - }, - { - "epoch": 0.019030386223248267, - "grad_norm": 5.1266984939575195, - "learning_rate": 0.0001987323474969323, - "loss": 0.2945, - "step": 1220 - }, - { - "epoch": 0.019186372995569977, - "grad_norm": 6.397432804107666, - "learning_rate": 0.00019872194837877748, - "loss": 0.4774, - "step": 1230 - }, - { - "epoch": 0.019342359767891684, - "grad_norm": 3.761129379272461, - "learning_rate": 0.0001987115492606227, - "loss": 0.4345, - "step": 1240 - }, - { - "epoch": 0.01949834654021339, - "grad_norm": 2.7355475425720215, - "learning_rate": 0.00019870115014246792, - "loss": 0.497, - "step": 1250 - }, - { - "epoch": 0.019654333312535098, - "grad_norm": 2.7893755435943604, - "learning_rate": 0.00019869075102431314, - "loss": 0.2169, - "step": 1260 - }, - { - "epoch": 0.019810320084856805, - "grad_norm": 10.722111701965332, - "learning_rate": 0.00019868035190615836, - "loss": 0.3867, - "step": 1270 - }, - { - "epoch": 0.019966306857178512, - "grad_norm": 11.647889137268066, - "learning_rate": 0.00019866995278800358, - "loss": 0.4138, - "step": 1280 - }, - { - "epoch": 0.02012229362950022, - "grad_norm": 1.4313205480575562, - "learning_rate": 0.0001986595536698488, - "loss": 0.3581, - "step": 1290 - }, - { - "epoch": 0.020278280401821926, - "grad_norm": 5.193609714508057, - "learning_rate": 0.00019864915455169402, - "loss": 0.5259, - "step": 1300 - }, - { - "epoch": 0.020434267174143633, - "grad_norm": 3.1939520835876465, - "learning_rate": 0.00019863875543353924, - "loss": 0.5391, - "step": 1310 - }, - { - "epoch": 0.02059025394646534, - "grad_norm": 7.716986179351807, - "learning_rate": 0.00019862835631538446, - "loss": 0.4773, - "step": 1320 - }, - { - "epoch": 0.020746240718787046, - "grad_norm": 3.191506862640381, - "learning_rate": 0.00019861795719722968, - "loss": 0.6051, - "step": 1330 - }, - { - "epoch": 0.020902227491108753, - "grad_norm": 5.67583703994751, - "learning_rate": 0.0001986075580790749, - "loss": 0.2716, - "step": 1340 - }, - { - "epoch": 0.02105821426343046, - "grad_norm": 3.1657073497772217, - "learning_rate": 0.00019859715896092012, - "loss": 0.4838, - "step": 1350 - }, - { - "epoch": 0.021214201035752167, - "grad_norm": 5.144644260406494, - "learning_rate": 0.00019858675984276534, - "loss": 0.3185, - "step": 1360 - }, - { - "epoch": 0.021370187808073874, - "grad_norm": 0.3853607177734375, - "learning_rate": 0.00019857636072461056, - "loss": 0.21, - "step": 1370 - }, - { - "epoch": 0.02152617458039558, - "grad_norm": 15.329313278198242, - "learning_rate": 0.00019856596160645578, - "loss": 0.4216, - "step": 1380 - }, - { - "epoch": 0.02168216135271729, - "grad_norm": 1.9162156581878662, - "learning_rate": 0.000198555562488301, - "loss": 0.3766, - "step": 1390 - }, - { - "epoch": 0.021838148125038998, - "grad_norm": 2.5800933837890625, - "learning_rate": 0.00019854516337014622, - "loss": 0.4478, - "step": 1400 - }, - { - "epoch": 0.021994134897360705, - "grad_norm": 12.898444175720215, - "learning_rate": 0.00019853476425199144, - "loss": 0.7923, - "step": 1410 - }, - { - "epoch": 0.022150121669682412, - "grad_norm": 11.576621055603027, - "learning_rate": 0.00019852436513383666, - "loss": 0.3582, - "step": 1420 - }, - { - "epoch": 0.02230610844200412, - "grad_norm": 10.80225658416748, - "learning_rate": 0.00019851396601568188, - "loss": 0.4698, - "step": 1430 - }, - { - "epoch": 0.022462095214325826, - "grad_norm": 5.019324779510498, - "learning_rate": 0.0001985035668975271, - "loss": 0.4463, - "step": 1440 - }, - { - "epoch": 0.022618081986647533, - "grad_norm": 5.366154670715332, - "learning_rate": 0.00019849316777937232, - "loss": 0.4288, - "step": 1450 - }, - { - "epoch": 0.02277406875896924, - "grad_norm": 3.945873498916626, - "learning_rate": 0.00019848276866121754, - "loss": 0.5625, - "step": 1460 - }, - { - "epoch": 0.022930055531290947, - "grad_norm": 1.8938435316085815, - "learning_rate": 0.00019847236954306276, - "loss": 0.4298, - "step": 1470 - }, - { - "epoch": 0.023086042303612653, - "grad_norm": 7.040036678314209, - "learning_rate": 0.00019846197042490798, - "loss": 0.5806, - "step": 1480 - }, - { - "epoch": 0.02324202907593436, - "grad_norm": 12.009136199951172, - "learning_rate": 0.0001984515713067532, - "loss": 0.3603, - "step": 1490 - }, - { - "epoch": 0.023398015848256067, - "grad_norm": 3.157111883163452, - "learning_rate": 0.00019844117218859841, - "loss": 0.3811, - "step": 1500 - }, - { - "epoch": 0.023554002620577774, - "grad_norm": 1.0227491855621338, - "learning_rate": 0.00019843077307044363, - "loss": 0.3459, - "step": 1510 - }, - { - "epoch": 0.02370998939289948, - "grad_norm": 2.699263334274292, - "learning_rate": 0.00019842037395228885, - "loss": 0.4047, - "step": 1520 - }, - { - "epoch": 0.023865976165221188, - "grad_norm": 3.879258871078491, - "learning_rate": 0.00019840997483413407, - "loss": 0.2083, - "step": 1530 - }, - { - "epoch": 0.024021962937542895, - "grad_norm": 1.0468460321426392, - "learning_rate": 0.0001983995757159793, - "loss": 0.3146, - "step": 1540 - }, - { - "epoch": 0.024177949709864602, - "grad_norm": 3.5738143920898438, - "learning_rate": 0.0001983891765978245, - "loss": 0.3181, - "step": 1550 - }, - { - "epoch": 0.024333936482186312, - "grad_norm": 6.3198113441467285, - "learning_rate": 0.00019837877747966973, - "loss": 0.3196, - "step": 1560 - }, - { - "epoch": 0.02448992325450802, - "grad_norm": 1.6974947452545166, - "learning_rate": 0.00019836837836151495, - "loss": 0.3832, - "step": 1570 - }, - { - "epoch": 0.024645910026829726, - "grad_norm": 6.220627307891846, - "learning_rate": 0.00019835797924336017, - "loss": 0.3445, - "step": 1580 - }, - { - "epoch": 0.024801896799151433, - "grad_norm": 3.9202497005462646, - "learning_rate": 0.0001983475801252054, - "loss": 0.4304, - "step": 1590 - }, - { - "epoch": 0.02495788357147314, - "grad_norm": 2.2945311069488525, - "learning_rate": 0.0001983371810070506, - "loss": 0.4487, - "step": 1600 - }, - { - "epoch": 0.025113870343794847, - "grad_norm": 9.262955665588379, - "learning_rate": 0.00019832678188889583, - "loss": 0.3691, - "step": 1610 - }, - { - "epoch": 0.025269857116116554, - "grad_norm": 2.6905221939086914, - "learning_rate": 0.00019831638277074105, - "loss": 0.5117, - "step": 1620 - }, - { - "epoch": 0.02542584388843826, - "grad_norm": 4.1330060958862305, - "learning_rate": 0.00019830598365258627, - "loss": 0.5721, - "step": 1630 - }, - { - "epoch": 0.025581830660759967, - "grad_norm": 1.0681806802749634, - "learning_rate": 0.0001982955845344315, - "loss": 0.4794, - "step": 1640 - }, - { - "epoch": 0.025737817433081674, - "grad_norm": 2.6171841621398926, - "learning_rate": 0.0001982851854162767, - "loss": 0.4596, - "step": 1650 - }, - { - "epoch": 0.02589380420540338, - "grad_norm": 3.3451762199401855, - "learning_rate": 0.00019827478629812193, - "loss": 0.4745, - "step": 1660 - }, - { - "epoch": 0.026049790977725088, - "grad_norm": 4.152349472045898, - "learning_rate": 0.00019826438717996715, - "loss": 0.4452, - "step": 1670 - }, - { - "epoch": 0.026205777750046795, - "grad_norm": 2.7109954357147217, - "learning_rate": 0.00019825398806181237, - "loss": 0.3966, - "step": 1680 - }, - { - "epoch": 0.026361764522368502, - "grad_norm": 5.731235980987549, - "learning_rate": 0.0001982435889436576, - "loss": 0.4423, - "step": 1690 - }, - { - "epoch": 0.02651775129469021, - "grad_norm": 2.370852470397949, - "learning_rate": 0.0001982331898255028, - "loss": 0.5619, - "step": 1700 - }, - { - "epoch": 0.026673738067011916, - "grad_norm": 2.5500876903533936, - "learning_rate": 0.00019822279070734803, - "loss": 0.2658, - "step": 1710 - }, - { - "epoch": 0.026829724839333626, - "grad_norm": 0.20157204568386078, - "learning_rate": 0.00019821239158919325, - "loss": 0.2939, - "step": 1720 - }, - { - "epoch": 0.026985711611655333, - "grad_norm": 2.6305084228515625, - "learning_rate": 0.00019820199247103847, - "loss": 0.5339, - "step": 1730 - }, - { - "epoch": 0.02714169838397704, - "grad_norm": 6.241075038909912, - "learning_rate": 0.00019819159335288369, - "loss": 0.3625, - "step": 1740 - }, - { - "epoch": 0.027297685156298747, - "grad_norm": 5.584851264953613, - "learning_rate": 0.0001981811942347289, - "loss": 0.6158, - "step": 1750 - }, - { - "epoch": 0.027453671928620454, - "grad_norm": 6.756530284881592, - "learning_rate": 0.00019817079511657413, - "loss": 0.3513, - "step": 1760 - }, - { - "epoch": 0.02760965870094216, - "grad_norm": 5.153111457824707, - "learning_rate": 0.00019816039599841935, - "loss": 0.2885, - "step": 1770 - }, - { - "epoch": 0.027765645473263868, - "grad_norm": 6.222873210906982, - "learning_rate": 0.00019814999688026456, - "loss": 0.3291, - "step": 1780 - }, - { - "epoch": 0.027921632245585574, - "grad_norm": 4.744391918182373, - "learning_rate": 0.00019813959776210978, - "loss": 0.5053, - "step": 1790 - }, - { - "epoch": 0.02807761901790728, - "grad_norm": 3.7181131839752197, - "learning_rate": 0.000198129198643955, - "loss": 0.3983, - "step": 1800 - }, - { - "epoch": 0.02823360579022899, - "grad_norm": 1.8923834562301636, - "learning_rate": 0.00019811879952580022, - "loss": 0.4465, - "step": 1810 - }, - { - "epoch": 0.028389592562550695, - "grad_norm": 6.049956321716309, - "learning_rate": 0.00019810840040764544, - "loss": 0.5028, - "step": 1820 - }, - { - "epoch": 0.028545579334872402, - "grad_norm": 8.034428596496582, - "learning_rate": 0.00019809800128949066, - "loss": 0.4799, - "step": 1830 - }, - { - "epoch": 0.02870156610719411, - "grad_norm": 3.916890859603882, - "learning_rate": 0.00019808760217133588, - "loss": 0.3109, - "step": 1840 - }, - { - "epoch": 0.028857552879515816, - "grad_norm": 3.351285457611084, - "learning_rate": 0.0001980772030531811, - "loss": 0.5981, - "step": 1850 - }, - { - "epoch": 0.029013539651837523, - "grad_norm": 4.443714141845703, - "learning_rate": 0.00019806680393502632, - "loss": 0.3243, - "step": 1860 - }, - { - "epoch": 0.02916952642415923, - "grad_norm": 0.6998550295829773, - "learning_rate": 0.00019805640481687154, - "loss": 0.2057, - "step": 1870 - }, - { - "epoch": 0.02932551319648094, - "grad_norm": 12.252562522888184, - "learning_rate": 0.00019804600569871676, - "loss": 0.3806, - "step": 1880 - }, - { - "epoch": 0.029481499968802647, - "grad_norm": 9.03868293762207, - "learning_rate": 0.00019803560658056198, - "loss": 0.3351, - "step": 1890 - }, - { - "epoch": 0.029637486741124354, - "grad_norm": 2.3400044441223145, - "learning_rate": 0.0001980252074624072, - "loss": 0.3103, - "step": 1900 - }, - { - "epoch": 0.02979347351344606, - "grad_norm": 2.399489164352417, - "learning_rate": 0.00019801480834425242, - "loss": 0.4792, - "step": 1910 - }, - { - "epoch": 0.029949460285767768, - "grad_norm": 1.5152381658554077, - "learning_rate": 0.00019800440922609764, - "loss": 0.2873, - "step": 1920 - }, - { - "epoch": 0.030105447058089475, - "grad_norm": 2.7013959884643555, - "learning_rate": 0.00019799401010794286, - "loss": 0.381, - "step": 1930 - }, - { - "epoch": 0.03026143383041118, - "grad_norm": 0.7135261297225952, - "learning_rate": 0.00019798361098978808, - "loss": 0.4001, - "step": 1940 - }, - { - "epoch": 0.03041742060273289, - "grad_norm": 15.61109447479248, - "learning_rate": 0.0001979732118716333, - "loss": 0.474, - "step": 1950 - }, - { - "epoch": 0.030573407375054595, - "grad_norm": 3.236668348312378, - "learning_rate": 0.00019796281275347852, - "loss": 0.2149, - "step": 1960 - }, - { - "epoch": 0.030729394147376302, - "grad_norm": 4.722929954528809, - "learning_rate": 0.00019795241363532374, - "loss": 0.5474, - "step": 1970 - }, - { - "epoch": 0.03088538091969801, - "grad_norm": 6.002335548400879, - "learning_rate": 0.00019794201451716896, - "loss": 0.5196, - "step": 1980 - }, - { - "epoch": 0.031041367692019716, - "grad_norm": 5.210022926330566, - "learning_rate": 0.00019793161539901418, - "loss": 0.4741, - "step": 1990 - }, - { - "epoch": 0.031197354464341423, - "grad_norm": 7.984034538269043, - "learning_rate": 0.0001979212162808594, - "loss": 0.5399, - "step": 2000 - }, - { - "epoch": 0.03135334123666313, - "grad_norm": 4.008991241455078, - "learning_rate": 0.00019791081716270462, - "loss": 0.4788, - "step": 2010 - }, - { - "epoch": 0.03150932800898484, - "grad_norm": 4.395316123962402, - "learning_rate": 0.00019790041804454984, - "loss": 0.3456, - "step": 2020 - }, - { - "epoch": 0.03166531478130655, - "grad_norm": 3.9784998893737793, - "learning_rate": 0.00019789001892639506, - "loss": 0.326, - "step": 2030 - }, - { - "epoch": 0.03182130155362825, - "grad_norm": 5.040256023406982, - "learning_rate": 0.00019787961980824028, - "loss": 0.5264, - "step": 2040 - }, - { - "epoch": 0.03197728832594996, - "grad_norm": 2.686203956604004, - "learning_rate": 0.0001978692206900855, - "loss": 0.5461, - "step": 2050 - }, - { - "epoch": 0.032133275098271664, - "grad_norm": 1.8417892456054688, - "learning_rate": 0.00019785882157193071, - "loss": 0.4736, - "step": 2060 - }, - { - "epoch": 0.032289261870593375, - "grad_norm": 3.623887300491333, - "learning_rate": 0.00019784842245377593, - "loss": 0.3742, - "step": 2070 - }, - { - "epoch": 0.03244524864291508, - "grad_norm": 0.5184723734855652, - "learning_rate": 0.00019783802333562115, - "loss": 0.3941, - "step": 2080 - }, - { - "epoch": 0.03260123541523679, - "grad_norm": 7.751046180725098, - "learning_rate": 0.00019782762421746637, - "loss": 0.4609, - "step": 2090 - }, - { - "epoch": 0.03275722218755849, - "grad_norm": 2.547175168991089, - "learning_rate": 0.0001978172250993116, - "loss": 0.3089, - "step": 2100 - }, - { - "epoch": 0.0329132089598802, - "grad_norm": 1.8453168869018555, - "learning_rate": 0.0001978068259811568, - "loss": 0.4371, - "step": 2110 - }, - { - "epoch": 0.033069195732201906, - "grad_norm": 7.033441543579102, - "learning_rate": 0.00019779642686300203, - "loss": 0.3842, - "step": 2120 - }, - { - "epoch": 0.033225182504523616, - "grad_norm": 2.2642929553985596, - "learning_rate": 0.00019778602774484725, - "loss": 0.2069, - "step": 2130 - }, - { - "epoch": 0.03338116927684533, - "grad_norm": 4.043938636779785, - "learning_rate": 0.00019777562862669244, - "loss": 0.4148, - "step": 2140 - }, - { - "epoch": 0.03353715604916703, - "grad_norm": 4.227573871612549, - "learning_rate": 0.0001977652295085377, - "loss": 0.4785, - "step": 2150 - }, - { - "epoch": 0.03369314282148874, - "grad_norm": 2.0808868408203125, - "learning_rate": 0.00019775483039038288, - "loss": 0.3073, - "step": 2160 - }, - { - "epoch": 0.033849129593810444, - "grad_norm": 4.518038749694824, - "learning_rate": 0.00019774443127222813, - "loss": 0.4197, - "step": 2170 - }, - { - "epoch": 0.034005116366132154, - "grad_norm": 2.3731303215026855, - "learning_rate": 0.00019773403215407332, - "loss": 0.2753, - "step": 2180 - }, - { - "epoch": 0.03416110313845386, - "grad_norm": 1.7613986730575562, - "learning_rate": 0.00019772363303591857, - "loss": 0.3834, - "step": 2190 - }, - { - "epoch": 0.03431708991077557, - "grad_norm": 2.9518485069274902, - "learning_rate": 0.00019771323391776376, - "loss": 0.3544, - "step": 2200 - }, - { - "epoch": 0.03447307668309727, - "grad_norm": 4.424034118652344, - "learning_rate": 0.000197702834799609, - "loss": 0.4046, - "step": 2210 - }, - { - "epoch": 0.03462906345541898, - "grad_norm": 2.307690382003784, - "learning_rate": 0.0001976924356814542, - "loss": 0.1445, - "step": 2220 - }, - { - "epoch": 0.034785050227740685, - "grad_norm": 2.536533832550049, - "learning_rate": 0.00019768203656329945, - "loss": 0.2815, - "step": 2230 - }, - { - "epoch": 0.034941037000062396, - "grad_norm": 2.006728172302246, - "learning_rate": 0.00019767163744514464, - "loss": 0.5314, - "step": 2240 - }, - { - "epoch": 0.0350970237723841, - "grad_norm": 4.729264259338379, - "learning_rate": 0.0001976612383269899, - "loss": 0.4547, - "step": 2250 - }, - { - "epoch": 0.03525301054470581, - "grad_norm": 4.600561141967773, - "learning_rate": 0.00019765083920883508, - "loss": 0.3967, - "step": 2260 - }, - { - "epoch": 0.03540899731702751, - "grad_norm": 0.6419815421104431, - "learning_rate": 0.00019764044009068033, - "loss": 0.336, - "step": 2270 - }, - { - "epoch": 0.03556498408934922, - "grad_norm": 5.567695140838623, - "learning_rate": 0.00019763004097252552, - "loss": 0.5644, - "step": 2280 - }, - { - "epoch": 0.035720970861670934, - "grad_norm": 0.5767773389816284, - "learning_rate": 0.00019761964185437077, - "loss": 0.2643, - "step": 2290 - }, - { - "epoch": 0.03587695763399264, - "grad_norm": 4.91269063949585, - "learning_rate": 0.00019760924273621596, - "loss": 0.4198, - "step": 2300 - }, - { - "epoch": 0.03603294440631435, - "grad_norm": 4.422239780426025, - "learning_rate": 0.0001975988436180612, - "loss": 0.5338, - "step": 2310 - }, - { - "epoch": 0.03618893117863605, - "grad_norm": 1.672713041305542, - "learning_rate": 0.0001975884444999064, - "loss": 0.3214, - "step": 2320 - }, - { - "epoch": 0.03634491795095776, - "grad_norm": 4.033807277679443, - "learning_rate": 0.00019757804538175164, - "loss": 0.3593, - "step": 2330 - }, - { - "epoch": 0.036500904723279465, - "grad_norm": 3.198880910873413, - "learning_rate": 0.00019756764626359684, - "loss": 0.3657, - "step": 2340 - }, - { - "epoch": 0.036656891495601175, - "grad_norm": 2.640443801879883, - "learning_rate": 0.00019755724714544208, - "loss": 0.3243, - "step": 2350 - }, - { - "epoch": 0.03681287826792288, - "grad_norm": 1.562350869178772, - "learning_rate": 0.00019754684802728728, - "loss": 0.3362, - "step": 2360 - }, - { - "epoch": 0.03696886504024459, - "grad_norm": 4.379998683929443, - "learning_rate": 0.00019753644890913252, - "loss": 0.541, - "step": 2370 - }, - { - "epoch": 0.03712485181256629, - "grad_norm": 2.529301881790161, - "learning_rate": 0.00019752604979097774, - "loss": 0.5396, - "step": 2380 - }, - { - "epoch": 0.037280838584888, - "grad_norm": 1.7864131927490234, - "learning_rate": 0.00019751565067282296, - "loss": 0.3165, - "step": 2390 - }, - { - "epoch": 0.037436825357209706, - "grad_norm": 1.0566092729568481, - "learning_rate": 0.00019750525155466818, - "loss": 0.4225, - "step": 2400 - }, - { - "epoch": 0.03759281212953142, - "grad_norm": 2.4530510902404785, - "learning_rate": 0.0001974948524365134, - "loss": 0.4281, - "step": 2410 - }, - { - "epoch": 0.03774879890185312, - "grad_norm": 5.363385200500488, - "learning_rate": 0.00019748445331835862, - "loss": 0.4746, - "step": 2420 - }, - { - "epoch": 0.03790478567417483, - "grad_norm": 6.038578510284424, - "learning_rate": 0.00019747405420020384, - "loss": 0.3182, - "step": 2430 - }, - { - "epoch": 0.038060772446496534, - "grad_norm": 9.37709903717041, - "learning_rate": 0.00019746365508204906, - "loss": 0.3288, - "step": 2440 - }, - { - "epoch": 0.038216759218818244, - "grad_norm": 28.81134033203125, - "learning_rate": 0.00019745325596389428, - "loss": 0.2646, - "step": 2450 - }, - { - "epoch": 0.038372745991139955, - "grad_norm": 0.616172730922699, - "learning_rate": 0.0001974428568457395, - "loss": 0.3848, - "step": 2460 - }, - { - "epoch": 0.03852873276346166, - "grad_norm": 2.9943106174468994, - "learning_rate": 0.00019743245772758472, - "loss": 0.4097, - "step": 2470 - }, - { - "epoch": 0.03868471953578337, - "grad_norm": 4.001022815704346, - "learning_rate": 0.00019742205860942994, - "loss": 0.2359, - "step": 2480 - }, - { - "epoch": 0.03884070630810507, - "grad_norm": 3.587982416152954, - "learning_rate": 0.00019741165949127516, - "loss": 0.5829, - "step": 2490 - }, - { - "epoch": 0.03899669308042678, - "grad_norm": 1.906042218208313, - "learning_rate": 0.00019740126037312038, - "loss": 0.336, - "step": 2500 - }, - { - "epoch": 0.039152679852748486, - "grad_norm": 9.536376953125, - "learning_rate": 0.0001973908612549656, - "loss": 0.3468, - "step": 2510 - }, - { - "epoch": 0.039308666625070196, - "grad_norm": 1.582669734954834, - "learning_rate": 0.00019738046213681082, - "loss": 0.3534, - "step": 2520 - }, - { - "epoch": 0.0394646533973919, - "grad_norm": 2.59112286567688, - "learning_rate": 0.00019737006301865604, - "loss": 0.239, - "step": 2530 - }, - { - "epoch": 0.03962064016971361, - "grad_norm": 2.091029167175293, - "learning_rate": 0.00019735966390050126, - "loss": 0.5045, - "step": 2540 - }, - { - "epoch": 0.03977662694203531, - "grad_norm": 5.665968894958496, - "learning_rate": 0.00019734926478234648, - "loss": 0.3729, - "step": 2550 - }, - { - "epoch": 0.039932613714357024, - "grad_norm": 2.2713849544525146, - "learning_rate": 0.0001973388656641917, - "loss": 0.3598, - "step": 2560 - }, - { - "epoch": 0.04008860048667873, - "grad_norm": 3.3854405879974365, - "learning_rate": 0.00019732846654603692, - "loss": 0.2485, - "step": 2570 - }, - { - "epoch": 0.04024458725900044, - "grad_norm": 7.780918598175049, - "learning_rate": 0.00019731806742788214, - "loss": 0.3557, - "step": 2580 - }, - { - "epoch": 0.04040057403132214, - "grad_norm": 2.7871644496917725, - "learning_rate": 0.00019730766830972733, - "loss": 0.7119, - "step": 2590 - }, - { - "epoch": 0.04055656080364385, - "grad_norm": 2.53657603263855, - "learning_rate": 0.00019729726919157258, - "loss": 0.3575, - "step": 2600 - }, - { - "epoch": 0.040712547575965555, - "grad_norm": 0.3270189166069031, - "learning_rate": 0.00019728687007341777, - "loss": 0.3733, - "step": 2610 - }, - { - "epoch": 0.040868534348287265, - "grad_norm": 1.2280256748199463, - "learning_rate": 0.00019727647095526301, - "loss": 0.458, - "step": 2620 - }, - { - "epoch": 0.041024521120608975, - "grad_norm": 0.7863756418228149, - "learning_rate": 0.0001972660718371082, - "loss": 0.6961, - "step": 2630 - }, - { - "epoch": 0.04118050789293068, - "grad_norm": 3.562969207763672, - "learning_rate": 0.00019725567271895345, - "loss": 0.3823, - "step": 2640 - }, - { - "epoch": 0.04133649466525239, - "grad_norm": 2.5304505825042725, - "learning_rate": 0.00019724527360079865, - "loss": 0.4938, - "step": 2650 - }, - { - "epoch": 0.04149248143757409, - "grad_norm": 1.8894665241241455, - "learning_rate": 0.0001972348744826439, - "loss": 0.4397, - "step": 2660 - }, - { - "epoch": 0.0416484682098958, - "grad_norm": 3.460160970687866, - "learning_rate": 0.00019722447536448909, - "loss": 0.3571, - "step": 2670 - }, - { - "epoch": 0.04180445498221751, - "grad_norm": 4.1218061447143555, - "learning_rate": 0.00019721407624633433, - "loss": 0.2929, - "step": 2680 - }, - { - "epoch": 0.04196044175453922, - "grad_norm": 2.9370386600494385, - "learning_rate": 0.00019720367712817952, - "loss": 0.4142, - "step": 2690 - }, - { - "epoch": 0.04211642852686092, - "grad_norm": 5.4618401527404785, - "learning_rate": 0.00019719327801002477, - "loss": 0.3288, - "step": 2700 - }, - { - "epoch": 0.04227241529918263, - "grad_norm": 3.1975386142730713, - "learning_rate": 0.00019718287889186996, - "loss": 0.3373, - "step": 2710 - }, - { - "epoch": 0.042428402071504334, - "grad_norm": 4.580129623413086, - "learning_rate": 0.0001971724797737152, - "loss": 0.7003, - "step": 2720 - }, - { - "epoch": 0.042584388843826045, - "grad_norm": 1.7264137268066406, - "learning_rate": 0.0001971620806555604, - "loss": 0.3375, - "step": 2730 - }, - { - "epoch": 0.04274037561614775, - "grad_norm": 8.90623950958252, - "learning_rate": 0.00019715168153740565, - "loss": 0.4492, - "step": 2740 - }, - { - "epoch": 0.04289636238846946, - "grad_norm": 8.407225608825684, - "learning_rate": 0.00019714128241925084, - "loss": 0.3827, - "step": 2750 - }, - { - "epoch": 0.04305234916079116, - "grad_norm": 5.202430725097656, - "learning_rate": 0.0001971308833010961, - "loss": 0.2806, - "step": 2760 - }, - { - "epoch": 0.04320833593311287, - "grad_norm": 1.9569600820541382, - "learning_rate": 0.00019712048418294128, - "loss": 0.2506, - "step": 2770 - }, - { - "epoch": 0.04336432270543458, - "grad_norm": 3.2171106338500977, - "learning_rate": 0.00019711008506478653, - "loss": 0.3344, - "step": 2780 - }, - { - "epoch": 0.043520309477756286, - "grad_norm": 0.7588744163513184, - "learning_rate": 0.00019709968594663172, - "loss": 0.3656, - "step": 2790 - }, - { - "epoch": 0.043676296250077996, - "grad_norm": 1.4201767444610596, - "learning_rate": 0.00019708928682847697, - "loss": 0.4033, - "step": 2800 - }, - { - "epoch": 0.0438322830223997, - "grad_norm": 2.439276933670044, - "learning_rate": 0.00019707888771032216, - "loss": 0.2828, - "step": 2810 - }, - { - "epoch": 0.04398826979472141, - "grad_norm": 18.436410903930664, - "learning_rate": 0.0001970684885921674, - "loss": 0.6879, - "step": 2820 - }, - { - "epoch": 0.044144256567043114, - "grad_norm": 3.1035194396972656, - "learning_rate": 0.0001970580894740126, - "loss": 0.3022, - "step": 2830 - }, - { - "epoch": 0.044300243339364824, - "grad_norm": 4.849295139312744, - "learning_rate": 0.00019704769035585785, - "loss": 0.4078, - "step": 2840 - }, - { - "epoch": 0.04445623011168653, - "grad_norm": 3.680732488632202, - "learning_rate": 0.00019703729123770304, - "loss": 0.4738, - "step": 2850 - }, - { - "epoch": 0.04461221688400824, - "grad_norm": 2.8644354343414307, - "learning_rate": 0.00019702689211954829, - "loss": 0.386, - "step": 2860 - }, - { - "epoch": 0.04476820365632994, - "grad_norm": 0.7621586918830872, - "learning_rate": 0.00019701649300139348, - "loss": 0.2032, - "step": 2870 - }, - { - "epoch": 0.04492419042865165, - "grad_norm": 0.9556747078895569, - "learning_rate": 0.00019700609388323873, - "loss": 0.5366, - "step": 2880 - }, - { - "epoch": 0.045080177200973355, - "grad_norm": 4.282146453857422, - "learning_rate": 0.00019699569476508392, - "loss": 0.4814, - "step": 2890 - }, - { - "epoch": 0.045236163973295065, - "grad_norm": 1.905647873878479, - "learning_rate": 0.00019698529564692916, - "loss": 0.5273, - "step": 2900 - }, - { - "epoch": 0.04539215074561677, - "grad_norm": 0.7890440821647644, - "learning_rate": 0.00019697489652877436, - "loss": 0.3434, - "step": 2910 - }, - { - "epoch": 0.04554813751793848, - "grad_norm": 1.717529296875, - "learning_rate": 0.0001969644974106196, - "loss": 0.6094, - "step": 2920 - }, - { - "epoch": 0.04570412429026018, - "grad_norm": 3.9639952182769775, - "learning_rate": 0.0001969540982924648, - "loss": 0.2958, - "step": 2930 - }, - { - "epoch": 0.04586011106258189, - "grad_norm": 1.4309637546539307, - "learning_rate": 0.00019694369917431004, - "loss": 0.2879, - "step": 2940 - }, - { - "epoch": 0.0460160978349036, - "grad_norm": 8.747053146362305, - "learning_rate": 0.00019693330005615524, - "loss": 0.5953, - "step": 2950 - }, - { - "epoch": 0.04617208460722531, - "grad_norm": 0.027920352295041084, - "learning_rate": 0.00019692290093800048, - "loss": 0.2181, - "step": 2960 - }, - { - "epoch": 0.04632807137954702, - "grad_norm": 1.9787837266921997, - "learning_rate": 0.00019691250181984567, - "loss": 0.2764, - "step": 2970 - }, - { - "epoch": 0.04648405815186872, - "grad_norm": 10.669599533081055, - "learning_rate": 0.00019690210270169092, - "loss": 0.5824, - "step": 2980 - }, - { - "epoch": 0.04664004492419043, - "grad_norm": 2.9555306434631348, - "learning_rate": 0.00019689170358353611, - "loss": 0.5993, - "step": 2990 - }, - { - "epoch": 0.046796031696512135, - "grad_norm": 2.168975353240967, - "learning_rate": 0.00019688130446538136, - "loss": 0.2988, - "step": 3000 - }, - { - "epoch": 0.046952018468833845, - "grad_norm": 1.4161510467529297, - "learning_rate": 0.00019687090534722655, - "loss": 0.2984, - "step": 3010 - }, - { - "epoch": 0.04710800524115555, - "grad_norm": 4.282196521759033, - "learning_rate": 0.0001968605062290718, - "loss": 0.2822, - "step": 3020 - }, - { - "epoch": 0.04726399201347726, - "grad_norm": 2.5621113777160645, - "learning_rate": 0.000196850107110917, - "loss": 0.3886, - "step": 3030 - }, - { - "epoch": 0.04741997878579896, - "grad_norm": 0.7100577354431152, - "learning_rate": 0.0001968397079927622, - "loss": 0.2256, - "step": 3040 - }, - { - "epoch": 0.04757596555812067, - "grad_norm": 2.2435333728790283, - "learning_rate": 0.00019682930887460743, - "loss": 0.2155, - "step": 3050 - }, - { - "epoch": 0.047731952330442376, - "grad_norm": 1.9456003904342651, - "learning_rate": 0.00019681890975645265, - "loss": 0.2938, - "step": 3060 - }, - { - "epoch": 0.047887939102764086, - "grad_norm": 8.851899147033691, - "learning_rate": 0.00019680851063829787, - "loss": 0.3822, - "step": 3070 - }, - { - "epoch": 0.04804392587508579, - "grad_norm": 3.1571285724639893, - "learning_rate": 0.0001967981115201431, - "loss": 0.4913, - "step": 3080 - }, - { - "epoch": 0.0481999126474075, - "grad_norm": 6.932368755340576, - "learning_rate": 0.0001967877124019883, - "loss": 0.2643, - "step": 3090 - }, - { - "epoch": 0.048355899419729204, - "grad_norm": 4.671253204345703, - "learning_rate": 0.00019677731328383353, - "loss": 0.3736, - "step": 3100 - }, - { - "epoch": 0.048511886192050914, - "grad_norm": 4.238051891326904, - "learning_rate": 0.00019676691416567875, - "loss": 0.2905, - "step": 3110 - }, - { - "epoch": 0.048667872964372624, - "grad_norm": 4.185165882110596, - "learning_rate": 0.00019675651504752397, - "loss": 0.5799, - "step": 3120 - }, - { - "epoch": 0.04882385973669433, - "grad_norm": 5.000885486602783, - "learning_rate": 0.0001967461159293692, - "loss": 0.5353, - "step": 3130 - }, - { - "epoch": 0.04897984650901604, - "grad_norm": 2.1454551219940186, - "learning_rate": 0.0001967357168112144, - "loss": 0.2612, - "step": 3140 - }, - { - "epoch": 0.04913583328133774, - "grad_norm": 5.727521896362305, - "learning_rate": 0.00019672531769305963, - "loss": 0.3584, - "step": 3150 - }, - { - "epoch": 0.04929182005365945, - "grad_norm": 2.14788556098938, - "learning_rate": 0.00019671491857490485, - "loss": 0.248, - "step": 3160 - }, - { - "epoch": 0.049447806825981155, - "grad_norm": 5.505538463592529, - "learning_rate": 0.00019670451945675007, - "loss": 0.5042, - "step": 3170 - }, - { - "epoch": 0.049603793598302866, - "grad_norm": 0.45676565170288086, - "learning_rate": 0.0001966941203385953, - "loss": 0.3119, - "step": 3180 - }, - { - "epoch": 0.04975978037062457, - "grad_norm": 2.446408748626709, - "learning_rate": 0.0001966837212204405, - "loss": 0.1164, - "step": 3190 - }, - { - "epoch": 0.04991576714294628, - "grad_norm": 0.8342123031616211, - "learning_rate": 0.00019667332210228573, - "loss": 0.3749, - "step": 3200 - }, - { - "epoch": 0.05007175391526798, - "grad_norm": 3.0448403358459473, - "learning_rate": 0.00019666292298413095, - "loss": 0.4866, - "step": 3210 - }, - { - "epoch": 0.05022774068758969, - "grad_norm": 0.2389092594385147, - "learning_rate": 0.00019665252386597617, - "loss": 0.4043, - "step": 3220 - }, - { - "epoch": 0.0503837274599114, - "grad_norm": 5.156858921051025, - "learning_rate": 0.00019664212474782139, - "loss": 0.2329, - "step": 3230 - }, - { - "epoch": 0.05053971423223311, - "grad_norm": 3.149811029434204, - "learning_rate": 0.0001966317256296666, - "loss": 0.3312, - "step": 3240 - }, - { - "epoch": 0.05069570100455481, - "grad_norm": 0.45096033811569214, - "learning_rate": 0.00019662132651151185, - "loss": 0.3754, - "step": 3250 - }, - { - "epoch": 0.05085168777687652, - "grad_norm": 0.986663818359375, - "learning_rate": 0.00019661092739335704, - "loss": 0.3235, - "step": 3260 - }, - { - "epoch": 0.05100767454919823, - "grad_norm": 2.9946937561035156, - "learning_rate": 0.0001966005282752023, - "loss": 0.2073, - "step": 3270 - }, - { - "epoch": 0.051163661321519935, - "grad_norm": 3.995633125305176, - "learning_rate": 0.00019659012915704748, - "loss": 0.3481, - "step": 3280 - }, - { - "epoch": 0.051319648093841645, - "grad_norm": 1.7635934352874756, - "learning_rate": 0.00019657973003889273, - "loss": 0.1994, - "step": 3290 - }, - { - "epoch": 0.05147563486616335, - "grad_norm": 4.386302471160889, - "learning_rate": 0.00019656933092073792, - "loss": 0.3384, - "step": 3300 - }, - { - "epoch": 0.05163162163848506, - "grad_norm": 2.7013816833496094, - "learning_rate": 0.00019655893180258317, - "loss": 0.2823, - "step": 3310 - }, - { - "epoch": 0.05178760841080676, - "grad_norm": 7.045024871826172, - "learning_rate": 0.00019654853268442836, - "loss": 0.2112, - "step": 3320 - }, - { - "epoch": 0.05194359518312847, - "grad_norm": 2.587613105773926, - "learning_rate": 0.0001965381335662736, - "loss": 0.3959, - "step": 3330 - }, - { - "epoch": 0.052099581955450176, - "grad_norm": 0.9898849129676819, - "learning_rate": 0.0001965277344481188, - "loss": 0.3295, - "step": 3340 - }, - { - "epoch": 0.05225556872777189, - "grad_norm": 8.89372444152832, - "learning_rate": 0.00019651733532996405, - "loss": 0.3443, - "step": 3350 - }, - { - "epoch": 0.05241155550009359, - "grad_norm": 3.0839977264404297, - "learning_rate": 0.00019650693621180924, - "loss": 0.4376, - "step": 3360 - }, - { - "epoch": 0.0525675422724153, - "grad_norm": 5.862309455871582, - "learning_rate": 0.0001964965370936545, - "loss": 0.5446, - "step": 3370 - }, - { - "epoch": 0.052723529044737004, - "grad_norm": 5.274170398712158, - "learning_rate": 0.00019648613797549968, - "loss": 0.3987, - "step": 3380 - }, - { - "epoch": 0.052879515817058714, - "grad_norm": 4.794046878814697, - "learning_rate": 0.00019647573885734493, - "loss": 0.5267, - "step": 3390 - }, - { - "epoch": 0.05303550258938042, - "grad_norm": 1.6140331029891968, - "learning_rate": 0.00019646533973919012, - "loss": 0.6051, - "step": 3400 - }, - { - "epoch": 0.05319148936170213, - "grad_norm": 6.158959865570068, - "learning_rate": 0.00019645494062103537, - "loss": 0.3056, - "step": 3410 - }, - { - "epoch": 0.05334747613402383, - "grad_norm": 1.6302224397659302, - "learning_rate": 0.00019644454150288056, - "loss": 0.3473, - "step": 3420 - }, - { - "epoch": 0.05350346290634554, - "grad_norm": 4.881719589233398, - "learning_rate": 0.0001964341423847258, - "loss": 0.4032, - "step": 3430 - }, - { - "epoch": 0.05365944967866725, - "grad_norm": 1.1205174922943115, - "learning_rate": 0.000196423743266571, - "loss": 0.4288, - "step": 3440 - }, - { - "epoch": 0.053815436450988956, - "grad_norm": 6.742442607879639, - "learning_rate": 0.00019641334414841624, - "loss": 0.3299, - "step": 3450 - }, - { - "epoch": 0.053971423223310666, - "grad_norm": 0.43702423572540283, - "learning_rate": 0.00019640294503026144, - "loss": 0.4711, - "step": 3460 - }, - { - "epoch": 0.05412740999563237, - "grad_norm": 0.10137466341257095, - "learning_rate": 0.00019639254591210668, - "loss": 0.3355, - "step": 3470 - }, - { - "epoch": 0.05428339676795408, - "grad_norm": 3.7362818717956543, - "learning_rate": 0.00019638214679395188, - "loss": 0.3051, - "step": 3480 - }, - { - "epoch": 0.05443938354027578, - "grad_norm": 2.639213800430298, - "learning_rate": 0.0001963717476757971, - "loss": 0.3746, - "step": 3490 - }, - { - "epoch": 0.054595370312597494, - "grad_norm": 0.7321776151657104, - "learning_rate": 0.00019636134855764232, - "loss": 0.4572, - "step": 3500 - }, - { - "epoch": 0.0547513570849192, - "grad_norm": 1.803120732307434, - "learning_rate": 0.00019635094943948754, - "loss": 0.2899, - "step": 3510 - }, - { - "epoch": 0.05490734385724091, - "grad_norm": 0.9119880795478821, - "learning_rate": 0.00019634055032133276, - "loss": 0.2634, - "step": 3520 - }, - { - "epoch": 0.05506333062956261, - "grad_norm": 14.491175651550293, - "learning_rate": 0.00019633015120317797, - "loss": 0.4814, - "step": 3530 - }, - { - "epoch": 0.05521931740188432, - "grad_norm": 1.126238465309143, - "learning_rate": 0.0001963197520850232, - "loss": 0.3872, - "step": 3540 - }, - { - "epoch": 0.055375304174206025, - "grad_norm": 1.937894582748413, - "learning_rate": 0.00019630935296686841, - "loss": 0.4281, - "step": 3550 - }, - { - "epoch": 0.055531290946527735, - "grad_norm": 3.4557878971099854, - "learning_rate": 0.00019629895384871363, - "loss": 0.3678, - "step": 3560 - }, - { - "epoch": 0.05568727771884944, - "grad_norm": 0.7918877005577087, - "learning_rate": 0.00019628855473055885, - "loss": 0.1537, - "step": 3570 - }, - { - "epoch": 0.05584326449117115, - "grad_norm": 0.9775216579437256, - "learning_rate": 0.00019627815561240407, - "loss": 0.3411, - "step": 3580 - }, - { - "epoch": 0.05599925126349285, - "grad_norm": 3.5335514545440674, - "learning_rate": 0.0001962677564942493, - "loss": 0.4431, - "step": 3590 - }, - { - "epoch": 0.05615523803581456, - "grad_norm": 5.357288360595703, - "learning_rate": 0.0001962573573760945, - "loss": 0.5486, - "step": 3600 - }, - { - "epoch": 0.05631122480813627, - "grad_norm": 2.541792869567871, - "learning_rate": 0.00019624695825793973, - "loss": 0.2704, - "step": 3610 - }, - { - "epoch": 0.05646721158045798, - "grad_norm": 1.9300107955932617, - "learning_rate": 0.00019623655913978495, - "loss": 0.536, - "step": 3620 - }, - { - "epoch": 0.05662319835277969, - "grad_norm": 1.5524200201034546, - "learning_rate": 0.00019622616002163017, - "loss": 0.4837, - "step": 3630 - }, - { - "epoch": 0.05677918512510139, - "grad_norm": 4.605470180511475, - "learning_rate": 0.0001962157609034754, - "loss": 0.3935, - "step": 3640 - }, - { - "epoch": 0.0569351718974231, - "grad_norm": 1.5231540203094482, - "learning_rate": 0.0001962053617853206, - "loss": 0.289, - "step": 3650 - }, - { - "epoch": 0.057091158669744804, - "grad_norm": 1.3818199634552002, - "learning_rate": 0.00019619496266716583, - "loss": 0.3829, - "step": 3660 - }, - { - "epoch": 0.057247145442066515, - "grad_norm": 2.465237617492676, - "learning_rate": 0.00019618456354901105, - "loss": 0.4775, - "step": 3670 - }, - { - "epoch": 0.05740313221438822, - "grad_norm": 2.287116765975952, - "learning_rate": 0.00019617416443085627, - "loss": 0.6419, - "step": 3680 - }, - { - "epoch": 0.05755911898670993, - "grad_norm": 0.8104041218757629, - "learning_rate": 0.0001961637653127015, - "loss": 0.4125, - "step": 3690 - }, - { - "epoch": 0.05771510575903163, - "grad_norm": 6.393253803253174, - "learning_rate": 0.0001961533661945467, - "loss": 0.3983, - "step": 3700 - }, - { - "epoch": 0.05787109253135334, - "grad_norm": 1.9173061847686768, - "learning_rate": 0.00019614296707639193, - "loss": 0.3925, - "step": 3710 - }, - { - "epoch": 0.058027079303675046, - "grad_norm": 8.817895889282227, - "learning_rate": 0.00019613256795823715, - "loss": 0.6599, - "step": 3720 - }, - { - "epoch": 0.058183066075996756, - "grad_norm": 1.728727102279663, - "learning_rate": 0.00019612216884008237, - "loss": 0.4401, - "step": 3730 - }, - { - "epoch": 0.05833905284831846, - "grad_norm": 4.851232051849365, - "learning_rate": 0.0001961117697219276, - "loss": 0.1871, - "step": 3740 - }, - { - "epoch": 0.05849503962064017, - "grad_norm": 4.049133777618408, - "learning_rate": 0.0001961013706037728, - "loss": 0.3421, - "step": 3750 - }, - { - "epoch": 0.05865102639296188, - "grad_norm": 4.8147149085998535, - "learning_rate": 0.00019609097148561803, - "loss": 0.3399, - "step": 3760 - }, - { - "epoch": 0.058807013165283584, - "grad_norm": 2.2623703479766846, - "learning_rate": 0.00019608057236746325, - "loss": 0.3337, - "step": 3770 - }, - { - "epoch": 0.058962999937605294, - "grad_norm": 3.290314197540283, - "learning_rate": 0.00019607017324930847, - "loss": 0.3188, - "step": 3780 - }, - { - "epoch": 0.059118986709927, - "grad_norm": 6.715724945068359, - "learning_rate": 0.00019605977413115369, - "loss": 0.4027, - "step": 3790 - }, - { - "epoch": 0.05927497348224871, - "grad_norm": 0.1159973219037056, - "learning_rate": 0.0001960493750129989, - "loss": 0.3586, - "step": 3800 - }, - { - "epoch": 0.05943096025457041, - "grad_norm": 6.8083882331848145, - "learning_rate": 0.00019603897589484412, - "loss": 0.436, - "step": 3810 - }, - { - "epoch": 0.05958694702689212, - "grad_norm": 2.4548165798187256, - "learning_rate": 0.00019602857677668934, - "loss": 0.3744, - "step": 3820 - }, - { - "epoch": 0.059742933799213825, - "grad_norm": 7.0474629402160645, - "learning_rate": 0.00019601817765853456, - "loss": 0.4615, - "step": 3830 - }, - { - "epoch": 0.059898920571535536, - "grad_norm": 9.740936279296875, - "learning_rate": 0.00019600777854037978, - "loss": 0.3168, - "step": 3840 - }, - { - "epoch": 0.06005490734385724, - "grad_norm": 1.5505999326705933, - "learning_rate": 0.000195997379422225, - "loss": 0.2565, - "step": 3850 - }, - { - "epoch": 0.06021089411617895, - "grad_norm": 1.97008216381073, - "learning_rate": 0.00019598698030407022, - "loss": 0.4208, - "step": 3860 - }, - { - "epoch": 0.06036688088850065, - "grad_norm": 1.0527325868606567, - "learning_rate": 0.00019597658118591544, - "loss": 0.2885, - "step": 3870 - }, - { - "epoch": 0.06052286766082236, - "grad_norm": 3.9389073848724365, - "learning_rate": 0.00019596618206776066, - "loss": 0.3585, - "step": 3880 - }, - { - "epoch": 0.06067885443314407, - "grad_norm": 1.5215762853622437, - "learning_rate": 0.00019595578294960588, - "loss": 0.2961, - "step": 3890 - }, - { - "epoch": 0.06083484120546578, - "grad_norm": 3.200901985168457, - "learning_rate": 0.0001959453838314511, - "loss": 0.2842, - "step": 3900 - }, - { - "epoch": 0.06099082797778748, - "grad_norm": 3.8170034885406494, - "learning_rate": 0.00019593498471329632, - "loss": 0.3067, - "step": 3910 - }, - { - "epoch": 0.06114681475010919, - "grad_norm": 6.250919342041016, - "learning_rate": 0.00019592458559514154, - "loss": 0.3787, - "step": 3920 - }, - { - "epoch": 0.0613028015224309, - "grad_norm": 0.3816400468349457, - "learning_rate": 0.00019591418647698676, - "loss": 0.3611, - "step": 3930 - }, - { - "epoch": 0.061458788294752605, - "grad_norm": 0.548915445804596, - "learning_rate": 0.00019590378735883198, - "loss": 0.3225, - "step": 3940 - }, - { - "epoch": 0.061614775067074315, - "grad_norm": 3.9601895809173584, - "learning_rate": 0.0001958933882406772, - "loss": 0.3119, - "step": 3950 - }, - { - "epoch": 0.06177076183939602, - "grad_norm": 3.092454195022583, - "learning_rate": 0.00019588298912252242, - "loss": 0.2339, - "step": 3960 - }, - { - "epoch": 0.06192674861171773, - "grad_norm": 3.365994691848755, - "learning_rate": 0.00019587259000436764, - "loss": 0.3607, - "step": 3970 - }, - { - "epoch": 0.06208273538403943, - "grad_norm": 3.847018241882324, - "learning_rate": 0.00019586219088621286, - "loss": 0.2026, - "step": 3980 - }, - { - "epoch": 0.06223872215636114, - "grad_norm": 2.2113258838653564, - "learning_rate": 0.00019585179176805808, - "loss": 0.3062, - "step": 3990 - }, - { - "epoch": 0.062394708928682846, - "grad_norm": 2.4622628688812256, - "learning_rate": 0.0001958413926499033, - "loss": 0.1639, - "step": 4000 - }, - { - "epoch": 0.06255069570100455, - "grad_norm": 1.026604413986206, - "learning_rate": 0.00019583099353174852, - "loss": 0.3698, - "step": 4010 - }, - { - "epoch": 0.06270668247332627, - "grad_norm": 2.184748649597168, - "learning_rate": 0.00019582059441359374, - "loss": 0.4276, - "step": 4020 - }, - { - "epoch": 0.06286266924564797, - "grad_norm": 0.38206374645233154, - "learning_rate": 0.00019581019529543896, - "loss": 0.1884, - "step": 4030 - }, - { - "epoch": 0.06301865601796967, - "grad_norm": 3.477356195449829, - "learning_rate": 0.00019579979617728418, - "loss": 0.4619, - "step": 4040 - }, - { - "epoch": 0.06317464279029138, - "grad_norm": 1.1069543361663818, - "learning_rate": 0.0001957893970591294, - "loss": 0.5837, - "step": 4050 - }, - { - "epoch": 0.0633306295626131, - "grad_norm": 1.015366554260254, - "learning_rate": 0.00019577899794097462, - "loss": 0.2466, - "step": 4060 - }, - { - "epoch": 0.0634866163349348, - "grad_norm": 3.1485137939453125, - "learning_rate": 0.00019576859882281984, - "loss": 0.3343, - "step": 4070 - }, - { - "epoch": 0.0636426031072565, - "grad_norm": 1.8550643920898438, - "learning_rate": 0.00019575819970466505, - "loss": 0.2985, - "step": 4080 - }, - { - "epoch": 0.06379858987957822, - "grad_norm": 5.412409782409668, - "learning_rate": 0.00019574780058651027, - "loss": 0.212, - "step": 4090 - }, - { - "epoch": 0.06395457665189992, - "grad_norm": 6.8558526039123535, - "learning_rate": 0.0001957374014683555, - "loss": 0.7002, - "step": 4100 - }, - { - "epoch": 0.06411056342422163, - "grad_norm": 2.9232373237609863, - "learning_rate": 0.00019572700235020071, - "loss": 0.4317, - "step": 4110 - }, - { - "epoch": 0.06426655019654333, - "grad_norm": 2.8058922290802, - "learning_rate": 0.00019571660323204593, - "loss": 0.3056, - "step": 4120 - }, - { - "epoch": 0.06442253696886505, - "grad_norm": 1.4379881620407104, - "learning_rate": 0.00019570620411389115, - "loss": 0.2719, - "step": 4130 - }, - { - "epoch": 0.06457852374118675, - "grad_norm": 6.666982173919678, - "learning_rate": 0.00019569580499573637, - "loss": 0.2342, - "step": 4140 - }, - { - "epoch": 0.06473451051350845, - "grad_norm": 1.3467093706130981, - "learning_rate": 0.0001956854058775816, - "loss": 0.276, - "step": 4150 - }, - { - "epoch": 0.06489049728583016, - "grad_norm": 2.596644401550293, - "learning_rate": 0.0001956750067594268, - "loss": 0.4239, - "step": 4160 - }, - { - "epoch": 0.06504648405815187, - "grad_norm": 4.598370552062988, - "learning_rate": 0.00019566460764127203, - "loss": 0.4242, - "step": 4170 - }, - { - "epoch": 0.06520247083047358, - "grad_norm": 5.2616424560546875, - "learning_rate": 0.00019565420852311725, - "loss": 0.2431, - "step": 4180 - }, - { - "epoch": 0.06535845760279528, - "grad_norm": 1.9798760414123535, - "learning_rate": 0.00019564380940496247, - "loss": 0.6246, - "step": 4190 - }, - { - "epoch": 0.06551444437511698, - "grad_norm": 1.605912685394287, - "learning_rate": 0.0001956334102868077, - "loss": 0.4617, - "step": 4200 - }, - { - "epoch": 0.0656704311474387, - "grad_norm": 3.7138144969940186, - "learning_rate": 0.0001956230111686529, - "loss": 0.5119, - "step": 4210 - }, - { - "epoch": 0.0658264179197604, - "grad_norm": 2.1792988777160645, - "learning_rate": 0.00019561261205049813, - "loss": 0.4112, - "step": 4220 - }, - { - "epoch": 0.06598240469208211, - "grad_norm": 2.197249174118042, - "learning_rate": 0.00019560221293234335, - "loss": 0.4402, - "step": 4230 - }, - { - "epoch": 0.06613839146440381, - "grad_norm": 2.6887547969818115, - "learning_rate": 0.00019559181381418857, - "loss": 0.2802, - "step": 4240 - }, - { - "epoch": 0.06629437823672553, - "grad_norm": 6.97922945022583, - "learning_rate": 0.0001955814146960338, - "loss": 0.3786, - "step": 4250 - }, - { - "epoch": 0.06645036500904723, - "grad_norm": 2.674954414367676, - "learning_rate": 0.000195571015577879, - "loss": 0.3124, - "step": 4260 - }, - { - "epoch": 0.06660635178136894, - "grad_norm": 1.5316659212112427, - "learning_rate": 0.00019556061645972423, - "loss": 0.2674, - "step": 4270 - }, - { - "epoch": 0.06676233855369065, - "grad_norm": 3.402186393737793, - "learning_rate": 0.00019555021734156945, - "loss": 0.1722, - "step": 4280 - }, - { - "epoch": 0.06691832532601236, - "grad_norm": 4.810988903045654, - "learning_rate": 0.00019553981822341467, - "loss": 0.4486, - "step": 4290 - }, - { - "epoch": 0.06707431209833406, - "grad_norm": 4.834427833557129, - "learning_rate": 0.0001955294191052599, - "loss": 0.3412, - "step": 4300 - }, - { - "epoch": 0.06723029887065576, - "grad_norm": 1.7941811084747314, - "learning_rate": 0.0001955190199871051, - "loss": 0.2732, - "step": 4310 - }, - { - "epoch": 0.06738628564297748, - "grad_norm": 4.941607475280762, - "learning_rate": 0.00019550862086895033, - "loss": 0.3954, - "step": 4320 - }, - { - "epoch": 0.06754227241529918, - "grad_norm": 1.0881506204605103, - "learning_rate": 0.00019549822175079555, - "loss": 0.3729, - "step": 4330 - }, - { - "epoch": 0.06769825918762089, - "grad_norm": 2.2694144248962402, - "learning_rate": 0.00019548782263264077, - "loss": 0.4538, - "step": 4340 - }, - { - "epoch": 0.06785424595994259, - "grad_norm": 2.897859573364258, - "learning_rate": 0.00019547742351448599, - "loss": 0.3101, - "step": 4350 - }, - { - "epoch": 0.06801023273226431, - "grad_norm": 2.17456316947937, - "learning_rate": 0.0001954670243963312, - "loss": 0.5781, - "step": 4360 - }, - { - "epoch": 0.06816621950458601, - "grad_norm": 5.235560417175293, - "learning_rate": 0.00019545662527817642, - "loss": 0.3683, - "step": 4370 - }, - { - "epoch": 0.06832220627690772, - "grad_norm": 3.0128090381622314, - "learning_rate": 0.00019544622616002164, - "loss": 0.2558, - "step": 4380 - }, - { - "epoch": 0.06847819304922942, - "grad_norm": 1.6238577365875244, - "learning_rate": 0.00019543582704186684, - "loss": 0.3038, - "step": 4390 - }, - { - "epoch": 0.06863417982155114, - "grad_norm": 2.9384729862213135, - "learning_rate": 0.00019542542792371208, - "loss": 0.3866, - "step": 4400 - }, - { - "epoch": 0.06879016659387284, - "grad_norm": 0.7549554109573364, - "learning_rate": 0.00019541502880555728, - "loss": 0.2905, - "step": 4410 - }, - { - "epoch": 0.06894615336619454, - "grad_norm": 3.4662294387817383, - "learning_rate": 0.00019540462968740252, - "loss": 0.257, - "step": 4420 - }, - { - "epoch": 0.06910214013851626, - "grad_norm": 4.433900356292725, - "learning_rate": 0.00019539423056924772, - "loss": 0.4666, - "step": 4430 - }, - { - "epoch": 0.06925812691083796, - "grad_norm": 2.6317667961120605, - "learning_rate": 0.00019538383145109296, - "loss": 0.2396, - "step": 4440 - }, - { - "epoch": 0.06941411368315967, - "grad_norm": 0.9813170433044434, - "learning_rate": 0.00019537343233293815, - "loss": 0.2148, - "step": 4450 - }, - { - "epoch": 0.06957010045548137, - "grad_norm": 2.8227760791778564, - "learning_rate": 0.0001953630332147834, - "loss": 0.6245, - "step": 4460 - }, - { - "epoch": 0.06972608722780309, - "grad_norm": 1.489005446434021, - "learning_rate": 0.0001953526340966286, - "loss": 0.5559, - "step": 4470 - }, - { - "epoch": 0.06988207400012479, - "grad_norm": 1.7438796758651733, - "learning_rate": 0.00019534223497847384, - "loss": 0.2688, - "step": 4480 - }, - { - "epoch": 0.0700380607724465, - "grad_norm": 1.6185518503189087, - "learning_rate": 0.00019533183586031903, - "loss": 0.386, - "step": 4490 - }, - { - "epoch": 0.0701940475447682, - "grad_norm": 2.74615740776062, - "learning_rate": 0.00019532143674216428, - "loss": 0.4799, - "step": 4500 - }, - { - "epoch": 0.07035003431708992, - "grad_norm": 4.399687767028809, - "learning_rate": 0.00019531103762400947, - "loss": 0.4494, - "step": 4510 - }, - { - "epoch": 0.07050602108941162, - "grad_norm": 2.602672815322876, - "learning_rate": 0.00019530063850585472, - "loss": 0.2798, - "step": 4520 - }, - { - "epoch": 0.07066200786173332, - "grad_norm": 8.301713943481445, - "learning_rate": 0.0001952902393876999, - "loss": 0.3375, - "step": 4530 - }, - { - "epoch": 0.07081799463405503, - "grad_norm": 4.576874732971191, - "learning_rate": 0.00019527984026954516, - "loss": 0.3633, - "step": 4540 - }, - { - "epoch": 0.07097398140637674, - "grad_norm": 3.145695447921753, - "learning_rate": 0.00019526944115139035, - "loss": 0.4248, - "step": 4550 - }, - { - "epoch": 0.07112996817869845, - "grad_norm": 2.894092082977295, - "learning_rate": 0.0001952590420332356, - "loss": 0.3215, - "step": 4560 - }, - { - "epoch": 0.07128595495102015, - "grad_norm": 0.13097500801086426, - "learning_rate": 0.00019524864291508082, - "loss": 0.3269, - "step": 4570 - }, - { - "epoch": 0.07144194172334187, - "grad_norm": 5.0817341804504395, - "learning_rate": 0.00019523824379692604, - "loss": 0.4399, - "step": 4580 - }, - { - "epoch": 0.07159792849566357, - "grad_norm": 1.6493229866027832, - "learning_rate": 0.00019522784467877126, - "loss": 0.3025, - "step": 4590 - }, - { - "epoch": 0.07175391526798527, - "grad_norm": 6.487491130828857, - "learning_rate": 0.00019521744556061648, - "loss": 0.5133, - "step": 4600 - }, - { - "epoch": 0.07190990204030698, - "grad_norm": 4.299659252166748, - "learning_rate": 0.0001952070464424617, - "loss": 1.2018, - "step": 4610 - }, - { - "epoch": 0.0720658888126287, - "grad_norm": 0.8001397252082825, - "learning_rate": 0.00019519664732430692, - "loss": 0.3898, - "step": 4620 - }, - { - "epoch": 0.0722218755849504, - "grad_norm": 3.345963716506958, - "learning_rate": 0.00019518624820615214, - "loss": 0.3375, - "step": 4630 - }, - { - "epoch": 0.0723778623572721, - "grad_norm": 1.9037283658981323, - "learning_rate": 0.00019517584908799735, - "loss": 0.4469, - "step": 4640 - }, - { - "epoch": 0.0725338491295938, - "grad_norm": 0.64785236120224, - "learning_rate": 0.00019516544996984257, - "loss": 0.2919, - "step": 4650 - }, - { - "epoch": 0.07268983590191552, - "grad_norm": 4.346297740936279, - "learning_rate": 0.0001951550508516878, - "loss": 0.2624, - "step": 4660 - }, - { - "epoch": 0.07284582267423723, - "grad_norm": 1.2748740911483765, - "learning_rate": 0.00019514465173353301, - "loss": 0.2201, - "step": 4670 - }, - { - "epoch": 0.07300180944655893, - "grad_norm": 2.5817744731903076, - "learning_rate": 0.00019513425261537823, - "loss": 0.3252, - "step": 4680 - }, - { - "epoch": 0.07315779621888063, - "grad_norm": 1.5413234233856201, - "learning_rate": 0.00019512385349722345, - "loss": 0.4234, - "step": 4690 - }, - { - "epoch": 0.07331378299120235, - "grad_norm": 1.6661428213119507, - "learning_rate": 0.00019511345437906867, - "loss": 0.5214, - "step": 4700 - }, - { - "epoch": 0.07346976976352405, - "grad_norm": 4.948998928070068, - "learning_rate": 0.0001951030552609139, - "loss": 0.5428, - "step": 4710 - }, - { - "epoch": 0.07362575653584576, - "grad_norm": 1.4078203439712524, - "learning_rate": 0.0001950926561427591, - "loss": 0.4, - "step": 4720 - }, - { - "epoch": 0.07378174330816746, - "grad_norm": 0.6059741973876953, - "learning_rate": 0.00019508225702460433, - "loss": 0.5198, - "step": 4730 - }, - { - "epoch": 0.07393773008048918, - "grad_norm": 0.24869346618652344, - "learning_rate": 0.00019507185790644955, - "loss": 0.3058, - "step": 4740 - }, - { - "epoch": 0.07409371685281088, - "grad_norm": 1.6327784061431885, - "learning_rate": 0.00019506145878829477, - "loss": 1.5225, - "step": 4750 - }, - { - "epoch": 0.07424970362513258, - "grad_norm": 3.5642881393432617, - "learning_rate": 0.00019505105967014, - "loss": 0.387, - "step": 4760 - }, - { - "epoch": 0.0744056903974543, - "grad_norm": 3.4709348678588867, - "learning_rate": 0.0001950406605519852, - "loss": 0.4065, - "step": 4770 - }, - { - "epoch": 0.074561677169776, - "grad_norm": 5.526008129119873, - "learning_rate": 0.00019503026143383043, - "loss": 0.4199, - "step": 4780 - }, - { - "epoch": 0.07471766394209771, - "grad_norm": 7.353578567504883, - "learning_rate": 0.00019501986231567565, - "loss": 0.2635, - "step": 4790 - }, - { - "epoch": 0.07487365071441941, - "grad_norm": 4.601809978485107, - "learning_rate": 0.00019500946319752087, - "loss": 0.4241, - "step": 4800 - }, - { - "epoch": 0.07502963748674113, - "grad_norm": 5.431839466094971, - "learning_rate": 0.0001949990640793661, - "loss": 0.2799, - "step": 4810 - }, - { - "epoch": 0.07518562425906283, - "grad_norm": 3.1551904678344727, - "learning_rate": 0.0001949886649612113, - "loss": 0.4654, - "step": 4820 - }, - { - "epoch": 0.07534161103138454, - "grad_norm": 1.8644407987594604, - "learning_rate": 0.00019497826584305653, - "loss": 0.3813, - "step": 4830 - }, - { - "epoch": 0.07549759780370624, - "grad_norm": 1.2464070320129395, - "learning_rate": 0.00019496786672490175, - "loss": 0.3293, - "step": 4840 - }, - { - "epoch": 0.07565358457602796, - "grad_norm": 3.8643407821655273, - "learning_rate": 0.00019495746760674697, - "loss": 0.5776, - "step": 4850 - }, - { - "epoch": 0.07580957134834966, - "grad_norm": 0.7169101238250732, - "learning_rate": 0.00019494706848859216, - "loss": 0.3103, - "step": 4860 - }, - { - "epoch": 0.07596555812067136, - "grad_norm": 0.9111189246177673, - "learning_rate": 0.0001949366693704374, - "loss": 0.5543, - "step": 4870 - }, - { - "epoch": 0.07612154489299307, - "grad_norm": 5.50398063659668, - "learning_rate": 0.0001949262702522826, - "loss": 0.5258, - "step": 4880 - }, - { - "epoch": 0.07627753166531479, - "grad_norm": 4.419070243835449, - "learning_rate": 0.00019491587113412785, - "loss": 0.6176, - "step": 4890 - }, - { - "epoch": 0.07643351843763649, - "grad_norm": 1.6785311698913574, - "learning_rate": 0.00019490547201597304, - "loss": 0.2027, - "step": 4900 - }, - { - "epoch": 0.07658950520995819, - "grad_norm": 2.111905097961426, - "learning_rate": 0.00019489507289781829, - "loss": 0.193, - "step": 4910 - }, - { - "epoch": 0.07674549198227991, - "grad_norm": 2.166006326675415, - "learning_rate": 0.00019488467377966348, - "loss": 0.1709, - "step": 4920 - }, - { - "epoch": 0.07690147875460161, - "grad_norm": 3.8289716243743896, - "learning_rate": 0.00019487427466150872, - "loss": 0.1576, - "step": 4930 - }, - { - "epoch": 0.07705746552692332, - "grad_norm": 2.0966336727142334, - "learning_rate": 0.00019486387554335392, - "loss": 0.628, - "step": 4940 - }, - { - "epoch": 0.07721345229924502, - "grad_norm": 3.5698137283325195, - "learning_rate": 0.00019485347642519916, - "loss": 0.3256, - "step": 4950 - }, - { - "epoch": 0.07736943907156674, - "grad_norm": 2.1274750232696533, - "learning_rate": 0.00019484307730704436, - "loss": 0.3941, - "step": 4960 - }, - { - "epoch": 0.07752542584388844, - "grad_norm": 4.426887512207031, - "learning_rate": 0.0001948326781888896, - "loss": 0.4366, - "step": 4970 - }, - { - "epoch": 0.07768141261621014, - "grad_norm": 5.6438307762146, - "learning_rate": 0.0001948222790707348, - "loss": 0.2416, - "step": 4980 - }, - { - "epoch": 0.07783739938853185, - "grad_norm": 4.310532093048096, - "learning_rate": 0.00019481187995258004, - "loss": 0.4948, - "step": 4990 - }, - { - "epoch": 0.07799338616085356, - "grad_norm": 1.9643551111221313, - "learning_rate": 0.00019480148083442523, - "loss": 0.2636, - "step": 5000 - }, - { - "epoch": 0.07814937293317527, - "grad_norm": 2.5053608417510986, - "learning_rate": 0.00019479108171627048, - "loss": 0.2981, - "step": 5010 - }, - { - "epoch": 0.07830535970549697, - "grad_norm": 1.290313482284546, - "learning_rate": 0.00019478068259811567, - "loss": 0.2891, - "step": 5020 - }, - { - "epoch": 0.07846134647781867, - "grad_norm": 0.16381753981113434, - "learning_rate": 0.00019477028347996092, - "loss": 0.3624, - "step": 5030 - }, - { - "epoch": 0.07861733325014039, - "grad_norm": 3.1533336639404297, - "learning_rate": 0.0001947598843618061, - "loss": 0.2191, - "step": 5040 - }, - { - "epoch": 0.0787733200224621, - "grad_norm": 0.9760684370994568, - "learning_rate": 0.00019474948524365136, - "loss": 0.464, - "step": 5050 - }, - { - "epoch": 0.0789293067947838, - "grad_norm": 2.2214372158050537, - "learning_rate": 0.00019473908612549655, - "loss": 0.4851, - "step": 5060 - }, - { - "epoch": 0.07908529356710552, - "grad_norm": 1.9488602876663208, - "learning_rate": 0.0001947286870073418, - "loss": 0.3167, - "step": 5070 - }, - { - "epoch": 0.07924128033942722, - "grad_norm": 3.411989212036133, - "learning_rate": 0.000194718287889187, - "loss": 0.3631, - "step": 5080 - }, - { - "epoch": 0.07939726711174892, - "grad_norm": 3.581641912460327, - "learning_rate": 0.00019470788877103224, - "loss": 0.1301, - "step": 5090 - }, - { - "epoch": 0.07955325388407063, - "grad_norm": 3.8233249187469482, - "learning_rate": 0.00019469748965287743, - "loss": 0.277, - "step": 5100 - }, - { - "epoch": 0.07970924065639234, - "grad_norm": 8.081124305725098, - "learning_rate": 0.00019468709053472268, - "loss": 0.2866, - "step": 5110 - }, - { - "epoch": 0.07986522742871405, - "grad_norm": 3.1644890308380127, - "learning_rate": 0.00019467669141656787, - "loss": 0.3535, - "step": 5120 - }, - { - "epoch": 0.08002121420103575, - "grad_norm": 0.956291913986206, - "learning_rate": 0.00019466629229841312, - "loss": 0.4388, - "step": 5130 - }, - { - "epoch": 0.08017720097335745, - "grad_norm": 2.458174467086792, - "learning_rate": 0.0001946558931802583, - "loss": 0.321, - "step": 5140 - }, - { - "epoch": 0.08033318774567917, - "grad_norm": 1.813386082649231, - "learning_rate": 0.00019464549406210356, - "loss": 0.4277, - "step": 5150 - }, - { - "epoch": 0.08048917451800087, - "grad_norm": 0.2468508780002594, - "learning_rate": 0.00019463509494394875, - "loss": 0.3927, - "step": 5160 - }, - { - "epoch": 0.08064516129032258, - "grad_norm": 0.5065591931343079, - "learning_rate": 0.000194624695825794, - "loss": 0.175, - "step": 5170 - }, - { - "epoch": 0.08080114806264428, - "grad_norm": 3.965261697769165, - "learning_rate": 0.0001946142967076392, - "loss": 0.2932, - "step": 5180 - }, - { - "epoch": 0.080957134834966, - "grad_norm": 8.131027221679688, - "learning_rate": 0.00019460389758948444, - "loss": 0.3341, - "step": 5190 - }, - { - "epoch": 0.0811131216072877, - "grad_norm": 3.0608482360839844, - "learning_rate": 0.00019459349847132963, - "loss": 0.337, - "step": 5200 - }, - { - "epoch": 0.0812691083796094, - "grad_norm": 1.2423006296157837, - "learning_rate": 0.00019458309935317487, - "loss": 0.439, - "step": 5210 - }, - { - "epoch": 0.08142509515193111, - "grad_norm": 2.6871418952941895, - "learning_rate": 0.00019457270023502007, - "loss": 0.3247, - "step": 5220 - }, - { - "epoch": 0.08158108192425283, - "grad_norm": 3.4826762676239014, - "learning_rate": 0.00019456230111686531, - "loss": 0.5614, - "step": 5230 - }, - { - "epoch": 0.08173706869657453, - "grad_norm": 1.9816076755523682, - "learning_rate": 0.0001945519019987105, - "loss": 0.5613, - "step": 5240 - }, - { - "epoch": 0.08189305546889623, - "grad_norm": 5.499810218811035, - "learning_rate": 0.00019454150288055575, - "loss": 0.4791, - "step": 5250 - }, - { - "epoch": 0.08204904224121795, - "grad_norm": 3.7328336238861084, - "learning_rate": 0.00019453110376240095, - "loss": 0.2752, - "step": 5260 - }, - { - "epoch": 0.08220502901353965, - "grad_norm": 1.268531084060669, - "learning_rate": 0.0001945207046442462, - "loss": 0.2227, - "step": 5270 - }, - { - "epoch": 0.08236101578586136, - "grad_norm": 4.293695449829102, - "learning_rate": 0.00019451030552609138, - "loss": 0.6035, - "step": 5280 - }, - { - "epoch": 0.08251700255818306, - "grad_norm": 3.009061574935913, - "learning_rate": 0.00019449990640793663, - "loss": 0.4032, - "step": 5290 - }, - { - "epoch": 0.08267298933050478, - "grad_norm": 2.217053174972534, - "learning_rate": 0.00019448950728978182, - "loss": 0.4399, - "step": 5300 - }, - { - "epoch": 0.08282897610282648, - "grad_norm": 2.3602845668792725, - "learning_rate": 0.00019447910817162704, - "loss": 0.6428, - "step": 5310 - }, - { - "epoch": 0.08298496287514819, - "grad_norm": 11.570981979370117, - "learning_rate": 0.00019446870905347226, - "loss": 0.4949, - "step": 5320 - }, - { - "epoch": 0.08314094964746989, - "grad_norm": 4.000115394592285, - "learning_rate": 0.00019445830993531748, - "loss": 0.7965, - "step": 5330 - }, - { - "epoch": 0.0832969364197916, - "grad_norm": 0.9110697507858276, - "learning_rate": 0.0001944479108171627, - "loss": 0.2803, - "step": 5340 - }, - { - "epoch": 0.08345292319211331, - "grad_norm": 2.672752618789673, - "learning_rate": 0.00019443751169900792, - "loss": 0.2366, - "step": 5350 - }, - { - "epoch": 0.08360890996443501, - "grad_norm": 4.964208126068115, - "learning_rate": 0.00019442711258085314, - "loss": 0.3474, - "step": 5360 - }, - { - "epoch": 0.08376489673675672, - "grad_norm": 1.1511766910552979, - "learning_rate": 0.00019441671346269836, - "loss": 0.2477, - "step": 5370 - }, - { - "epoch": 0.08392088350907843, - "grad_norm": 4.001786708831787, - "learning_rate": 0.00019440631434454358, - "loss": 0.349, - "step": 5380 - }, - { - "epoch": 0.08407687028140014, - "grad_norm": 1.4430139064788818, - "learning_rate": 0.0001943959152263888, - "loss": 0.4473, - "step": 5390 - }, - { - "epoch": 0.08423285705372184, - "grad_norm": 2.817662239074707, - "learning_rate": 0.00019438551610823402, - "loss": 1.5236, - "step": 5400 - }, - { - "epoch": 0.08438884382604356, - "grad_norm": 1.8846608400344849, - "learning_rate": 0.00019437511699007924, - "loss": 0.3512, - "step": 5410 - }, - { - "epoch": 0.08454483059836526, - "grad_norm": 0.6527857780456543, - "learning_rate": 0.00019436471787192446, - "loss": 0.3563, - "step": 5420 - }, - { - "epoch": 0.08470081737068696, - "grad_norm": 3.056812047958374, - "learning_rate": 0.00019435431875376968, - "loss": 0.3094, - "step": 5430 - }, - { - "epoch": 0.08485680414300867, - "grad_norm": 1.3599086999893188, - "learning_rate": 0.00019434391963561493, - "loss": 0.3397, - "step": 5440 - }, - { - "epoch": 0.08501279091533039, - "grad_norm": 6.885451316833496, - "learning_rate": 0.00019433352051746012, - "loss": 0.3929, - "step": 5450 - }, - { - "epoch": 0.08516877768765209, - "grad_norm": 0.5930185317993164, - "learning_rate": 0.00019432312139930537, - "loss": 0.4468, - "step": 5460 - }, - { - "epoch": 0.08532476445997379, - "grad_norm": 1.0316463708877563, - "learning_rate": 0.00019431272228115056, - "loss": 0.3909, - "step": 5470 - }, - { - "epoch": 0.0854807512322955, - "grad_norm": 1.9080007076263428, - "learning_rate": 0.0001943023231629958, - "loss": 0.4185, - "step": 5480 - }, - { - "epoch": 0.08563673800461721, - "grad_norm": 1.182185173034668, - "learning_rate": 0.000194291924044841, - "loss": 0.3496, - "step": 5490 - }, - { - "epoch": 0.08579272477693892, - "grad_norm": 1.7894487380981445, - "learning_rate": 0.00019428152492668624, - "loss": 0.2275, - "step": 5500 - }, - { - "epoch": 0.08594871154926062, - "grad_norm": 7.630679130554199, - "learning_rate": 0.00019427112580853144, - "loss": 0.3358, - "step": 5510 - }, - { - "epoch": 0.08610469832158232, - "grad_norm": 4.102541923522949, - "learning_rate": 0.00019426072669037668, - "loss": 0.3494, - "step": 5520 - }, - { - "epoch": 0.08626068509390404, - "grad_norm": 2.618604898452759, - "learning_rate": 0.00019425032757222188, - "loss": 0.2152, - "step": 5530 - }, - { - "epoch": 0.08641667186622574, - "grad_norm": 7.212046146392822, - "learning_rate": 0.00019423992845406712, - "loss": 0.4545, - "step": 5540 - }, - { - "epoch": 0.08657265863854745, - "grad_norm": 4.502930164337158, - "learning_rate": 0.00019422952933591232, - "loss": 0.484, - "step": 5550 - }, - { - "epoch": 0.08672864541086917, - "grad_norm": 1.1972264051437378, - "learning_rate": 0.00019421913021775756, - "loss": 0.2682, - "step": 5560 - }, - { - "epoch": 0.08688463218319087, - "grad_norm": 4.042606830596924, - "learning_rate": 0.00019420873109960275, - "loss": 0.299, - "step": 5570 - }, - { - "epoch": 0.08704061895551257, - "grad_norm": 1.233870506286621, - "learning_rate": 0.000194198331981448, - "loss": 0.2416, - "step": 5580 - }, - { - "epoch": 0.08719660572783428, - "grad_norm": 2.3368418216705322, - "learning_rate": 0.0001941879328632932, - "loss": 0.3265, - "step": 5590 - }, - { - "epoch": 0.08735259250015599, - "grad_norm": 1.1264872550964355, - "learning_rate": 0.00019417753374513844, - "loss": 0.3293, - "step": 5600 - }, - { - "epoch": 0.0875085792724777, - "grad_norm": 0.9745358228683472, - "learning_rate": 0.00019416713462698363, - "loss": 0.3, - "step": 5610 - }, - { - "epoch": 0.0876645660447994, - "grad_norm": 5.301463603973389, - "learning_rate": 0.00019415673550882888, - "loss": 0.5055, - "step": 5620 - }, - { - "epoch": 0.0878205528171211, - "grad_norm": 2.5696167945861816, - "learning_rate": 0.00019414633639067407, - "loss": 0.3489, - "step": 5630 - }, - { - "epoch": 0.08797653958944282, - "grad_norm": 1.5095734596252441, - "learning_rate": 0.00019413593727251932, - "loss": 0.4118, - "step": 5640 - }, - { - "epoch": 0.08813252636176452, - "grad_norm": 2.604001522064209, - "learning_rate": 0.0001941255381543645, - "loss": 0.2596, - "step": 5650 - }, - { - "epoch": 0.08828851313408623, - "grad_norm": 4.1579365730285645, - "learning_rate": 0.00019411513903620976, - "loss": 0.2804, - "step": 5660 - }, - { - "epoch": 0.08844449990640793, - "grad_norm": 1.092995524406433, - "learning_rate": 0.00019410473991805495, - "loss": 0.524, - "step": 5670 - }, - { - "epoch": 0.08860048667872965, - "grad_norm": 3.346893787384033, - "learning_rate": 0.0001940943407999002, - "loss": 0.3079, - "step": 5680 - }, - { - "epoch": 0.08875647345105135, - "grad_norm": 5.878612995147705, - "learning_rate": 0.0001940839416817454, - "loss": 0.5345, - "step": 5690 - }, - { - "epoch": 0.08891246022337305, - "grad_norm": 5.466742515563965, - "learning_rate": 0.00019407354256359064, - "loss": 0.2776, - "step": 5700 - }, - { - "epoch": 0.08906844699569476, - "grad_norm": 1.969839096069336, - "learning_rate": 0.00019406314344543583, - "loss": 0.1536, - "step": 5710 - }, - { - "epoch": 0.08922443376801648, - "grad_norm": 0.32240307331085205, - "learning_rate": 0.00019405274432728108, - "loss": 0.0561, - "step": 5720 - }, - { - "epoch": 0.08938042054033818, - "grad_norm": 2.4503121376037598, - "learning_rate": 0.00019404234520912627, - "loss": 0.457, - "step": 5730 - }, - { - "epoch": 0.08953640731265988, - "grad_norm": 2.636259078979492, - "learning_rate": 0.00019403194609097152, - "loss": 0.2433, - "step": 5740 - }, - { - "epoch": 0.0896923940849816, - "grad_norm": 1.139318585395813, - "learning_rate": 0.0001940215469728167, - "loss": 0.1501, - "step": 5750 - }, - { - "epoch": 0.0898483808573033, - "grad_norm": 3.3219454288482666, - "learning_rate": 0.00019401114785466193, - "loss": 0.4036, - "step": 5760 - }, - { - "epoch": 0.090004367629625, - "grad_norm": 5.504943370819092, - "learning_rate": 0.00019400074873650715, - "loss": 0.3269, - "step": 5770 - }, - { - "epoch": 0.09016035440194671, - "grad_norm": 0.14088541269302368, - "learning_rate": 0.00019399034961835237, - "loss": 0.4199, - "step": 5780 - }, - { - "epoch": 0.09031634117426843, - "grad_norm": 2.2649993896484375, - "learning_rate": 0.00019397995050019759, - "loss": 0.5362, - "step": 5790 - }, - { - "epoch": 0.09047232794659013, - "grad_norm": 0.7106287479400635, - "learning_rate": 0.0001939695513820428, - "loss": 0.2285, - "step": 5800 - }, - { - "epoch": 0.09062831471891183, - "grad_norm": 5.814300060272217, - "learning_rate": 0.00019395915226388803, - "loss": 0.5139, - "step": 5810 - }, - { - "epoch": 0.09078430149123354, - "grad_norm": 4.537754058837891, - "learning_rate": 0.00019394875314573325, - "loss": 0.2923, - "step": 5820 - }, - { - "epoch": 0.09094028826355526, - "grad_norm": 4.533511638641357, - "learning_rate": 0.00019393835402757846, - "loss": 0.2571, - "step": 5830 - }, - { - "epoch": 0.09109627503587696, - "grad_norm": 1.4505037069320679, - "learning_rate": 0.00019392795490942368, - "loss": 0.2398, - "step": 5840 - }, - { - "epoch": 0.09125226180819866, - "grad_norm": 2.154186248779297, - "learning_rate": 0.0001939175557912689, - "loss": 0.2248, - "step": 5850 - }, - { - "epoch": 0.09140824858052037, - "grad_norm": 1.9082036018371582, - "learning_rate": 0.00019390715667311412, - "loss": 0.4279, - "step": 5860 - }, - { - "epoch": 0.09156423535284208, - "grad_norm": 3.025155782699585, - "learning_rate": 0.00019389675755495934, - "loss": 0.322, - "step": 5870 - }, - { - "epoch": 0.09172022212516379, - "grad_norm": 3.047150135040283, - "learning_rate": 0.00019388635843680456, - "loss": 0.3582, - "step": 5880 - }, - { - "epoch": 0.09187620889748549, - "grad_norm": 2.638498067855835, - "learning_rate": 0.00019387595931864978, - "loss": 0.3079, - "step": 5890 - }, - { - "epoch": 0.0920321956698072, - "grad_norm": 0.13002626597881317, - "learning_rate": 0.000193865560200495, - "loss": 0.3355, - "step": 5900 - }, - { - "epoch": 0.09218818244212891, - "grad_norm": 5.473140716552734, - "learning_rate": 0.00019385516108234022, - "loss": 0.4944, - "step": 5910 - }, - { - "epoch": 0.09234416921445061, - "grad_norm": 1.5157524347305298, - "learning_rate": 0.00019384476196418544, - "loss": 0.2071, - "step": 5920 - }, - { - "epoch": 0.09250015598677232, - "grad_norm": 1.8442835807800293, - "learning_rate": 0.00019383436284603066, - "loss": 0.2215, - "step": 5930 - }, - { - "epoch": 0.09265614275909403, - "grad_norm": 2.5450761318206787, - "learning_rate": 0.00019382396372787588, - "loss": 0.3973, - "step": 5940 - }, - { - "epoch": 0.09281212953141574, - "grad_norm": 0.356424480676651, - "learning_rate": 0.0001938135646097211, - "loss": 0.2876, - "step": 5950 - }, - { - "epoch": 0.09296811630373744, - "grad_norm": 4.990633487701416, - "learning_rate": 0.00019380316549156632, - "loss": 0.1775, - "step": 5960 - }, - { - "epoch": 0.09312410307605914, - "grad_norm": 8.785608291625977, - "learning_rate": 0.00019379276637341154, - "loss": 0.4081, - "step": 5970 - }, - { - "epoch": 0.09328008984838086, - "grad_norm": 1.021690845489502, - "learning_rate": 0.00019378236725525676, - "loss": 0.1727, - "step": 5980 - }, - { - "epoch": 0.09343607662070257, - "grad_norm": 0.6903918385505676, - "learning_rate": 0.00019377196813710198, - "loss": 0.4199, - "step": 5990 - }, - { - "epoch": 0.09359206339302427, - "grad_norm": 6.114228248596191, - "learning_rate": 0.0001937615690189472, - "loss": 0.2502, - "step": 6000 - }, - { - "epoch": 0.09374805016534597, - "grad_norm": 4.526939868927002, - "learning_rate": 0.00019375116990079242, - "loss": 0.3677, - "step": 6010 - }, - { - "epoch": 0.09390403693766769, - "grad_norm": 3.766489028930664, - "learning_rate": 0.00019374077078263764, - "loss": 0.4365, - "step": 6020 - }, - { - "epoch": 0.0940600237099894, - "grad_norm": 0.553138256072998, - "learning_rate": 0.00019373037166448286, - "loss": 0.2203, - "step": 6030 - }, - { - "epoch": 0.0942160104823111, - "grad_norm": 8.94212818145752, - "learning_rate": 0.00019371997254632808, - "loss": 0.2823, - "step": 6040 - }, - { - "epoch": 0.09437199725463281, - "grad_norm": 4.464921951293945, - "learning_rate": 0.0001937095734281733, - "loss": 0.3549, - "step": 6050 - }, - { - "epoch": 0.09452798402695452, - "grad_norm": 0.8132085204124451, - "learning_rate": 0.00019369917431001852, - "loss": 0.4469, - "step": 6060 - }, - { - "epoch": 0.09468397079927622, - "grad_norm": 2.162436008453369, - "learning_rate": 0.00019368877519186374, - "loss": 0.2974, - "step": 6070 - }, - { - "epoch": 0.09483995757159792, - "grad_norm": 3.0433995723724365, - "learning_rate": 0.00019367837607370896, - "loss": 0.4088, - "step": 6080 - }, - { - "epoch": 0.09499594434391964, - "grad_norm": 0.16140206158161163, - "learning_rate": 0.00019366797695555418, - "loss": 0.2823, - "step": 6090 - }, - { - "epoch": 0.09515193111624135, - "grad_norm": 0.8597097992897034, - "learning_rate": 0.0001936575778373994, - "loss": 0.2961, - "step": 6100 - }, - { - "epoch": 0.09530791788856305, - "grad_norm": 0.8245694637298584, - "learning_rate": 0.00019364717871924461, - "loss": 0.1174, - "step": 6110 - }, - { - "epoch": 0.09546390466088475, - "grad_norm": 4.541038513183594, - "learning_rate": 0.00019363677960108983, - "loss": 0.2867, - "step": 6120 - }, - { - "epoch": 0.09561989143320647, - "grad_norm": 10.14047622680664, - "learning_rate": 0.00019362638048293505, - "loss": 0.7289, - "step": 6130 - }, - { - "epoch": 0.09577587820552817, - "grad_norm": 2.899733304977417, - "learning_rate": 0.00019361598136478027, - "loss": 0.3055, - "step": 6140 - }, - { - "epoch": 0.09593186497784988, - "grad_norm": 3.6699278354644775, - "learning_rate": 0.0001936055822466255, - "loss": 0.285, - "step": 6150 - }, - { - "epoch": 0.09608785175017158, - "grad_norm": 1.7877542972564697, - "learning_rate": 0.0001935951831284707, - "loss": 0.3871, - "step": 6160 - }, - { - "epoch": 0.0962438385224933, - "grad_norm": 2.966464042663574, - "learning_rate": 0.00019358478401031593, - "loss": 0.3512, - "step": 6170 - }, - { - "epoch": 0.096399825294815, - "grad_norm": 2.51312255859375, - "learning_rate": 0.00019357438489216115, - "loss": 0.3996, - "step": 6180 - }, - { - "epoch": 0.0965558120671367, - "grad_norm": 5.688634872436523, - "learning_rate": 0.00019356398577400637, - "loss": 0.3632, - "step": 6190 - }, - { - "epoch": 0.09671179883945841, - "grad_norm": 6.877967834472656, - "learning_rate": 0.0001935535866558516, - "loss": 0.3548, - "step": 6200 - }, - { - "epoch": 0.09686778561178012, - "grad_norm": 3.846853256225586, - "learning_rate": 0.0001935431875376968, - "loss": 0.3466, - "step": 6210 - }, - { - "epoch": 0.09702377238410183, - "grad_norm": 5.721917629241943, - "learning_rate": 0.00019353278841954203, - "loss": 0.3567, - "step": 6220 - }, - { - "epoch": 0.09717975915642353, - "grad_norm": 2.47615647315979, - "learning_rate": 0.00019352238930138725, - "loss": 0.358, - "step": 6230 - }, - { - "epoch": 0.09733574592874525, - "grad_norm": 0.843755304813385, - "learning_rate": 0.00019351199018323247, - "loss": 0.3825, - "step": 6240 - }, - { - "epoch": 0.09749173270106695, - "grad_norm": 2.846238136291504, - "learning_rate": 0.0001935015910650777, - "loss": 0.1483, - "step": 6250 - }, - { - "epoch": 0.09764771947338866, - "grad_norm": 4.587934494018555, - "learning_rate": 0.0001934911919469229, - "loss": 0.3881, - "step": 6260 - }, - { - "epoch": 0.09780370624571036, - "grad_norm": 2.4678304195404053, - "learning_rate": 0.00019348079282876813, - "loss": 0.3414, - "step": 6270 - }, - { - "epoch": 0.09795969301803208, - "grad_norm": 0.5573933124542236, - "learning_rate": 0.00019347039371061335, - "loss": 0.2158, - "step": 6280 - }, - { - "epoch": 0.09811567979035378, - "grad_norm": 2.4001994132995605, - "learning_rate": 0.00019345999459245857, - "loss": 0.3069, - "step": 6290 - }, - { - "epoch": 0.09827166656267548, - "grad_norm": 1.2843972444534302, - "learning_rate": 0.0001934495954743038, - "loss": 0.2112, - "step": 6300 - }, - { - "epoch": 0.09842765333499719, - "grad_norm": 6.185412406921387, - "learning_rate": 0.000193439196356149, - "loss": 0.3721, - "step": 6310 - }, - { - "epoch": 0.0985836401073189, - "grad_norm": 2.1380372047424316, - "learning_rate": 0.00019342879723799423, - "loss": 0.304, - "step": 6320 - }, - { - "epoch": 0.09873962687964061, - "grad_norm": 4.583189964294434, - "learning_rate": 0.00019341839811983945, - "loss": 0.3399, - "step": 6330 - }, - { - "epoch": 0.09889561365196231, - "grad_norm": 1.0530366897583008, - "learning_rate": 0.00019340799900168467, - "loss": 0.2154, - "step": 6340 - }, - { - "epoch": 0.09905160042428401, - "grad_norm": 7.6255202293396, - "learning_rate": 0.00019339759988352989, - "loss": 0.4958, - "step": 6350 - }, - { - "epoch": 0.09920758719660573, - "grad_norm": 9.250993728637695, - "learning_rate": 0.0001933872007653751, - "loss": 0.8084, - "step": 6360 - }, - { - "epoch": 0.09936357396892743, - "grad_norm": 1.315562129020691, - "learning_rate": 0.00019337680164722033, - "loss": 0.4946, - "step": 6370 - }, - { - "epoch": 0.09951956074124914, - "grad_norm": 0.964072048664093, - "learning_rate": 0.00019336640252906555, - "loss": 0.2992, - "step": 6380 - }, - { - "epoch": 0.09967554751357086, - "grad_norm": 2.9351625442504883, - "learning_rate": 0.00019335600341091076, - "loss": 0.3194, - "step": 6390 - }, - { - "epoch": 0.09983153428589256, - "grad_norm": 4.353033542633057, - "learning_rate": 0.00019334560429275598, - "loss": 0.4996, - "step": 6400 - }, - { - "epoch": 0.09998752105821426, - "grad_norm": 6.008812427520752, - "learning_rate": 0.0001933352051746012, - "loss": 0.5374, - "step": 6410 - }, - { - "epoch": 0.10014350783053597, - "grad_norm": 3.2766318321228027, - "learning_rate": 0.00019332480605644642, - "loss": 0.3351, - "step": 6420 - }, - { - "epoch": 0.10029949460285768, - "grad_norm": 0.8730254769325256, - "learning_rate": 0.00019331440693829164, - "loss": 0.3071, - "step": 6430 - }, - { - "epoch": 0.10045548137517939, - "grad_norm": 2.073676347732544, - "learning_rate": 0.00019330400782013686, - "loss": 0.3931, - "step": 6440 - }, - { - "epoch": 0.10061146814750109, - "grad_norm": 10.372605323791504, - "learning_rate": 0.00019329360870198208, - "loss": 0.4514, - "step": 6450 - }, - { - "epoch": 0.1007674549198228, - "grad_norm": 5.290284633636475, - "learning_rate": 0.0001932832095838273, - "loss": 0.3721, - "step": 6460 - }, - { - "epoch": 0.10092344169214451, - "grad_norm": 1.0684601068496704, - "learning_rate": 0.00019327281046567252, - "loss": 0.2702, - "step": 6470 - }, - { - "epoch": 0.10107942846446621, - "grad_norm": 1.874252200126648, - "learning_rate": 0.00019326241134751774, - "loss": 0.4048, - "step": 6480 - }, - { - "epoch": 0.10123541523678792, - "grad_norm": 1.0442904233932495, - "learning_rate": 0.00019325201222936296, - "loss": 0.0949, - "step": 6490 - }, - { - "epoch": 0.10139140200910962, - "grad_norm": 4.850639820098877, - "learning_rate": 0.00019324161311120818, - "loss": 0.5827, - "step": 6500 - }, - { - "epoch": 0.10154738878143134, - "grad_norm": 1.0578861236572266, - "learning_rate": 0.0001932312139930534, - "loss": 0.1895, - "step": 6510 - }, - { - "epoch": 0.10170337555375304, - "grad_norm": 2.8853342533111572, - "learning_rate": 0.00019322081487489862, - "loss": 0.3856, - "step": 6520 - }, - { - "epoch": 0.10185936232607475, - "grad_norm": 2.4124507904052734, - "learning_rate": 0.00019321041575674384, - "loss": 0.4219, - "step": 6530 - }, - { - "epoch": 0.10201534909839646, - "grad_norm": 3.3235878944396973, - "learning_rate": 0.00019320001663858906, - "loss": 0.4442, - "step": 6540 - }, - { - "epoch": 0.10217133587071817, - "grad_norm": 2.4012255668640137, - "learning_rate": 0.00019318961752043428, - "loss": 0.3567, - "step": 6550 - }, - { - "epoch": 0.10232732264303987, - "grad_norm": 2.174771785736084, - "learning_rate": 0.0001931792184022795, - "loss": 0.4336, - "step": 6560 - }, - { - "epoch": 0.10248330941536157, - "grad_norm": 1.151854395866394, - "learning_rate": 0.00019316881928412472, - "loss": 0.2289, - "step": 6570 - }, - { - "epoch": 0.10263929618768329, - "grad_norm": 3.706355571746826, - "learning_rate": 0.00019315842016596994, - "loss": 0.4042, - "step": 6580 - }, - { - "epoch": 0.102795282960005, - "grad_norm": 1.293565273284912, - "learning_rate": 0.00019314802104781516, - "loss": 0.3473, - "step": 6590 - }, - { - "epoch": 0.1029512697323267, - "grad_norm": 1.179175615310669, - "learning_rate": 0.00019313762192966038, - "loss": 0.2029, - "step": 6600 - }, - { - "epoch": 0.1031072565046484, - "grad_norm": 1.1723263263702393, - "learning_rate": 0.0001931272228115056, - "loss": 0.568, - "step": 6610 - }, - { - "epoch": 0.10326324327697012, - "grad_norm": 4.858260154724121, - "learning_rate": 0.00019311682369335082, - "loss": 0.3813, - "step": 6620 - }, - { - "epoch": 0.10341923004929182, - "grad_norm": 1.0854183435440063, - "learning_rate": 0.00019310642457519604, - "loss": 0.3765, - "step": 6630 - }, - { - "epoch": 0.10357521682161352, - "grad_norm": 2.6507794857025146, - "learning_rate": 0.00019309602545704126, - "loss": 0.2726, - "step": 6640 - }, - { - "epoch": 0.10373120359393523, - "grad_norm": 0.2786160111427307, - "learning_rate": 0.00019308562633888648, - "loss": 0.4565, - "step": 6650 - }, - { - "epoch": 0.10388719036625695, - "grad_norm": 3.4936556816101074, - "learning_rate": 0.00019307522722073167, - "loss": 0.25, - "step": 6660 - }, - { - "epoch": 0.10404317713857865, - "grad_norm": 2.4455699920654297, - "learning_rate": 0.00019306482810257691, - "loss": 0.4754, - "step": 6670 - }, - { - "epoch": 0.10419916391090035, - "grad_norm": 2.3813774585723877, - "learning_rate": 0.0001930544289844221, - "loss": 0.2868, - "step": 6680 - }, - { - "epoch": 0.10435515068322206, - "grad_norm": 3.8882644176483154, - "learning_rate": 0.00019304402986626735, - "loss": 0.2338, - "step": 6690 - }, - { - "epoch": 0.10451113745554377, - "grad_norm": 2.1813762187957764, - "learning_rate": 0.00019303363074811255, - "loss": 0.2803, - "step": 6700 - }, - { - "epoch": 0.10466712422786548, - "grad_norm": 4.09452486038208, - "learning_rate": 0.0001930232316299578, - "loss": 0.3288, - "step": 6710 - }, - { - "epoch": 0.10482311100018718, - "grad_norm": 2.630009412765503, - "learning_rate": 0.00019301283251180299, - "loss": 0.3096, - "step": 6720 - }, - { - "epoch": 0.1049790977725089, - "grad_norm": 1.8648737668991089, - "learning_rate": 0.00019300243339364823, - "loss": 0.5055, - "step": 6730 - }, - { - "epoch": 0.1051350845448306, - "grad_norm": 1.4091843366622925, - "learning_rate": 0.00019299203427549343, - "loss": 0.2747, - "step": 6740 - }, - { - "epoch": 0.1052910713171523, - "grad_norm": 1.8997440338134766, - "learning_rate": 0.00019298163515733867, - "loss": 0.4575, - "step": 6750 - }, - { - "epoch": 0.10544705808947401, - "grad_norm": 3.2947158813476562, - "learning_rate": 0.0001929712360391839, - "loss": 0.251, - "step": 6760 - }, - { - "epoch": 0.10560304486179573, - "grad_norm": 1.6467453241348267, - "learning_rate": 0.0001929608369210291, - "loss": 0.2899, - "step": 6770 - }, - { - "epoch": 0.10575903163411743, - "grad_norm": 1.930840015411377, - "learning_rate": 0.00019295043780287433, - "loss": 0.1265, - "step": 6780 - }, - { - "epoch": 0.10591501840643913, - "grad_norm": 9.051993370056152, - "learning_rate": 0.00019294003868471955, - "loss": 0.3849, - "step": 6790 - }, - { - "epoch": 0.10607100517876084, - "grad_norm": 2.5743913650512695, - "learning_rate": 0.00019292963956656477, - "loss": 0.2501, - "step": 6800 - }, - { - "epoch": 0.10622699195108255, - "grad_norm": 1.333158016204834, - "learning_rate": 0.00019291924044841, - "loss": 0.415, - "step": 6810 - }, - { - "epoch": 0.10638297872340426, - "grad_norm": 4.270011901855469, - "learning_rate": 0.0001929088413302552, - "loss": 0.2877, - "step": 6820 - }, - { - "epoch": 0.10653896549572596, - "grad_norm": 2.4836020469665527, - "learning_rate": 0.00019289844221210043, - "loss": 0.5365, - "step": 6830 - }, - { - "epoch": 0.10669495226804766, - "grad_norm": 0.6873673796653748, - "learning_rate": 0.00019288804309394565, - "loss": 0.3269, - "step": 6840 - }, - { - "epoch": 0.10685093904036938, - "grad_norm": 3.456326723098755, - "learning_rate": 0.00019287764397579087, - "loss": 0.3082, - "step": 6850 - }, - { - "epoch": 0.10700692581269108, - "grad_norm": 1.6058659553527832, - "learning_rate": 0.0001928672448576361, - "loss": 0.2203, - "step": 6860 - }, - { - "epoch": 0.10716291258501279, - "grad_norm": 10.284086227416992, - "learning_rate": 0.0001928568457394813, - "loss": 0.4912, - "step": 6870 - }, - { - "epoch": 0.1073188993573345, - "grad_norm": 2.4946155548095703, - "learning_rate": 0.00019284644662132653, - "loss": 0.169, - "step": 6880 - }, - { - "epoch": 0.10747488612965621, - "grad_norm": 3.717564105987549, - "learning_rate": 0.00019283604750317175, - "loss": 0.3952, - "step": 6890 - }, - { - "epoch": 0.10763087290197791, - "grad_norm": 1.3209753036499023, - "learning_rate": 0.00019282564838501697, - "loss": 0.2958, - "step": 6900 - }, - { - "epoch": 0.10778685967429961, - "grad_norm": 2.270832061767578, - "learning_rate": 0.00019281524926686219, - "loss": 0.3197, - "step": 6910 - }, - { - "epoch": 0.10794284644662133, - "grad_norm": 3.6317005157470703, - "learning_rate": 0.0001928048501487074, - "loss": 0.234, - "step": 6920 - }, - { - "epoch": 0.10809883321894304, - "grad_norm": 1.7942756414413452, - "learning_rate": 0.00019279445103055263, - "loss": 0.3435, - "step": 6930 - }, - { - "epoch": 0.10825481999126474, - "grad_norm": 0.7844222784042358, - "learning_rate": 0.00019278405191239785, - "loss": 0.3194, - "step": 6940 - }, - { - "epoch": 0.10841080676358644, - "grad_norm": 4.149482727050781, - "learning_rate": 0.00019277365279424306, - "loss": 0.2947, - "step": 6950 - }, - { - "epoch": 0.10856679353590816, - "grad_norm": 5.077116966247559, - "learning_rate": 0.00019276325367608828, - "loss": 0.3631, - "step": 6960 - }, - { - "epoch": 0.10872278030822986, - "grad_norm": 1.661909580230713, - "learning_rate": 0.0001927528545579335, - "loss": 0.2069, - "step": 6970 - }, - { - "epoch": 0.10887876708055157, - "grad_norm": 4.791439533233643, - "learning_rate": 0.00019274245543977872, - "loss": 0.3677, - "step": 6980 - }, - { - "epoch": 0.10903475385287327, - "grad_norm": 5.538137912750244, - "learning_rate": 0.00019273205632162394, - "loss": 0.2138, - "step": 6990 - }, - { - "epoch": 0.10919074062519499, - "grad_norm": 2.9961178302764893, - "learning_rate": 0.00019272165720346916, - "loss": 0.3518, - "step": 7000 - }, - { - "epoch": 0.10934672739751669, - "grad_norm": 2.286130666732788, - "learning_rate": 0.00019271125808531438, - "loss": 0.4359, - "step": 7010 - }, - { - "epoch": 0.1095027141698384, - "grad_norm": 1.5699574947357178, - "learning_rate": 0.0001927008589671596, - "loss": 0.3327, - "step": 7020 - }, - { - "epoch": 0.10965870094216011, - "grad_norm": 0.44791361689567566, - "learning_rate": 0.00019269045984900482, - "loss": 0.2382, - "step": 7030 - }, - { - "epoch": 0.10981468771448182, - "grad_norm": 1.4798070192337036, - "learning_rate": 0.00019268006073085004, - "loss": 0.3653, - "step": 7040 - }, - { - "epoch": 0.10997067448680352, - "grad_norm": 1.1877557039260864, - "learning_rate": 0.00019266966161269526, - "loss": 0.3338, - "step": 7050 - }, - { - "epoch": 0.11012666125912522, - "grad_norm": 0.3134145140647888, - "learning_rate": 0.00019265926249454048, - "loss": 0.3832, - "step": 7060 - }, - { - "epoch": 0.11028264803144694, - "grad_norm": 1.1449509859085083, - "learning_rate": 0.0001926488633763857, - "loss": 0.27, - "step": 7070 - }, - { - "epoch": 0.11043863480376864, - "grad_norm": 0.8753034472465515, - "learning_rate": 0.00019263846425823092, - "loss": 0.3869, - "step": 7080 - }, - { - "epoch": 0.11059462157609035, - "grad_norm": 4.871153831481934, - "learning_rate": 0.00019262806514007614, - "loss": 0.3679, - "step": 7090 - }, - { - "epoch": 0.11075060834841205, - "grad_norm": 2.3015692234039307, - "learning_rate": 0.00019261766602192136, - "loss": 0.2879, - "step": 7100 - }, - { - "epoch": 0.11090659512073377, - "grad_norm": 10.451770782470703, - "learning_rate": 0.00019260726690376655, - "loss": 0.4762, - "step": 7110 - }, - { - "epoch": 0.11106258189305547, - "grad_norm": 0.6290777325630188, - "learning_rate": 0.0001925968677856118, - "loss": 0.2412, - "step": 7120 - }, - { - "epoch": 0.11121856866537717, - "grad_norm": 3.2147696018218994, - "learning_rate": 0.000192586468667457, - "loss": 0.3651, - "step": 7130 - }, - { - "epoch": 0.11137455543769888, - "grad_norm": 2.770252227783203, - "learning_rate": 0.00019257606954930224, - "loss": 0.2206, - "step": 7140 - }, - { - "epoch": 0.1115305422100206, - "grad_norm": 5.446028232574463, - "learning_rate": 0.00019256567043114743, - "loss": 0.2899, - "step": 7150 - }, - { - "epoch": 0.1116865289823423, - "grad_norm": 1.5418429374694824, - "learning_rate": 0.00019255527131299268, - "loss": 0.2798, - "step": 7160 - }, - { - "epoch": 0.111842515754664, - "grad_norm": 0.7440494298934937, - "learning_rate": 0.00019254487219483787, - "loss": 0.5074, - "step": 7170 - }, - { - "epoch": 0.1119985025269857, - "grad_norm": 3.4700686931610107, - "learning_rate": 0.00019253447307668312, - "loss": 0.2211, - "step": 7180 - }, - { - "epoch": 0.11215448929930742, - "grad_norm": 3.707090139389038, - "learning_rate": 0.0001925240739585283, - "loss": 0.4275, - "step": 7190 - }, - { - "epoch": 0.11231047607162913, - "grad_norm": 4.110064506530762, - "learning_rate": 0.00019251367484037356, - "loss": 0.2144, - "step": 7200 - }, - { - "epoch": 0.11246646284395083, - "grad_norm": 2.0909435749053955, - "learning_rate": 0.00019250327572221875, - "loss": 0.2126, - "step": 7210 - }, - { - "epoch": 0.11262244961627255, - "grad_norm": 1.880205512046814, - "learning_rate": 0.000192492876604064, - "loss": 0.2941, - "step": 7220 - }, - { - "epoch": 0.11277843638859425, - "grad_norm": 2.2292237281799316, - "learning_rate": 0.0001924824774859092, - "loss": 0.3682, - "step": 7230 - }, - { - "epoch": 0.11293442316091595, - "grad_norm": 1.1748030185699463, - "learning_rate": 0.00019247207836775443, - "loss": 0.3069, - "step": 7240 - }, - { - "epoch": 0.11309040993323766, - "grad_norm": 3.5196616649627686, - "learning_rate": 0.00019246167924959963, - "loss": 0.2736, - "step": 7250 - }, - { - "epoch": 0.11324639670555937, - "grad_norm": 0.2772327959537506, - "learning_rate": 0.00019245128013144487, - "loss": 0.3013, - "step": 7260 - }, - { - "epoch": 0.11340238347788108, - "grad_norm": 3.669743537902832, - "learning_rate": 0.00019244088101329007, - "loss": 0.6027, - "step": 7270 - }, - { - "epoch": 0.11355837025020278, - "grad_norm": 0.13140790164470673, - "learning_rate": 0.0001924304818951353, - "loss": 0.1937, - "step": 7280 - }, - { - "epoch": 0.11371435702252448, - "grad_norm": 5.190703392028809, - "learning_rate": 0.0001924200827769805, - "loss": 0.228, - "step": 7290 - }, - { - "epoch": 0.1138703437948462, - "grad_norm": 0.5808380246162415, - "learning_rate": 0.00019240968365882575, - "loss": 0.5241, - "step": 7300 - }, - { - "epoch": 0.1140263305671679, - "grad_norm": 1.059551477432251, - "learning_rate": 0.00019239928454067094, - "loss": 0.1928, - "step": 7310 - }, - { - "epoch": 0.11418231733948961, - "grad_norm": 0.9926179051399231, - "learning_rate": 0.0001923888854225162, - "loss": 0.5882, - "step": 7320 - }, - { - "epoch": 0.11433830411181131, - "grad_norm": 3.8654983043670654, - "learning_rate": 0.00019237848630436138, - "loss": 0.2472, - "step": 7330 - }, - { - "epoch": 0.11449429088413303, - "grad_norm": 1.716522455215454, - "learning_rate": 0.00019236808718620663, - "loss": 0.4283, - "step": 7340 - }, - { - "epoch": 0.11465027765645473, - "grad_norm": 0.9728891849517822, - "learning_rate": 0.00019235768806805182, - "loss": 0.183, - "step": 7350 - }, - { - "epoch": 0.11480626442877644, - "grad_norm": 0.9542707204818726, - "learning_rate": 0.00019234728894989707, - "loss": 0.0839, - "step": 7360 - }, - { - "epoch": 0.11496225120109815, - "grad_norm": 0.8934445977210999, - "learning_rate": 0.00019233688983174226, - "loss": 0.3306, - "step": 7370 - }, - { - "epoch": 0.11511823797341986, - "grad_norm": 3.0257279872894287, - "learning_rate": 0.0001923264907135875, - "loss": 0.3541, - "step": 7380 - }, - { - "epoch": 0.11527422474574156, - "grad_norm": 1.0677869319915771, - "learning_rate": 0.0001923160915954327, - "loss": 0.2027, - "step": 7390 - }, - { - "epoch": 0.11543021151806326, - "grad_norm": 2.3273229598999023, - "learning_rate": 0.00019230569247727795, - "loss": 0.4166, - "step": 7400 - }, - { - "epoch": 0.11558619829038498, - "grad_norm": 0.7801164984703064, - "learning_rate": 0.00019229529335912314, - "loss": 0.2957, - "step": 7410 - }, - { - "epoch": 0.11574218506270668, - "grad_norm": 2.9796864986419678, - "learning_rate": 0.0001922848942409684, - "loss": 0.2919, - "step": 7420 - }, - { - "epoch": 0.11589817183502839, - "grad_norm": 3.8021762371063232, - "learning_rate": 0.00019227449512281358, - "loss": 0.2455, - "step": 7430 - }, - { - "epoch": 0.11605415860735009, - "grad_norm": 0.5679472088813782, - "learning_rate": 0.00019226409600465883, - "loss": 0.1924, - "step": 7440 - }, - { - "epoch": 0.11621014537967181, - "grad_norm": 2.501598596572876, - "learning_rate": 0.00019225369688650402, - "loss": 0.2746, - "step": 7450 - }, - { - "epoch": 0.11636613215199351, - "grad_norm": 9.79799747467041, - "learning_rate": 0.00019224329776834927, - "loss": 0.5196, - "step": 7460 - }, - { - "epoch": 0.11652211892431522, - "grad_norm": 1.9365442991256714, - "learning_rate": 0.00019223289865019446, - "loss": 0.4821, - "step": 7470 - }, - { - "epoch": 0.11667810569663692, - "grad_norm": 2.2277321815490723, - "learning_rate": 0.0001922224995320397, - "loss": 0.262, - "step": 7480 - }, - { - "epoch": 0.11683409246895864, - "grad_norm": 2.1314620971679688, - "learning_rate": 0.0001922121004138849, - "loss": 0.2229, - "step": 7490 - }, - { - "epoch": 0.11699007924128034, - "grad_norm": 0.2625826299190521, - "learning_rate": 0.00019220170129573014, - "loss": 0.2179, - "step": 7500 - }, - { - "epoch": 0.11714606601360204, - "grad_norm": 4.7705183029174805, - "learning_rate": 0.00019219130217757534, - "loss": 0.2437, - "step": 7510 - }, - { - "epoch": 0.11730205278592376, - "grad_norm": 2.144707679748535, - "learning_rate": 0.00019218090305942058, - "loss": 0.331, - "step": 7520 - }, - { - "epoch": 0.11745803955824546, - "grad_norm": 1.4071444272994995, - "learning_rate": 0.00019217050394126578, - "loss": 0.4584, - "step": 7530 - }, - { - "epoch": 0.11761402633056717, - "grad_norm": 1.5593106746673584, - "learning_rate": 0.00019216010482311102, - "loss": 0.5354, - "step": 7540 - }, - { - "epoch": 0.11777001310288887, - "grad_norm": 2.466290235519409, - "learning_rate": 0.00019214970570495622, - "loss": 0.3469, - "step": 7550 - }, - { - "epoch": 0.11792599987521059, - "grad_norm": 11.850781440734863, - "learning_rate": 0.00019213930658680144, - "loss": 0.2365, - "step": 7560 - }, - { - "epoch": 0.11808198664753229, - "grad_norm": 3.969515562057495, - "learning_rate": 0.00019212890746864666, - "loss": 0.2725, - "step": 7570 - }, - { - "epoch": 0.118237973419854, - "grad_norm": 2.140038251876831, - "learning_rate": 0.00019211850835049188, - "loss": 0.265, - "step": 7580 - }, - { - "epoch": 0.1183939601921757, - "grad_norm": 5.333108901977539, - "learning_rate": 0.0001921081092323371, - "loss": 0.3605, - "step": 7590 - }, - { - "epoch": 0.11854994696449742, - "grad_norm": 1.0600502490997314, - "learning_rate": 0.00019209771011418231, - "loss": 0.3654, - "step": 7600 - }, - { - "epoch": 0.11870593373681912, - "grad_norm": 0.09935326874256134, - "learning_rate": 0.00019208731099602753, - "loss": 0.1326, - "step": 7610 - }, - { - "epoch": 0.11886192050914082, - "grad_norm": 0.39839038252830505, - "learning_rate": 0.00019207691187787275, - "loss": 0.4367, - "step": 7620 - }, - { - "epoch": 0.11901790728146253, - "grad_norm": 0.08571303635835648, - "learning_rate": 0.000192066512759718, - "loss": 0.5292, - "step": 7630 - }, - { - "epoch": 0.11917389405378424, - "grad_norm": 2.0398521423339844, - "learning_rate": 0.0001920561136415632, - "loss": 0.2453, - "step": 7640 - }, - { - "epoch": 0.11932988082610595, - "grad_norm": 1.871935486793518, - "learning_rate": 0.00019204571452340844, - "loss": 0.2797, - "step": 7650 - }, - { - "epoch": 0.11948586759842765, - "grad_norm": 8.323614120483398, - "learning_rate": 0.00019203531540525363, - "loss": 0.1233, - "step": 7660 - }, - { - "epoch": 0.11964185437074935, - "grad_norm": 4.91874885559082, - "learning_rate": 0.00019202491628709888, - "loss": 0.3747, - "step": 7670 - }, - { - "epoch": 0.11979784114307107, - "grad_norm": 1.5453379154205322, - "learning_rate": 0.00019201451716894407, - "loss": 0.3069, - "step": 7680 - }, - { - "epoch": 0.11995382791539277, - "grad_norm": 0.7079108953475952, - "learning_rate": 0.00019200411805078932, - "loss": 0.2985, - "step": 7690 - }, - { - "epoch": 0.12010981468771448, - "grad_norm": 1.6853049993515015, - "learning_rate": 0.0001919937189326345, - "loss": 0.3572, - "step": 7700 - }, - { - "epoch": 0.1202658014600362, - "grad_norm": 0.4013264775276184, - "learning_rate": 0.00019198331981447976, - "loss": 0.2229, - "step": 7710 - }, - { - "epoch": 0.1204217882323579, - "grad_norm": 6.26057767868042, - "learning_rate": 0.00019197292069632495, - "loss": 0.3661, - "step": 7720 - }, - { - "epoch": 0.1205777750046796, - "grad_norm": 4.295512676239014, - "learning_rate": 0.0001919625215781702, - "loss": 0.2982, - "step": 7730 - }, - { - "epoch": 0.1207337617770013, - "grad_norm": 3.1866257190704346, - "learning_rate": 0.0001919521224600154, - "loss": 0.3566, - "step": 7740 - }, - { - "epoch": 0.12088974854932302, - "grad_norm": 3.4306435585021973, - "learning_rate": 0.00019194172334186064, - "loss": 0.403, - "step": 7750 - }, - { - "epoch": 0.12104573532164473, - "grad_norm": 5.156405448913574, - "learning_rate": 0.00019193132422370583, - "loss": 0.3088, - "step": 7760 - }, - { - "epoch": 0.12120172209396643, - "grad_norm": 2.769991397857666, - "learning_rate": 0.00019192092510555108, - "loss": 0.224, - "step": 7770 - }, - { - "epoch": 0.12135770886628813, - "grad_norm": 4.933774471282959, - "learning_rate": 0.00019191052598739627, - "loss": 0.5199, - "step": 7780 - }, - { - "epoch": 0.12151369563860985, - "grad_norm": 2.2560410499572754, - "learning_rate": 0.00019190012686924151, - "loss": 0.4144, - "step": 7790 - }, - { - "epoch": 0.12166968241093155, - "grad_norm": 1.0731757879257202, - "learning_rate": 0.0001918897277510867, - "loss": 0.2327, - "step": 7800 - }, - { - "epoch": 0.12182566918325326, - "grad_norm": 1.9298815727233887, - "learning_rate": 0.00019187932863293195, - "loss": 0.3142, - "step": 7810 - }, - { - "epoch": 0.12198165595557496, - "grad_norm": 2.4373934268951416, - "learning_rate": 0.00019186892951477715, - "loss": 0.3267, - "step": 7820 - }, - { - "epoch": 0.12213764272789668, - "grad_norm": 2.707430839538574, - "learning_rate": 0.0001918585303966224, - "loss": 0.3451, - "step": 7830 - }, - { - "epoch": 0.12229362950021838, - "grad_norm": 2.4040186405181885, - "learning_rate": 0.00019184813127846759, - "loss": 0.3833, - "step": 7840 - }, - { - "epoch": 0.12244961627254008, - "grad_norm": 1.0734840631484985, - "learning_rate": 0.00019183773216031283, - "loss": 0.3856, - "step": 7850 - }, - { - "epoch": 0.1226056030448618, - "grad_norm": 2.8232712745666504, - "learning_rate": 0.00019182733304215802, - "loss": 0.3998, - "step": 7860 - }, - { - "epoch": 0.1227615898171835, - "grad_norm": 2.6121976375579834, - "learning_rate": 0.00019181693392400327, - "loss": 0.3892, - "step": 7870 - }, - { - "epoch": 0.12291757658950521, - "grad_norm": 3.365650177001953, - "learning_rate": 0.00019180653480584846, - "loss": 0.2566, - "step": 7880 - }, - { - "epoch": 0.12307356336182691, - "grad_norm": 0.292998731136322, - "learning_rate": 0.0001917961356876937, - "loss": 0.2939, - "step": 7890 - }, - { - "epoch": 0.12322955013414863, - "grad_norm": 0.3753792941570282, - "learning_rate": 0.0001917857365695389, - "loss": 0.2036, - "step": 7900 - }, - { - "epoch": 0.12338553690647033, - "grad_norm": 0.28090208768844604, - "learning_rate": 0.00019177533745138415, - "loss": 0.2517, - "step": 7910 - }, - { - "epoch": 0.12354152367879204, - "grad_norm": 1.4824659824371338, - "learning_rate": 0.00019176493833322934, - "loss": 0.9074, - "step": 7920 - }, - { - "epoch": 0.12369751045111374, - "grad_norm": 5.32909631729126, - "learning_rate": 0.0001917545392150746, - "loss": 0.428, - "step": 7930 - }, - { - "epoch": 0.12385349722343546, - "grad_norm": 4.6401824951171875, - "learning_rate": 0.00019174414009691978, - "loss": 0.2052, - "step": 7940 - }, - { - "epoch": 0.12400948399575716, - "grad_norm": 1.4302258491516113, - "learning_rate": 0.00019173374097876503, - "loss": 0.1912, - "step": 7950 - }, - { - "epoch": 0.12416547076807886, - "grad_norm": 0.6276291608810425, - "learning_rate": 0.00019172334186061022, - "loss": 0.2532, - "step": 7960 - }, - { - "epoch": 0.12432145754040057, - "grad_norm": 0.4935452342033386, - "learning_rate": 0.00019171294274245547, - "loss": 0.2772, - "step": 7970 - }, - { - "epoch": 0.12447744431272229, - "grad_norm": 1.2430734634399414, - "learning_rate": 0.00019170254362430066, - "loss": 0.2626, - "step": 7980 - }, - { - "epoch": 0.12463343108504399, - "grad_norm": 2.41579270362854, - "learning_rate": 0.0001916921445061459, - "loss": 0.2515, - "step": 7990 - }, - { - "epoch": 0.12478941785736569, - "grad_norm": 16.254776000976562, - "learning_rate": 0.0001916817453879911, - "loss": 0.4629, - "step": 8000 - }, - { - "epoch": 0.12494540462968741, - "grad_norm": 0.9846019148826599, - "learning_rate": 0.00019167134626983632, - "loss": 0.1909, - "step": 8010 - }, - { - "epoch": 0.1251013914020091, - "grad_norm": 0.2564077079296112, - "learning_rate": 0.00019166094715168154, - "loss": 0.2582, - "step": 8020 - }, - { - "epoch": 0.12525737817433083, - "grad_norm": 4.15382719039917, - "learning_rate": 0.00019165054803352676, - "loss": 0.3407, - "step": 8030 - }, - { - "epoch": 0.12541336494665253, - "grad_norm": 1.3758265972137451, - "learning_rate": 0.00019164014891537198, - "loss": 0.368, - "step": 8040 - }, - { - "epoch": 0.12556935171897424, - "grad_norm": 0.429997980594635, - "learning_rate": 0.0001916297497972172, - "loss": 0.4058, - "step": 8050 - }, - { - "epoch": 0.12572533849129594, - "grad_norm": 1.7166649103164673, - "learning_rate": 0.00019161935067906242, - "loss": 0.4811, - "step": 8060 - }, - { - "epoch": 0.12588132526361764, - "grad_norm": 1.644769310951233, - "learning_rate": 0.00019160895156090764, - "loss": 0.2566, - "step": 8070 - }, - { - "epoch": 0.12603731203593935, - "grad_norm": 1.0035743713378906, - "learning_rate": 0.00019159855244275286, - "loss": 0.2881, - "step": 8080 - }, - { - "epoch": 0.12619329880826105, - "grad_norm": 11.098752975463867, - "learning_rate": 0.00019158815332459808, - "loss": 0.2669, - "step": 8090 - }, - { - "epoch": 0.12634928558058275, - "grad_norm": 2.0310935974121094, - "learning_rate": 0.0001915777542064433, - "loss": 0.6095, - "step": 8100 - }, - { - "epoch": 0.12650527235290449, - "grad_norm": 3.2923743724823, - "learning_rate": 0.00019156735508828852, - "loss": 0.3604, - "step": 8110 - }, - { - "epoch": 0.1266612591252262, - "grad_norm": 2.754688262939453, - "learning_rate": 0.00019155695597013374, - "loss": 0.373, - "step": 8120 - }, - { - "epoch": 0.1268172458975479, - "grad_norm": 1.6191695928573608, - "learning_rate": 0.00019154655685197896, - "loss": 0.342, - "step": 8130 - }, - { - "epoch": 0.1269732326698696, - "grad_norm": 4.7524285316467285, - "learning_rate": 0.00019153615773382417, - "loss": 0.2754, - "step": 8140 - }, - { - "epoch": 0.1271292194421913, - "grad_norm": 0.20522356033325195, - "learning_rate": 0.0001915257586156694, - "loss": 0.1422, - "step": 8150 - }, - { - "epoch": 0.127285206214513, - "grad_norm": 0.21395255625247955, - "learning_rate": 0.00019151535949751461, - "loss": 0.3286, - "step": 8160 - }, - { - "epoch": 0.1274411929868347, - "grad_norm": 1.322213888168335, - "learning_rate": 0.00019150496037935983, - "loss": 0.393, - "step": 8170 - }, - { - "epoch": 0.12759717975915644, - "grad_norm": 2.434922933578491, - "learning_rate": 0.00019149456126120505, - "loss": 0.2596, - "step": 8180 - }, - { - "epoch": 0.12775316653147814, - "grad_norm": 6.111631393432617, - "learning_rate": 0.00019148416214305027, - "loss": 0.7321, - "step": 8190 - }, - { - "epoch": 0.12790915330379984, - "grad_norm": 8.627694129943848, - "learning_rate": 0.0001914737630248955, - "loss": 0.2527, - "step": 8200 - }, - { - "epoch": 0.12806514007612155, - "grad_norm": 7.098100662231445, - "learning_rate": 0.0001914633639067407, - "loss": 0.3481, - "step": 8210 - }, - { - "epoch": 0.12822112684844325, - "grad_norm": 2.6047921180725098, - "learning_rate": 0.00019145296478858593, - "loss": 0.3985, - "step": 8220 - }, - { - "epoch": 0.12837711362076495, - "grad_norm": 2.3601999282836914, - "learning_rate": 0.00019144256567043115, - "loss": 0.3182, - "step": 8230 - }, - { - "epoch": 0.12853310039308666, - "grad_norm": 5.261406421661377, - "learning_rate": 0.00019143216655227637, - "loss": 0.2836, - "step": 8240 - }, - { - "epoch": 0.12868908716540836, - "grad_norm": 2.3208975791931152, - "learning_rate": 0.0001914217674341216, - "loss": 0.1436, - "step": 8250 - }, - { - "epoch": 0.1288450739377301, - "grad_norm": 0.5465450286865234, - "learning_rate": 0.0001914113683159668, - "loss": 0.1902, - "step": 8260 - }, - { - "epoch": 0.1290010607100518, - "grad_norm": 1.4398523569107056, - "learning_rate": 0.00019140096919781203, - "loss": 0.3946, - "step": 8270 - }, - { - "epoch": 0.1291570474823735, - "grad_norm": 0.4072510302066803, - "learning_rate": 0.00019139057007965725, - "loss": 0.2422, - "step": 8280 - }, - { - "epoch": 0.1293130342546952, - "grad_norm": 1.2237650156021118, - "learning_rate": 0.00019138017096150247, - "loss": 0.166, - "step": 8290 - }, - { - "epoch": 0.1294690210270169, - "grad_norm": 0.7359814047813416, - "learning_rate": 0.0001913697718433477, - "loss": 0.2737, - "step": 8300 - }, - { - "epoch": 0.1296250077993386, - "grad_norm": 2.178809642791748, - "learning_rate": 0.0001913593727251929, - "loss": 0.1729, - "step": 8310 - }, - { - "epoch": 0.1297809945716603, - "grad_norm": 3.504578113555908, - "learning_rate": 0.00019134897360703813, - "loss": 0.3232, - "step": 8320 - }, - { - "epoch": 0.12993698134398202, - "grad_norm": 1.777565360069275, - "learning_rate": 0.00019133857448888335, - "loss": 0.3073, - "step": 8330 - }, - { - "epoch": 0.13009296811630375, - "grad_norm": 2.342390298843384, - "learning_rate": 0.00019132817537072857, - "loss": 0.4191, - "step": 8340 - }, - { - "epoch": 0.13024895488862545, - "grad_norm": 1.5919382572174072, - "learning_rate": 0.0001913177762525738, - "loss": 0.361, - "step": 8350 - }, - { - "epoch": 0.13040494166094715, - "grad_norm": 5.767488956451416, - "learning_rate": 0.000191307377134419, - "loss": 0.2514, - "step": 8360 - }, - { - "epoch": 0.13056092843326886, - "grad_norm": 0.2066197395324707, - "learning_rate": 0.00019129697801626423, - "loss": 0.2475, - "step": 8370 - }, - { - "epoch": 0.13071691520559056, - "grad_norm": 11.444001197814941, - "learning_rate": 0.00019128657889810945, - "loss": 0.3819, - "step": 8380 - }, - { - "epoch": 0.13087290197791226, - "grad_norm": 2.607598304748535, - "learning_rate": 0.00019127617977995467, - "loss": 0.4103, - "step": 8390 - }, - { - "epoch": 0.13102888875023397, - "grad_norm": 4.767722129821777, - "learning_rate": 0.00019126578066179989, - "loss": 0.3511, - "step": 8400 - }, - { - "epoch": 0.1311848755225557, - "grad_norm": 3.6843106746673584, - "learning_rate": 0.0001912553815436451, - "loss": 0.4402, - "step": 8410 - }, - { - "epoch": 0.1313408622948774, - "grad_norm": 3.7585389614105225, - "learning_rate": 0.00019124498242549032, - "loss": 0.4182, - "step": 8420 - }, - { - "epoch": 0.1314968490671991, - "grad_norm": 3.5597290992736816, - "learning_rate": 0.00019123458330733554, - "loss": 0.3161, - "step": 8430 - }, - { - "epoch": 0.1316528358395208, - "grad_norm": 0.20346300303936005, - "learning_rate": 0.00019122418418918076, - "loss": 0.1689, - "step": 8440 - }, - { - "epoch": 0.1318088226118425, - "grad_norm": 3.307145118713379, - "learning_rate": 0.00019121378507102598, - "loss": 0.3883, - "step": 8450 - }, - { - "epoch": 0.13196480938416422, - "grad_norm": 0.09858717024326324, - "learning_rate": 0.0001912033859528712, - "loss": 0.1792, - "step": 8460 - }, - { - "epoch": 0.13212079615648592, - "grad_norm": 3.1137168407440186, - "learning_rate": 0.00019119298683471642, - "loss": 0.2061, - "step": 8470 - }, - { - "epoch": 0.13227678292880762, - "grad_norm": 2.4133315086364746, - "learning_rate": 0.00019118258771656164, - "loss": 0.2278, - "step": 8480 - }, - { - "epoch": 0.13243276970112935, - "grad_norm": 1.2268180847167969, - "learning_rate": 0.00019117218859840686, - "loss": 0.2337, - "step": 8490 - }, - { - "epoch": 0.13258875647345106, - "grad_norm": 2.3650662899017334, - "learning_rate": 0.00019116178948025208, - "loss": 0.4079, - "step": 8500 - }, - { - "epoch": 0.13274474324577276, - "grad_norm": 1.009161353111267, - "learning_rate": 0.0001911513903620973, - "loss": 0.4019, - "step": 8510 - }, - { - "epoch": 0.13290073001809447, - "grad_norm": 1.7153061628341675, - "learning_rate": 0.00019114099124394252, - "loss": 0.2673, - "step": 8520 - }, - { - "epoch": 0.13305671679041617, - "grad_norm": 3.887568950653076, - "learning_rate": 0.00019113059212578774, - "loss": 0.2689, - "step": 8530 - }, - { - "epoch": 0.13321270356273787, - "grad_norm": 1.7054911851882935, - "learning_rate": 0.00019112019300763296, - "loss": 0.4124, - "step": 8540 - }, - { - "epoch": 0.13336869033505958, - "grad_norm": 2.063857316970825, - "learning_rate": 0.00019110979388947818, - "loss": 0.2602, - "step": 8550 - }, - { - "epoch": 0.1335246771073813, - "grad_norm": 1.704604148864746, - "learning_rate": 0.0001910993947713234, - "loss": 0.4358, - "step": 8560 - }, - { - "epoch": 0.133680663879703, - "grad_norm": 5.169620513916016, - "learning_rate": 0.00019108899565316862, - "loss": 0.4096, - "step": 8570 - }, - { - "epoch": 0.1338366506520247, - "grad_norm": 1.4623101949691772, - "learning_rate": 0.00019107859653501384, - "loss": 0.1843, - "step": 8580 - }, - { - "epoch": 0.13399263742434642, - "grad_norm": 2.4411351680755615, - "learning_rate": 0.00019106819741685906, - "loss": 0.1862, - "step": 8590 - }, - { - "epoch": 0.13414862419666812, - "grad_norm": 2.140017032623291, - "learning_rate": 0.00019105779829870428, - "loss": 0.4326, - "step": 8600 - }, - { - "epoch": 0.13430461096898982, - "grad_norm": 3.320019483566284, - "learning_rate": 0.0001910473991805495, - "loss": 0.2574, - "step": 8610 - }, - { - "epoch": 0.13446059774131153, - "grad_norm": 1.0196399688720703, - "learning_rate": 0.00019103700006239472, - "loss": 0.1434, - "step": 8620 - }, - { - "epoch": 0.13461658451363323, - "grad_norm": 0.6640204191207886, - "learning_rate": 0.00019102660094423994, - "loss": 0.5926, - "step": 8630 - }, - { - "epoch": 0.13477257128595496, - "grad_norm": 2.0378332138061523, - "learning_rate": 0.00019101620182608516, - "loss": 0.202, - "step": 8640 - }, - { - "epoch": 0.13492855805827667, - "grad_norm": 0.19052188098430634, - "learning_rate": 0.00019100580270793038, - "loss": 0.3694, - "step": 8650 - }, - { - "epoch": 0.13508454483059837, - "grad_norm": 1.1991510391235352, - "learning_rate": 0.0001909954035897756, - "loss": 0.2576, - "step": 8660 - }, - { - "epoch": 0.13524053160292007, - "grad_norm": 0.26283520460128784, - "learning_rate": 0.00019098500447162082, - "loss": 0.1626, - "step": 8670 - }, - { - "epoch": 0.13539651837524178, - "grad_norm": 3.249889612197876, - "learning_rate": 0.00019097460535346604, - "loss": 0.3994, - "step": 8680 - }, - { - "epoch": 0.13555250514756348, - "grad_norm": 6.312870502471924, - "learning_rate": 0.00019096420623531126, - "loss": 0.4395, - "step": 8690 - }, - { - "epoch": 0.13570849191988518, - "grad_norm": 0.6005007028579712, - "learning_rate": 0.00019095380711715647, - "loss": 0.195, - "step": 8700 - }, - { - "epoch": 0.1358644786922069, - "grad_norm": 2.2957708835601807, - "learning_rate": 0.0001909434079990017, - "loss": 0.2567, - "step": 8710 - }, - { - "epoch": 0.13602046546452862, - "grad_norm": 2.172750949859619, - "learning_rate": 0.00019093300888084691, - "loss": 0.4404, - "step": 8720 - }, - { - "epoch": 0.13617645223685032, - "grad_norm": 6.653365135192871, - "learning_rate": 0.00019092260976269213, - "loss": 0.2268, - "step": 8730 - }, - { - "epoch": 0.13633243900917202, - "grad_norm": 2.227572441101074, - "learning_rate": 0.00019091221064453735, - "loss": 0.3324, - "step": 8740 - }, - { - "epoch": 0.13648842578149373, - "grad_norm": 0.9583300352096558, - "learning_rate": 0.00019090181152638257, - "loss": 0.3938, - "step": 8750 - }, - { - "epoch": 0.13664441255381543, - "grad_norm": 0.9899696111679077, - "learning_rate": 0.0001908914124082278, - "loss": 0.2501, - "step": 8760 - }, - { - "epoch": 0.13680039932613713, - "grad_norm": 3.927495241165161, - "learning_rate": 0.000190881013290073, - "loss": 0.2983, - "step": 8770 - }, - { - "epoch": 0.13695638609845884, - "grad_norm": 2.9593987464904785, - "learning_rate": 0.00019087061417191823, - "loss": 0.3336, - "step": 8780 - }, - { - "epoch": 0.13711237287078057, - "grad_norm": 1.9015326499938965, - "learning_rate": 0.00019086021505376345, - "loss": 0.4253, - "step": 8790 - }, - { - "epoch": 0.13726835964310227, - "grad_norm": 0.7660776376724243, - "learning_rate": 0.00019084981593560867, - "loss": 0.367, - "step": 8800 - }, - { - "epoch": 0.13742434641542398, - "grad_norm": 4.850160121917725, - "learning_rate": 0.0001908394168174539, - "loss": 0.2792, - "step": 8810 - }, - { - "epoch": 0.13758033318774568, - "grad_norm": 2.631519317626953, - "learning_rate": 0.0001908290176992991, - "loss": 0.3198, - "step": 8820 - }, - { - "epoch": 0.13773631996006738, - "grad_norm": 0.9831134676933289, - "learning_rate": 0.00019081861858114433, - "loss": 0.2242, - "step": 8830 - }, - { - "epoch": 0.13789230673238909, - "grad_norm": 2.2504160404205322, - "learning_rate": 0.00019080821946298955, - "loss": 0.2806, - "step": 8840 - }, - { - "epoch": 0.1380482935047108, - "grad_norm": 3.266126871109009, - "learning_rate": 0.00019079782034483477, - "loss": 0.2641, - "step": 8850 - }, - { - "epoch": 0.13820428027703252, - "grad_norm": 0.21948297321796417, - "learning_rate": 0.00019078742122668, - "loss": 0.6415, - "step": 8860 - }, - { - "epoch": 0.13836026704935422, - "grad_norm": 0.23968499898910522, - "learning_rate": 0.0001907770221085252, - "loss": 0.2328, - "step": 8870 - }, - { - "epoch": 0.13851625382167593, - "grad_norm": 2.194678544998169, - "learning_rate": 0.00019076662299037043, - "loss": 0.4037, - "step": 8880 - }, - { - "epoch": 0.13867224059399763, - "grad_norm": 1.9633435010910034, - "learning_rate": 0.00019075622387221565, - "loss": 0.501, - "step": 8890 - }, - { - "epoch": 0.13882822736631933, - "grad_norm": 1.5285439491271973, - "learning_rate": 0.00019074582475406087, - "loss": 0.1977, - "step": 8900 - }, - { - "epoch": 0.13898421413864104, - "grad_norm": 0.5968760848045349, - "learning_rate": 0.0001907354256359061, - "loss": 0.2279, - "step": 8910 - }, - { - "epoch": 0.13914020091096274, - "grad_norm": 2.4202139377593994, - "learning_rate": 0.0001907250265177513, - "loss": 0.4585, - "step": 8920 - }, - { - "epoch": 0.13929618768328444, - "grad_norm": 6.815012454986572, - "learning_rate": 0.0001907146273995965, - "loss": 0.2049, - "step": 8930 - }, - { - "epoch": 0.13945217445560618, - "grad_norm": 2.510038375854492, - "learning_rate": 0.00019070422828144175, - "loss": 0.4618, - "step": 8940 - }, - { - "epoch": 0.13960816122792788, - "grad_norm": 5.081661224365234, - "learning_rate": 0.00019069382916328697, - "loss": 0.5324, - "step": 8950 - }, - { - "epoch": 0.13976414800024958, - "grad_norm": 0.5944671630859375, - "learning_rate": 0.00019068343004513219, - "loss": 0.3111, - "step": 8960 - }, - { - "epoch": 0.1399201347725713, - "grad_norm": 4.6881022453308105, - "learning_rate": 0.0001906730309269774, - "loss": 0.2048, - "step": 8970 - }, - { - "epoch": 0.140076121544893, - "grad_norm": 2.042879104614258, - "learning_rate": 0.00019066263180882262, - "loss": 0.271, - "step": 8980 - }, - { - "epoch": 0.1402321083172147, - "grad_norm": 1.1326006650924683, - "learning_rate": 0.00019065223269066784, - "loss": 0.1859, - "step": 8990 - }, - { - "epoch": 0.1403880950895364, - "grad_norm": 2.0956993103027344, - "learning_rate": 0.00019064183357251306, - "loss": 0.1465, - "step": 9000 - }, - { - "epoch": 0.14054408186185813, - "grad_norm": 1.8095557689666748, - "learning_rate": 0.00019063143445435828, - "loss": 0.2947, - "step": 9010 - }, - { - "epoch": 0.14070006863417983, - "grad_norm": 2.5442774295806885, - "learning_rate": 0.0001906210353362035, - "loss": 0.3196, - "step": 9020 - }, - { - "epoch": 0.14085605540650153, - "grad_norm": 3.7610888481140137, - "learning_rate": 0.00019061063621804872, - "loss": 0.2794, - "step": 9030 - }, - { - "epoch": 0.14101204217882324, - "grad_norm": 0.2639630138874054, - "learning_rate": 0.00019060023709989394, - "loss": 0.339, - "step": 9040 - }, - { - "epoch": 0.14116802895114494, - "grad_norm": 1.3571760654449463, - "learning_rate": 0.00019058983798173916, - "loss": 0.295, - "step": 9050 - }, - { - "epoch": 0.14132401572346664, - "grad_norm": 1.9788237810134888, - "learning_rate": 0.00019057943886358438, - "loss": 0.145, - "step": 9060 - }, - { - "epoch": 0.14148000249578835, - "grad_norm": 1.98755943775177, - "learning_rate": 0.0001905690397454296, - "loss": 0.5241, - "step": 9070 - }, - { - "epoch": 0.14163598926811005, - "grad_norm": 11.126473426818848, - "learning_rate": 0.00019055864062727482, - "loss": 0.2084, - "step": 9080 - }, - { - "epoch": 0.14179197604043178, - "grad_norm": 6.048513412475586, - "learning_rate": 0.00019054824150912004, - "loss": 0.4383, - "step": 9090 - }, - { - "epoch": 0.1419479628127535, - "grad_norm": 6.858469486236572, - "learning_rate": 0.00019053784239096526, - "loss": 0.5582, - "step": 9100 - }, - { - "epoch": 0.1421039495850752, - "grad_norm": 0.4861454665660858, - "learning_rate": 0.00019052744327281048, - "loss": 0.2774, - "step": 9110 - }, - { - "epoch": 0.1422599363573969, - "grad_norm": 3.1820757389068604, - "learning_rate": 0.0001905170441546557, - "loss": 0.2467, - "step": 9120 - }, - { - "epoch": 0.1424159231297186, - "grad_norm": 1.196014642715454, - "learning_rate": 0.00019050664503650092, - "loss": 0.2698, - "step": 9130 - }, - { - "epoch": 0.1425719099020403, - "grad_norm": 0.511761486530304, - "learning_rate": 0.00019049624591834614, - "loss": 0.1662, - "step": 9140 - }, - { - "epoch": 0.142727896674362, - "grad_norm": 4.00806188583374, - "learning_rate": 0.00019048584680019136, - "loss": 0.4401, - "step": 9150 - }, - { - "epoch": 0.14288388344668373, - "grad_norm": 1.0736618041992188, - "learning_rate": 0.00019047544768203658, - "loss": 0.239, - "step": 9160 - }, - { - "epoch": 0.14303987021900544, - "grad_norm": 3.9986984729766846, - "learning_rate": 0.0001904650485638818, - "loss": 0.4237, - "step": 9170 - }, - { - "epoch": 0.14319585699132714, - "grad_norm": 0.895958423614502, - "learning_rate": 0.00019045464944572702, - "loss": 0.2857, - "step": 9180 - }, - { - "epoch": 0.14335184376364885, - "grad_norm": 2.562671661376953, - "learning_rate": 0.00019044425032757224, - "loss": 0.3014, - "step": 9190 - }, - { - "epoch": 0.14350783053597055, - "grad_norm": 5.249929904937744, - "learning_rate": 0.00019043385120941746, - "loss": 0.3175, - "step": 9200 - }, - { - "epoch": 0.14366381730829225, - "grad_norm": 0.5497576594352722, - "learning_rate": 0.00019042345209126268, - "loss": 0.2828, - "step": 9210 - }, - { - "epoch": 0.14381980408061396, - "grad_norm": 3.966878890991211, - "learning_rate": 0.0001904130529731079, - "loss": 0.4266, - "step": 9220 - }, - { - "epoch": 0.14397579085293566, - "grad_norm": 0.5906422734260559, - "learning_rate": 0.00019040265385495312, - "loss": 0.2646, - "step": 9230 - }, - { - "epoch": 0.1441317776252574, - "grad_norm": 2.217254877090454, - "learning_rate": 0.00019039225473679834, - "loss": 0.1863, - "step": 9240 - }, - { - "epoch": 0.1442877643975791, - "grad_norm": 1.0814603567123413, - "learning_rate": 0.00019038185561864356, - "loss": 0.3919, - "step": 9250 - }, - { - "epoch": 0.1444437511699008, - "grad_norm": 3.348808765411377, - "learning_rate": 0.00019037145650048877, - "loss": 0.3517, - "step": 9260 - }, - { - "epoch": 0.1445997379422225, - "grad_norm": 2.634974479675293, - "learning_rate": 0.000190361057382334, - "loss": 0.332, - "step": 9270 - }, - { - "epoch": 0.1447557247145442, - "grad_norm": 5.290812969207764, - "learning_rate": 0.00019035065826417921, - "loss": 0.4049, - "step": 9280 - }, - { - "epoch": 0.1449117114868659, - "grad_norm": 1.3467392921447754, - "learning_rate": 0.00019034025914602443, - "loss": 0.4742, - "step": 9290 - }, - { - "epoch": 0.1450676982591876, - "grad_norm": 1.3368233442306519, - "learning_rate": 0.00019032986002786965, - "loss": 0.2514, - "step": 9300 - }, - { - "epoch": 0.14522368503150931, - "grad_norm": 0.37425053119659424, - "learning_rate": 0.00019031946090971487, - "loss": 0.2366, - "step": 9310 - }, - { - "epoch": 0.14537967180383105, - "grad_norm": 0.47371023893356323, - "learning_rate": 0.0001903090617915601, - "loss": 0.2965, - "step": 9320 - }, - { - "epoch": 0.14553565857615275, - "grad_norm": 1.9556888341903687, - "learning_rate": 0.0001902986626734053, - "loss": 0.2265, - "step": 9330 - }, - { - "epoch": 0.14569164534847445, - "grad_norm": 1.1136163473129272, - "learning_rate": 0.00019028826355525053, - "loss": 0.2852, - "step": 9340 - }, - { - "epoch": 0.14584763212079616, - "grad_norm": 1.7750933170318604, - "learning_rate": 0.00019027786443709575, - "loss": 0.2825, - "step": 9350 - }, - { - "epoch": 0.14600361889311786, - "grad_norm": 1.3628592491149902, - "learning_rate": 0.00019026746531894097, - "loss": 0.5666, - "step": 9360 - }, - { - "epoch": 0.14615960566543956, - "grad_norm": 2.016136646270752, - "learning_rate": 0.0001902570662007862, - "loss": 0.4292, - "step": 9370 - }, - { - "epoch": 0.14631559243776127, - "grad_norm": 1.0013788938522339, - "learning_rate": 0.00019024666708263138, - "loss": 0.3201, - "step": 9380 - }, - { - "epoch": 0.146471579210083, - "grad_norm": 2.918337106704712, - "learning_rate": 0.00019023626796447663, - "loss": 0.313, - "step": 9390 - }, - { - "epoch": 0.1466275659824047, - "grad_norm": 2.9638941287994385, - "learning_rate": 0.00019022586884632182, - "loss": 0.4592, - "step": 9400 - }, - { - "epoch": 0.1467835527547264, - "grad_norm": 2.4582252502441406, - "learning_rate": 0.00019021546972816707, - "loss": 0.2585, - "step": 9410 - }, - { - "epoch": 0.1469395395270481, - "grad_norm": 4.470787525177002, - "learning_rate": 0.00019020507061001226, - "loss": 0.4089, - "step": 9420 - }, - { - "epoch": 0.1470955262993698, - "grad_norm": 1.9958068132400513, - "learning_rate": 0.0001901946714918575, - "loss": 0.6056, - "step": 9430 - }, - { - "epoch": 0.14725151307169151, - "grad_norm": 0.7107248306274414, - "learning_rate": 0.0001901842723737027, - "loss": 0.3247, - "step": 9440 - }, - { - "epoch": 0.14740749984401322, - "grad_norm": 3.0590286254882812, - "learning_rate": 0.00019017387325554795, - "loss": 0.1213, - "step": 9450 - }, - { - "epoch": 0.14756348661633492, - "grad_norm": 1.6952658891677856, - "learning_rate": 0.00019016347413739314, - "loss": 0.1593, - "step": 9460 - }, - { - "epoch": 0.14771947338865665, - "grad_norm": 5.949721336364746, - "learning_rate": 0.0001901530750192384, - "loss": 0.271, - "step": 9470 - }, - { - "epoch": 0.14787546016097836, - "grad_norm": 0.35458824038505554, - "learning_rate": 0.00019014267590108358, - "loss": 0.1206, - "step": 9480 - }, - { - "epoch": 0.14803144693330006, - "grad_norm": 1.2978142499923706, - "learning_rate": 0.00019013227678292883, - "loss": 0.4097, - "step": 9490 - }, - { - "epoch": 0.14818743370562176, - "grad_norm": 2.7621631622314453, - "learning_rate": 0.00019012187766477402, - "loss": 0.3056, - "step": 9500 - }, - { - "epoch": 0.14834342047794347, - "grad_norm": 1.4229602813720703, - "learning_rate": 0.00019011147854661927, - "loss": 0.2489, - "step": 9510 - }, - { - "epoch": 0.14849940725026517, - "grad_norm": 1.638088345527649, - "learning_rate": 0.00019010107942846446, - "loss": 0.3914, - "step": 9520 - }, - { - "epoch": 0.14865539402258687, - "grad_norm": 1.4291104078292847, - "learning_rate": 0.0001900906803103097, - "loss": 0.2537, - "step": 9530 - }, - { - "epoch": 0.1488113807949086, - "grad_norm": 1.6094496250152588, - "learning_rate": 0.0001900802811921549, - "loss": 0.258, - "step": 9540 - }, - { - "epoch": 0.1489673675672303, - "grad_norm": 5.987236976623535, - "learning_rate": 0.00019006988207400014, - "loss": 0.2963, - "step": 9550 - }, - { - "epoch": 0.149123354339552, - "grad_norm": 4.741234302520752, - "learning_rate": 0.00019005948295584534, - "loss": 0.2403, - "step": 9560 - }, - { - "epoch": 0.14927934111187371, - "grad_norm": 1.7254575490951538, - "learning_rate": 0.00019004908383769058, - "loss": 0.2019, - "step": 9570 - }, - { - "epoch": 0.14943532788419542, - "grad_norm": 0.6806448698043823, - "learning_rate": 0.00019003868471953578, - "loss": 0.3646, - "step": 9580 - }, - { - "epoch": 0.14959131465651712, - "grad_norm": 1.5645960569381714, - "learning_rate": 0.00019002828560138102, - "loss": 0.363, - "step": 9590 - }, - { - "epoch": 0.14974730142883882, - "grad_norm": 3.121180534362793, - "learning_rate": 0.00019001788648322622, - "loss": 0.4002, - "step": 9600 - }, - { - "epoch": 0.14990328820116053, - "grad_norm": 4.279516696929932, - "learning_rate": 0.00019000748736507146, - "loss": 0.1717, - "step": 9610 - }, - { - "epoch": 0.15005927497348226, - "grad_norm": 0.7724094986915588, - "learning_rate": 0.00018999708824691665, - "loss": 0.2195, - "step": 9620 - }, - { - "epoch": 0.15021526174580396, - "grad_norm": 0.4113433361053467, - "learning_rate": 0.0001899866891287619, - "loss": 0.2139, - "step": 9630 - }, - { - "epoch": 0.15037124851812567, - "grad_norm": 0.6473997235298157, - "learning_rate": 0.0001899762900106071, - "loss": 0.4487, - "step": 9640 - }, - { - "epoch": 0.15052723529044737, - "grad_norm": 0.9015470147132874, - "learning_rate": 0.00018996589089245234, - "loss": 0.2524, - "step": 9650 - }, - { - "epoch": 0.15068322206276907, - "grad_norm": 3.411315679550171, - "learning_rate": 0.00018995549177429753, - "loss": 0.438, - "step": 9660 - }, - { - "epoch": 0.15083920883509078, - "grad_norm": 3.1558773517608643, - "learning_rate": 0.00018994509265614278, - "loss": 0.3188, - "step": 9670 - }, - { - "epoch": 0.15099519560741248, - "grad_norm": 1.3154019117355347, - "learning_rate": 0.00018993469353798797, - "loss": 0.3454, - "step": 9680 - }, - { - "epoch": 0.1511511823797342, - "grad_norm": 0.18959377706050873, - "learning_rate": 0.00018992429441983322, - "loss": 0.2291, - "step": 9690 - }, - { - "epoch": 0.15130716915205591, - "grad_norm": 3.3683054447174072, - "learning_rate": 0.0001899138953016784, - "loss": 0.5085, - "step": 9700 - }, - { - "epoch": 0.15146315592437762, - "grad_norm": 1.4889847040176392, - "learning_rate": 0.00018990349618352366, - "loss": 0.1919, - "step": 9710 - }, - { - "epoch": 0.15161914269669932, - "grad_norm": 3.172668933868408, - "learning_rate": 0.00018989309706536885, - "loss": 0.397, - "step": 9720 - }, - { - "epoch": 0.15177512946902103, - "grad_norm": 3.7615721225738525, - "learning_rate": 0.0001898826979472141, - "loss": 0.4941, - "step": 9730 - }, - { - "epoch": 0.15193111624134273, - "grad_norm": 2.056947946548462, - "learning_rate": 0.0001898722988290593, - "loss": 0.3971, - "step": 9740 - }, - { - "epoch": 0.15208710301366443, - "grad_norm": 1.5753828287124634, - "learning_rate": 0.00018986189971090454, - "loss": 0.4617, - "step": 9750 - }, - { - "epoch": 0.15224308978598614, - "grad_norm": 8.168540000915527, - "learning_rate": 0.00018985150059274973, - "loss": 0.2655, - "step": 9760 - }, - { - "epoch": 0.15239907655830787, - "grad_norm": 3.8417530059814453, - "learning_rate": 0.00018984110147459498, - "loss": 0.3281, - "step": 9770 - }, - { - "epoch": 0.15255506333062957, - "grad_norm": 2.040729522705078, - "learning_rate": 0.00018983070235644017, - "loss": 0.3449, - "step": 9780 - }, - { - "epoch": 0.15271105010295127, - "grad_norm": 2.696732997894287, - "learning_rate": 0.00018982030323828542, - "loss": 0.2323, - "step": 9790 - }, - { - "epoch": 0.15286703687527298, - "grad_norm": 0.111504465341568, - "learning_rate": 0.0001898099041201306, - "loss": 0.2174, - "step": 9800 - }, - { - "epoch": 0.15302302364759468, - "grad_norm": 2.0092639923095703, - "learning_rate": 0.00018979950500197585, - "loss": 0.2481, - "step": 9810 - }, - { - "epoch": 0.15317901041991638, - "grad_norm": 2.0963873863220215, - "learning_rate": 0.00018978910588382107, - "loss": 0.2394, - "step": 9820 - }, - { - "epoch": 0.1533349971922381, - "grad_norm": 1.9664541482925415, - "learning_rate": 0.00018977870676566627, - "loss": 0.3264, - "step": 9830 - }, - { - "epoch": 0.15349098396455982, - "grad_norm": 2.2150211334228516, - "learning_rate": 0.00018976830764751151, - "loss": 0.4045, - "step": 9840 - }, - { - "epoch": 0.15364697073688152, - "grad_norm": 0.4570525586605072, - "learning_rate": 0.0001897579085293567, - "loss": 0.1828, - "step": 9850 - }, - { - "epoch": 0.15380295750920323, - "grad_norm": 8.73476505279541, - "learning_rate": 0.00018974750941120195, - "loss": 0.3996, - "step": 9860 - }, - { - "epoch": 0.15395894428152493, - "grad_norm": 0.4760019779205322, - "learning_rate": 0.00018973711029304715, - "loss": 0.2336, - "step": 9870 - }, - { - "epoch": 0.15411493105384663, - "grad_norm": 4.7811808586120605, - "learning_rate": 0.0001897267111748924, - "loss": 0.4498, - "step": 9880 - }, - { - "epoch": 0.15427091782616834, - "grad_norm": 7.987946510314941, - "learning_rate": 0.00018971631205673758, - "loss": 0.2823, - "step": 9890 - }, - { - "epoch": 0.15442690459849004, - "grad_norm": 1.7480034828186035, - "learning_rate": 0.00018970591293858283, - "loss": 0.3611, - "step": 9900 - }, - { - "epoch": 0.15458289137081174, - "grad_norm": 3.877750873565674, - "learning_rate": 0.00018969551382042802, - "loss": 0.2978, - "step": 9910 - }, - { - "epoch": 0.15473887814313347, - "grad_norm": 2.572160243988037, - "learning_rate": 0.00018968511470227327, - "loss": 0.3594, - "step": 9920 - }, - { - "epoch": 0.15489486491545518, - "grad_norm": 2.632244348526001, - "learning_rate": 0.00018967471558411846, - "loss": 0.3699, - "step": 9930 - }, - { - "epoch": 0.15505085168777688, - "grad_norm": 3.1230275630950928, - "learning_rate": 0.0001896643164659637, - "loss": 0.2424, - "step": 9940 - }, - { - "epoch": 0.15520683846009858, - "grad_norm": 2.4213674068450928, - "learning_rate": 0.0001896539173478089, - "loss": 0.252, - "step": 9950 - }, - { - "epoch": 0.1553628252324203, - "grad_norm": 1.5046018362045288, - "learning_rate": 0.00018964351822965415, - "loss": 0.4245, - "step": 9960 - }, - { - "epoch": 0.155518812004742, - "grad_norm": 2.1277291774749756, - "learning_rate": 0.00018963311911149934, - "loss": 0.3583, - "step": 9970 - }, - { - "epoch": 0.1556747987770637, - "grad_norm": 0.7622117400169373, - "learning_rate": 0.0001896227199933446, - "loss": 0.2141, - "step": 9980 - }, - { - "epoch": 0.15583078554938543, - "grad_norm": 7.480571746826172, - "learning_rate": 0.00018961232087518978, - "loss": 0.4933, - "step": 9990 - }, - { - "epoch": 0.15598677232170713, - "grad_norm": 3.018193006515503, - "learning_rate": 0.00018960192175703503, - "loss": 0.258, - "step": 10000 - } - ], - "logging_steps": 10, - "max_steps": 192324, - "num_input_tokens_seen": 0, - "num_train_epochs": 3, - "save_steps": 500, - "stateful_callbacks": { - "TrainerControl": { - "args": { - "should_epoch_stop": false, - "should_evaluate": false, - "should_log": false, - "should_save": true, - "should_training_stop": false - }, - "attributes": {} - } - }, - "total_flos": 6089578905600000.0, - "train_batch_size": 1, - "trial_name": null, - "trial_params": null -}