| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 8.69703905020741, |
| "eval_steps": 400, |
| "global_step": 30400, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.014304105278214848, |
| "grad_norm": 3.5871307485399613, |
| "learning_rate": 1.6666666666666667e-06, |
| "loss": 4.2171, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.028608210556429696, |
| "grad_norm": 8.578020193760794, |
| "learning_rate": 3.3333333333333333e-06, |
| "loss": 3.5086, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.04291231583464454, |
| "grad_norm": 3.453360153921811, |
| "learning_rate": 5e-06, |
| "loss": 2.7411, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.05721642111285939, |
| "grad_norm": 6.637419959801433, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": 2.5587, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.07152052639107424, |
| "grad_norm": 5.39103050730964, |
| "learning_rate": 8.333333333333334e-06, |
| "loss": 2.4626, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.08582463166928908, |
| "grad_norm": 10.145361416938202, |
| "learning_rate": 1e-05, |
| "loss": 2.3934, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.10012873694750393, |
| "grad_norm": 7.783722807777708, |
| "learning_rate": 9.999953760295448e-06, |
| "loss": 2.272, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.11443284222571878, |
| "grad_norm": 3.65791345167478, |
| "learning_rate": 9.999815042132062e-06, |
| "loss": 2.2033, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.11443284222571878, |
| "eval_loss": 2.1356396675109863, |
| "eval_runtime": 13.2089, |
| "eval_samples_per_second": 75.707, |
| "eval_steps_per_second": 2.423, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.12873694750393364, |
| "grad_norm": 8.621408307551274, |
| "learning_rate": 9.999583848360633e-06, |
| "loss": 2.1457, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.1430410527821485, |
| "grad_norm": 6.2605099034671, |
| "learning_rate": 9.999260183732424e-06, |
| "loss": 2.1022, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.1573451580603633, |
| "grad_norm": 7.684157630671387, |
| "learning_rate": 9.998844054899058e-06, |
| "loss": 2.0793, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.17164926333857816, |
| "grad_norm": 1.7771086513044794, |
| "learning_rate": 9.998335470412393e-06, |
| "loss": 2.0783, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.185953368616793, |
| "grad_norm": 4.839674439593447, |
| "learning_rate": 9.997734440724333e-06, |
| "loss": 2.0455, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.20025747389500786, |
| "grad_norm": 1.7547598316419108, |
| "learning_rate": 9.997040978186633e-06, |
| "loss": 2.0206, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.21456157917322272, |
| "grad_norm": 2.5514452897242035, |
| "learning_rate": 9.996255097050624e-06, |
| "loss": 2.003, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.22886568445143757, |
| "grad_norm": 3.9407907366389976, |
| "learning_rate": 9.995376813466934e-06, |
| "loss": 1.9948, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.22886568445143757, |
| "eval_loss": 1.9659738540649414, |
| "eval_runtime": 13.202, |
| "eval_samples_per_second": 75.746, |
| "eval_steps_per_second": 2.424, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.24316978972965242, |
| "grad_norm": 1.7301759518540019, |
| "learning_rate": 9.994406145485151e-06, |
| "loss": 1.9861, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.25747389500786727, |
| "grad_norm": 4.606477014527872, |
| "learning_rate": 9.993343113053454e-06, |
| "loss": 1.9713, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.2717780002860821, |
| "grad_norm": 2.8102943343276183, |
| "learning_rate": 9.992187738018203e-06, |
| "loss": 1.9675, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.286082105564297, |
| "grad_norm": 1.6938293305410583, |
| "learning_rate": 9.99094004412348e-06, |
| "loss": 1.9556, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.3003862108425118, |
| "grad_norm": 3.4489466081512044, |
| "learning_rate": 9.989600057010625e-06, |
| "loss": 1.9417, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.3146903161207266, |
| "grad_norm": 1.454314969912362, |
| "learning_rate": 9.988167804217682e-06, |
| "loss": 1.9361, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.3289944213989415, |
| "grad_norm": 1.168052123243911, |
| "learning_rate": 9.986643315178848e-06, |
| "loss": 1.9264, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.3432985266771563, |
| "grad_norm": 1.7334339924315771, |
| "learning_rate": 9.98502662122387e-06, |
| "loss": 1.9308, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.3432985266771563, |
| "eval_loss": 1.8975478410720825, |
| "eval_runtime": 13.3147, |
| "eval_samples_per_second": 75.105, |
| "eval_steps_per_second": 2.403, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.3576026319553712, |
| "grad_norm": 1.5051784354611382, |
| "learning_rate": 9.983317755577392e-06, |
| "loss": 1.9208, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.371906737233586, |
| "grad_norm": 1.5926471600228091, |
| "learning_rate": 9.981516753358274e-06, |
| "loss": 1.9156, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.3862108425118009, |
| "grad_norm": 1.4680846710500992, |
| "learning_rate": 9.979623651578881e-06, |
| "loss": 1.9003, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.40051494779001573, |
| "grad_norm": 1.3763604778464051, |
| "learning_rate": 9.977638489144308e-06, |
| "loss": 1.9069, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.4148190530682306, |
| "grad_norm": 1.524436140637499, |
| "learning_rate": 9.975561306851585e-06, |
| "loss": 1.8973, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.42912315834644543, |
| "grad_norm": 1.6595586978206922, |
| "learning_rate": 9.973392147388847e-06, |
| "loss": 1.8994, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.4434272636246603, |
| "grad_norm": 1.5658370421245245, |
| "learning_rate": 9.971131055334445e-06, |
| "loss": 1.8875, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.45773136890287514, |
| "grad_norm": 1.9874689334717497, |
| "learning_rate": 9.968778077156035e-06, |
| "loss": 1.8885, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.45773136890287514, |
| "eval_loss": 1.8600050210952759, |
| "eval_runtime": 13.4365, |
| "eval_samples_per_second": 74.424, |
| "eval_steps_per_second": 2.382, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.47203547418109, |
| "grad_norm": 1.391686515937597, |
| "learning_rate": 9.966333261209625e-06, |
| "loss": 1.8825, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.48633957945930484, |
| "grad_norm": 1.250031324432951, |
| "learning_rate": 9.96379665773858e-06, |
| "loss": 1.8812, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.5006436847375196, |
| "grad_norm": 0.937500185481468, |
| "learning_rate": 9.961168318872583e-06, |
| "loss": 1.873, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.5149477900157345, |
| "grad_norm": 1.3831225145690145, |
| "learning_rate": 9.958448298626576e-06, |
| "loss": 1.8733, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.5292518952939493, |
| "grad_norm": 2.8853368812546845, |
| "learning_rate": 9.95563665289964e-06, |
| "loss": 1.8607, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.5435560005721642, |
| "grad_norm": 1.8333114897013831, |
| "learning_rate": 9.952733439473847e-06, |
| "loss": 1.867, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.557860105850379, |
| "grad_norm": 1.4955344425097794, |
| "learning_rate": 9.94973871801308e-06, |
| "loss": 1.8668, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.572164211128594, |
| "grad_norm": 1.058766324989062, |
| "learning_rate": 9.946652550061798e-06, |
| "loss": 1.8468, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.572164211128594, |
| "eval_loss": 1.8319525718688965, |
| "eval_runtime": 13.2128, |
| "eval_samples_per_second": 75.684, |
| "eval_steps_per_second": 2.422, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.5864683164068087, |
| "grad_norm": 1.061225891590699, |
| "learning_rate": 9.943474999043775e-06, |
| "loss": 1.8513, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.6007724216850236, |
| "grad_norm": 1.8307541279709998, |
| "learning_rate": 9.9402061302608e-06, |
| "loss": 1.8489, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.6150765269632384, |
| "grad_norm": 1.3493002065305932, |
| "learning_rate": 9.93684601089133e-06, |
| "loss": 1.8428, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.6293806322414532, |
| "grad_norm": 1.3269702867770274, |
| "learning_rate": 9.933394709989109e-06, |
| "loss": 1.8485, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.6436847375196681, |
| "grad_norm": 1.3587430620093466, |
| "learning_rate": 9.92985229848175e-06, |
| "loss": 1.8323, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.657988842797883, |
| "grad_norm": 1.6220456911234815, |
| "learning_rate": 9.926218849169284e-06, |
| "loss": 1.8417, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.6722929480760979, |
| "grad_norm": 1.3306733931984986, |
| "learning_rate": 9.922494436722653e-06, |
| "loss": 1.8319, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.6865970533543126, |
| "grad_norm": 1.841456096268586, |
| "learning_rate": 9.91867913768218e-06, |
| "loss": 1.8221, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.6865970533543126, |
| "eval_loss": 1.8098633289337158, |
| "eval_runtime": 13.2499, |
| "eval_samples_per_second": 75.472, |
| "eval_steps_per_second": 2.415, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.7009011586325276, |
| "grad_norm": 1.0743552491621962, |
| "learning_rate": 9.914773030456001e-06, |
| "loss": 1.8264, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.7152052639107424, |
| "grad_norm": 0.9817511549306134, |
| "learning_rate": 9.910776195318448e-06, |
| "loss": 1.8339, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.7295093691889573, |
| "grad_norm": 2.5586651091316166, |
| "learning_rate": 9.906688714408396e-06, |
| "loss": 1.8361, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.743813474467172, |
| "grad_norm": 1.3521713422261818, |
| "learning_rate": 9.902510671727583e-06, |
| "loss": 1.8172, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.758117579745387, |
| "grad_norm": 1.196264818533523, |
| "learning_rate": 9.898242153138882e-06, |
| "loss": 1.815, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.7724216850236018, |
| "grad_norm": 1.1236416672692373, |
| "learning_rate": 9.89388324636453e-06, |
| "loss": 1.8217, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.7867257903018167, |
| "grad_norm": 1.0482615125286014, |
| "learning_rate": 9.889434040984333e-06, |
| "loss": 1.8111, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.8010298955800315, |
| "grad_norm": 1.2877895138097581, |
| "learning_rate": 9.88489462843382e-06, |
| "loss": 1.8139, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.8010298955800315, |
| "eval_loss": 1.7917475700378418, |
| "eval_runtime": 13.2172, |
| "eval_samples_per_second": 75.659, |
| "eval_steps_per_second": 2.421, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.8153340008582464, |
| "grad_norm": 1.3263953998902482, |
| "learning_rate": 9.880265102002369e-06, |
| "loss": 1.7969, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.8296381061364612, |
| "grad_norm": 0.8011721107369111, |
| "learning_rate": 9.875545556831283e-06, |
| "loss": 1.8123, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.843942211414676, |
| "grad_norm": 1.452056398499391, |
| "learning_rate": 9.870736089911836e-06, |
| "loss": 1.8137, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.8582463166928909, |
| "grad_norm": 1.1618490357992515, |
| "learning_rate": 9.865836800083291e-06, |
| "loss": 1.812, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.8725504219711057, |
| "grad_norm": 1.2693408191050113, |
| "learning_rate": 9.860847788030852e-06, |
| "loss": 1.7989, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.8868545272493206, |
| "grad_norm": 1.1394593474503365, |
| "learning_rate": 9.855769156283604e-06, |
| "loss": 1.8068, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.9011586325275354, |
| "grad_norm": 1.739305216161417, |
| "learning_rate": 9.850601009212408e-06, |
| "loss": 1.8026, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.9154627378057503, |
| "grad_norm": 0.9509459350795362, |
| "learning_rate": 9.845343453027747e-06, |
| "loss": 1.8055, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.9154627378057503, |
| "eval_loss": 1.7801611423492432, |
| "eval_runtime": 13.218, |
| "eval_samples_per_second": 75.654, |
| "eval_steps_per_second": 2.421, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.9297668430839651, |
| "grad_norm": 1.279124709454881, |
| "learning_rate": 9.839996595777552e-06, |
| "loss": 1.7968, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.94407094836218, |
| "grad_norm": 1.092437156034456, |
| "learning_rate": 9.83456054734498e-06, |
| "loss": 1.7892, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.9583750536403948, |
| "grad_norm": 0.9619111877602855, |
| "learning_rate": 9.829035419446156e-06, |
| "loss": 1.7951, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.9726791589186097, |
| "grad_norm": 0.9476479377652314, |
| "learning_rate": 9.823421325627865e-06, |
| "loss": 1.8003, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.9869832641968245, |
| "grad_norm": 0.7761637732992411, |
| "learning_rate": 9.81771838126524e-06, |
| "loss": 1.7852, |
| "step": 3450 |
| }, |
| { |
| "epoch": 1.0012873694750393, |
| "grad_norm": 1.6427439742947594, |
| "learning_rate": 9.811926703559374e-06, |
| "loss": 1.7813, |
| "step": 3500 |
| }, |
| { |
| "epoch": 1.0155914747532542, |
| "grad_norm": 1.1995527718721077, |
| "learning_rate": 9.806046411534916e-06, |
| "loss": 1.7693, |
| "step": 3550 |
| }, |
| { |
| "epoch": 1.029895580031469, |
| "grad_norm": 0.6804399746581633, |
| "learning_rate": 9.800077626037633e-06, |
| "loss": 1.7779, |
| "step": 3600 |
| }, |
| { |
| "epoch": 1.029895580031469, |
| "eval_loss": 1.770551323890686, |
| "eval_runtime": 13.1907, |
| "eval_samples_per_second": 75.811, |
| "eval_steps_per_second": 2.426, |
| "step": 3600 |
| }, |
| { |
| "epoch": 1.044199685309684, |
| "grad_norm": 0.9739432549705861, |
| "learning_rate": 9.794020469731915e-06, |
| "loss": 1.7844, |
| "step": 3650 |
| }, |
| { |
| "epoch": 1.0585037905878987, |
| "grad_norm": 1.24072763349112, |
| "learning_rate": 9.787875067098257e-06, |
| "loss": 1.7826, |
| "step": 3700 |
| }, |
| { |
| "epoch": 1.0728078958661136, |
| "grad_norm": 1.0497808533387654, |
| "learning_rate": 9.781641544430703e-06, |
| "loss": 1.7875, |
| "step": 3750 |
| }, |
| { |
| "epoch": 1.0871120011443285, |
| "grad_norm": 1.608878092364201, |
| "learning_rate": 9.775320029834255e-06, |
| "loss": 1.7676, |
| "step": 3800 |
| }, |
| { |
| "epoch": 1.1014161064225432, |
| "grad_norm": 0.8265375857322264, |
| "learning_rate": 9.76891065322223e-06, |
| "loss": 1.7758, |
| "step": 3850 |
| }, |
| { |
| "epoch": 1.115720211700758, |
| "grad_norm": 0.8074509140095618, |
| "learning_rate": 9.762413546313597e-06, |
| "loss": 1.773, |
| "step": 3900 |
| }, |
| { |
| "epoch": 1.130024316978973, |
| "grad_norm": 1.2955529307297018, |
| "learning_rate": 9.755828842630269e-06, |
| "loss": 1.7653, |
| "step": 3950 |
| }, |
| { |
| "epoch": 1.144328422257188, |
| "grad_norm": 0.9723230317765355, |
| "learning_rate": 9.749156677494357e-06, |
| "loss": 1.7813, |
| "step": 4000 |
| }, |
| { |
| "epoch": 1.144328422257188, |
| "eval_loss": 1.7602086067199707, |
| "eval_runtime": 13.0715, |
| "eval_samples_per_second": 76.502, |
| "eval_steps_per_second": 2.448, |
| "step": 4000 |
| }, |
| { |
| "epoch": 1.1587755685881849, |
| "grad_norm": 1.6203576789230651, |
| "learning_rate": 9.742397188025394e-06, |
| "loss": 1.7709, |
| "step": 4050 |
| }, |
| { |
| "epoch": 1.1730796738663996, |
| "grad_norm": 1.1690135821413727, |
| "learning_rate": 9.735550513137513e-06, |
| "loss": 1.7583, |
| "step": 4100 |
| }, |
| { |
| "epoch": 1.1873837791446145, |
| "grad_norm": 1.2584964373503669, |
| "learning_rate": 9.728616793536588e-06, |
| "loss": 1.7653, |
| "step": 4150 |
| }, |
| { |
| "epoch": 1.2016878844228294, |
| "grad_norm": 1.022986798804273, |
| "learning_rate": 9.721596171717352e-06, |
| "loss": 1.7661, |
| "step": 4200 |
| }, |
| { |
| "epoch": 1.2159919897010443, |
| "grad_norm": 0.9715095488392225, |
| "learning_rate": 9.714488791960463e-06, |
| "loss": 1.7745, |
| "step": 4250 |
| }, |
| { |
| "epoch": 1.230296094979259, |
| "grad_norm": 1.1370122008173429, |
| "learning_rate": 9.707294800329536e-06, |
| "loss": 1.7684, |
| "step": 4300 |
| }, |
| { |
| "epoch": 1.2446002002574739, |
| "grad_norm": 0.9536116406695613, |
| "learning_rate": 9.700014344668152e-06, |
| "loss": 1.7606, |
| "step": 4350 |
| }, |
| { |
| "epoch": 1.2589043055356888, |
| "grad_norm": 0.7353885681465049, |
| "learning_rate": 9.692647574596803e-06, |
| "loss": 1.7633, |
| "step": 4400 |
| }, |
| { |
| "epoch": 1.2589043055356888, |
| "eval_loss": 1.752835988998413, |
| "eval_runtime": 13.2443, |
| "eval_samples_per_second": 75.504, |
| "eval_steps_per_second": 2.416, |
| "step": 4400 |
| }, |
| { |
| "epoch": 1.2732084108139037, |
| "grad_norm": 1.1802783991524877, |
| "learning_rate": 9.685194641509837e-06, |
| "loss": 1.7686, |
| "step": 4450 |
| }, |
| { |
| "epoch": 1.2875125160921184, |
| "grad_norm": 0.9318456023565893, |
| "learning_rate": 9.677655698572326e-06, |
| "loss": 1.7556, |
| "step": 4500 |
| }, |
| { |
| "epoch": 1.3018166213703333, |
| "grad_norm": 1.0437155245378358, |
| "learning_rate": 9.670030900716941e-06, |
| "loss": 1.7571, |
| "step": 4550 |
| }, |
| { |
| "epoch": 1.3161207266485482, |
| "grad_norm": 1.0912943460238356, |
| "learning_rate": 9.662320404640743e-06, |
| "loss": 1.7546, |
| "step": 4600 |
| }, |
| { |
| "epoch": 1.3304248319267629, |
| "grad_norm": 1.0063174963239234, |
| "learning_rate": 9.654524368801982e-06, |
| "loss": 1.7644, |
| "step": 4650 |
| }, |
| { |
| "epoch": 1.3447289372049778, |
| "grad_norm": 1.0417193249219145, |
| "learning_rate": 9.646642953416835e-06, |
| "loss": 1.7525, |
| "step": 4700 |
| }, |
| { |
| "epoch": 1.3590330424831927, |
| "grad_norm": 0.698422675064761, |
| "learning_rate": 9.638676320456109e-06, |
| "loss": 1.7586, |
| "step": 4750 |
| }, |
| { |
| "epoch": 1.3733371477614076, |
| "grad_norm": 1.1910335445073554, |
| "learning_rate": 9.630624633641918e-06, |
| "loss": 1.7528, |
| "step": 4800 |
| }, |
| { |
| "epoch": 1.3733371477614076, |
| "eval_loss": 1.7467565536499023, |
| "eval_runtime": 13.1696, |
| "eval_samples_per_second": 75.932, |
| "eval_steps_per_second": 2.43, |
| "step": 4800 |
| }, |
| { |
| "epoch": 1.3876412530396225, |
| "grad_norm": 0.860266713830524, |
| "learning_rate": 9.622488058444313e-06, |
| "loss": 1.7526, |
| "step": 4850 |
| }, |
| { |
| "epoch": 1.4019453583178372, |
| "grad_norm": 0.6381947527857678, |
| "learning_rate": 9.614266762077891e-06, |
| "loss": 1.7548, |
| "step": 4900 |
| }, |
| { |
| "epoch": 1.416249463596052, |
| "grad_norm": 0.7554075411255639, |
| "learning_rate": 9.605960913498342e-06, |
| "loss": 1.7481, |
| "step": 4950 |
| }, |
| { |
| "epoch": 1.430553568874267, |
| "grad_norm": 1.2154500150186984, |
| "learning_rate": 9.597570683398996e-06, |
| "loss": 1.7584, |
| "step": 5000 |
| }, |
| { |
| "epoch": 1.4448576741524817, |
| "grad_norm": 0.6439874478006935, |
| "learning_rate": 9.5890962442073e-06, |
| "loss": 1.7445, |
| "step": 5050 |
| }, |
| { |
| "epoch": 1.4591617794306966, |
| "grad_norm": 1.1984560510642397, |
| "learning_rate": 9.580537770081285e-06, |
| "loss": 1.7442, |
| "step": 5100 |
| }, |
| { |
| "epoch": 1.4734658847089115, |
| "grad_norm": 1.1871050477497047, |
| "learning_rate": 9.57189543690598e-06, |
| "loss": 1.7498, |
| "step": 5150 |
| }, |
| { |
| "epoch": 1.4877699899871262, |
| "grad_norm": 0.7133766609729304, |
| "learning_rate": 9.563169422289798e-06, |
| "loss": 1.7423, |
| "step": 5200 |
| }, |
| { |
| "epoch": 1.4877699899871262, |
| "eval_loss": 1.7412633895874023, |
| "eval_runtime": 13.1818, |
| "eval_samples_per_second": 75.862, |
| "eval_steps_per_second": 2.428, |
| "step": 5200 |
| }, |
| { |
| "epoch": 1.5020740952653413, |
| "grad_norm": 0.7612313099084432, |
| "learning_rate": 9.554359905560887e-06, |
| "loss": 1.748, |
| "step": 5250 |
| }, |
| { |
| "epoch": 1.516378200543556, |
| "grad_norm": 0.750224706069401, |
| "learning_rate": 9.54546706776345e-06, |
| "loss": 1.7435, |
| "step": 5300 |
| }, |
| { |
| "epoch": 1.530682305821771, |
| "grad_norm": 0.9331760005679197, |
| "learning_rate": 9.536491091654018e-06, |
| "loss": 1.7324, |
| "step": 5350 |
| }, |
| { |
| "epoch": 1.5449864110999858, |
| "grad_norm": 0.8613210624917503, |
| "learning_rate": 9.527432161697696e-06, |
| "loss": 1.751, |
| "step": 5400 |
| }, |
| { |
| "epoch": 1.5592905163782005, |
| "grad_norm": 4.862833366342641, |
| "learning_rate": 9.518290464064365e-06, |
| "loss": 1.7365, |
| "step": 5450 |
| }, |
| { |
| "epoch": 1.5735946216564154, |
| "grad_norm": 0.7838749485611254, |
| "learning_rate": 9.509066186624872e-06, |
| "loss": 1.7399, |
| "step": 5500 |
| }, |
| { |
| "epoch": 1.5878987269346303, |
| "grad_norm": 0.798441318649249, |
| "learning_rate": 9.499759518947156e-06, |
| "loss": 1.737, |
| "step": 5550 |
| }, |
| { |
| "epoch": 1.602202832212845, |
| "grad_norm": 0.7078515946606195, |
| "learning_rate": 9.490370652292357e-06, |
| "loss": 1.7412, |
| "step": 5600 |
| }, |
| { |
| "epoch": 1.602202832212845, |
| "eval_loss": 1.7350177764892578, |
| "eval_runtime": 13.1662, |
| "eval_samples_per_second": 75.952, |
| "eval_steps_per_second": 2.43, |
| "step": 5600 |
| }, |
| { |
| "epoch": 1.61650693749106, |
| "grad_norm": 0.8939699763780663, |
| "learning_rate": 9.480899779610883e-06, |
| "loss": 1.7485, |
| "step": 5650 |
| }, |
| { |
| "epoch": 1.6308110427692748, |
| "grad_norm": 0.7241948247907901, |
| "learning_rate": 9.471347095538448e-06, |
| "loss": 1.7338, |
| "step": 5700 |
| }, |
| { |
| "epoch": 1.6451151480474895, |
| "grad_norm": 0.7753555958595959, |
| "learning_rate": 9.461712796392067e-06, |
| "loss": 1.7423, |
| "step": 5750 |
| }, |
| { |
| "epoch": 1.6594192533257046, |
| "grad_norm": 0.9453545668822089, |
| "learning_rate": 9.45199708016603e-06, |
| "loss": 1.7411, |
| "step": 5800 |
| }, |
| { |
| "epoch": 1.6737233586039193, |
| "grad_norm": 0.8069354542585767, |
| "learning_rate": 9.442200146527824e-06, |
| "loss": 1.7341, |
| "step": 5850 |
| }, |
| { |
| "epoch": 1.6880274638821342, |
| "grad_norm": 0.9570095002200463, |
| "learning_rate": 9.432322196814032e-06, |
| "loss": 1.7309, |
| "step": 5900 |
| }, |
| { |
| "epoch": 1.7023315691603491, |
| "grad_norm": 0.778079342980594, |
| "learning_rate": 9.422363434026205e-06, |
| "loss": 1.7331, |
| "step": 5950 |
| }, |
| { |
| "epoch": 1.7166356744385638, |
| "grad_norm": 0.8582782820893037, |
| "learning_rate": 9.41232406282667e-06, |
| "loss": 1.7375, |
| "step": 6000 |
| }, |
| { |
| "epoch": 1.7166356744385638, |
| "eval_loss": 1.7310324907302856, |
| "eval_runtime": 13.1951, |
| "eval_samples_per_second": 75.786, |
| "eval_steps_per_second": 2.425, |
| "step": 6000 |
| }, |
| { |
| "epoch": 1.7309397797167787, |
| "grad_norm": 0.657112371502439, |
| "learning_rate": 9.402204289534344e-06, |
| "loss": 1.725, |
| "step": 6050 |
| }, |
| { |
| "epoch": 1.7452438849949936, |
| "grad_norm": 1.1123814501158173, |
| "learning_rate": 9.392004322120484e-06, |
| "loss": 1.7303, |
| "step": 6100 |
| }, |
| { |
| "epoch": 1.7595479902732083, |
| "grad_norm": 0.9499171364049573, |
| "learning_rate": 9.381724370204414e-06, |
| "loss": 1.7203, |
| "step": 6150 |
| }, |
| { |
| "epoch": 1.7738520955514234, |
| "grad_norm": 1.095712190889447, |
| "learning_rate": 9.371364645049216e-06, |
| "loss": 1.7291, |
| "step": 6200 |
| }, |
| { |
| "epoch": 1.788156200829638, |
| "grad_norm": 1.0199371888229702, |
| "learning_rate": 9.360925359557397e-06, |
| "loss": 1.7155, |
| "step": 6250 |
| }, |
| { |
| "epoch": 1.8024603061078528, |
| "grad_norm": 0.827904086276593, |
| "learning_rate": 9.3504067282665e-06, |
| "loss": 1.7404, |
| "step": 6300 |
| }, |
| { |
| "epoch": 1.816764411386068, |
| "grad_norm": 0.8252103968718575, |
| "learning_rate": 9.339808967344701e-06, |
| "loss": 1.7275, |
| "step": 6350 |
| }, |
| { |
| "epoch": 1.8310685166642826, |
| "grad_norm": 0.8608721928591242, |
| "learning_rate": 9.329132294586374e-06, |
| "loss": 1.7257, |
| "step": 6400 |
| }, |
| { |
| "epoch": 1.8310685166642826, |
| "eval_loss": 1.7276860475540161, |
| "eval_runtime": 13.1438, |
| "eval_samples_per_second": 76.082, |
| "eval_steps_per_second": 2.435, |
| "step": 6400 |
| }, |
| { |
| "epoch": 1.8453726219424975, |
| "grad_norm": 1.0655198477534613, |
| "learning_rate": 9.318376929407606e-06, |
| "loss": 1.7219, |
| "step": 6450 |
| }, |
| { |
| "epoch": 1.8596767272207124, |
| "grad_norm": 0.8202058864526146, |
| "learning_rate": 9.307543092841688e-06, |
| "loss": 1.7219, |
| "step": 6500 |
| }, |
| { |
| "epoch": 1.873980832498927, |
| "grad_norm": 0.9119156406563158, |
| "learning_rate": 9.296631007534576e-06, |
| "loss": 1.7232, |
| "step": 6550 |
| }, |
| { |
| "epoch": 1.888284937777142, |
| "grad_norm": 0.8561460550361113, |
| "learning_rate": 9.285640897740316e-06, |
| "loss": 1.718, |
| "step": 6600 |
| }, |
| { |
| "epoch": 1.902589043055357, |
| "grad_norm": 0.9549020817113293, |
| "learning_rate": 9.27457298931643e-06, |
| "loss": 1.7348, |
| "step": 6650 |
| }, |
| { |
| "epoch": 1.9168931483335716, |
| "grad_norm": 0.6729589207861246, |
| "learning_rate": 9.263427509719287e-06, |
| "loss": 1.7175, |
| "step": 6700 |
| }, |
| { |
| "epoch": 1.9311972536117867, |
| "grad_norm": 0.709786130314777, |
| "learning_rate": 9.252204687999401e-06, |
| "loss": 1.7287, |
| "step": 6750 |
| }, |
| { |
| "epoch": 1.9455013588900014, |
| "grad_norm": 0.9827727167496969, |
| "learning_rate": 9.240904754796767e-06, |
| "loss": 1.7241, |
| "step": 6800 |
| }, |
| { |
| "epoch": 1.9455013588900014, |
| "eval_loss": 1.7200103998184204, |
| "eval_runtime": 13.113, |
| "eval_samples_per_second": 76.26, |
| "eval_steps_per_second": 2.44, |
| "step": 6800 |
| }, |
| { |
| "epoch": 1.9598054641682163, |
| "grad_norm": 0.8218927694104888, |
| "learning_rate": 9.22952794233608e-06, |
| "loss": 1.7334, |
| "step": 6850 |
| }, |
| { |
| "epoch": 1.9741095694464312, |
| "grad_norm": 0.9202146449681148, |
| "learning_rate": 9.218074484421977e-06, |
| "loss": 1.7176, |
| "step": 6900 |
| }, |
| { |
| "epoch": 1.988413674724646, |
| "grad_norm": 0.9757466555134802, |
| "learning_rate": 9.206544616434249e-06, |
| "loss": 1.716, |
| "step": 6950 |
| }, |
| { |
| "epoch": 2.0027177800028606, |
| "grad_norm": 0.9665391581844135, |
| "learning_rate": 9.194938575322973e-06, |
| "loss": 1.7104, |
| "step": 7000 |
| }, |
| { |
| "epoch": 2.0170218852810757, |
| "grad_norm": 0.747601621726994, |
| "learning_rate": 9.183256599603672e-06, |
| "loss": 1.7045, |
| "step": 7050 |
| }, |
| { |
| "epoch": 2.0313259905592904, |
| "grad_norm": 0.590573479620199, |
| "learning_rate": 9.171498929352388e-06, |
| "loss": 1.7012, |
| "step": 7100 |
| }, |
| { |
| "epoch": 2.0456300958375055, |
| "grad_norm": 0.7932008874801502, |
| "learning_rate": 9.159665806200766e-06, |
| "loss": 1.7017, |
| "step": 7150 |
| }, |
| { |
| "epoch": 2.0599342011157202, |
| "grad_norm": 0.824718775314277, |
| "learning_rate": 9.147757473331082e-06, |
| "loss": 1.7019, |
| "step": 7200 |
| }, |
| { |
| "epoch": 2.0599342011157202, |
| "eval_loss": 1.7177115678787231, |
| "eval_runtime": 13.1014, |
| "eval_samples_per_second": 76.328, |
| "eval_steps_per_second": 2.442, |
| "step": 7200 |
| }, |
| { |
| "epoch": 2.074238306393935, |
| "grad_norm": 1.080101464877567, |
| "learning_rate": 9.135774175471244e-06, |
| "loss": 1.7056, |
| "step": 7250 |
| }, |
| { |
| "epoch": 2.08854241167215, |
| "grad_norm": 0.6161967559479131, |
| "learning_rate": 9.123716158889765e-06, |
| "loss": 1.6892, |
| "step": 7300 |
| }, |
| { |
| "epoch": 2.1028465169503647, |
| "grad_norm": 0.6259364978680784, |
| "learning_rate": 9.111583671390697e-06, |
| "loss": 1.6876, |
| "step": 7350 |
| }, |
| { |
| "epoch": 2.1171506222285794, |
| "grad_norm": 0.9054346826665011, |
| "learning_rate": 9.09937696230855e-06, |
| "loss": 1.7004, |
| "step": 7400 |
| }, |
| { |
| "epoch": 2.1314547275067945, |
| "grad_norm": 0.8115910709185014, |
| "learning_rate": 9.087096282503152e-06, |
| "loss": 1.6951, |
| "step": 7450 |
| }, |
| { |
| "epoch": 2.1457588327850092, |
| "grad_norm": 0.939188480700838, |
| "learning_rate": 9.074741884354507e-06, |
| "loss": 1.698, |
| "step": 7500 |
| }, |
| { |
| "epoch": 2.1600629380632244, |
| "grad_norm": 0.5461813129993677, |
| "learning_rate": 9.062314021757603e-06, |
| "loss": 1.6953, |
| "step": 7550 |
| }, |
| { |
| "epoch": 2.174367043341439, |
| "grad_norm": 0.8915398126133827, |
| "learning_rate": 9.049812950117191e-06, |
| "loss": 1.6838, |
| "step": 7600 |
| }, |
| { |
| "epoch": 2.174367043341439, |
| "eval_loss": 1.7138569355010986, |
| "eval_runtime": 13.1252, |
| "eval_samples_per_second": 76.189, |
| "eval_steps_per_second": 2.438, |
| "step": 7600 |
| }, |
| { |
| "epoch": 2.1886711486196537, |
| "grad_norm": 0.8119718442892704, |
| "learning_rate": 9.037238926342544e-06, |
| "loss": 1.7012, |
| "step": 7650 |
| }, |
| { |
| "epoch": 2.202975253897869, |
| "grad_norm": 0.6869354062120884, |
| "learning_rate": 9.02459220884217e-06, |
| "loss": 1.6918, |
| "step": 7700 |
| }, |
| { |
| "epoch": 2.2172793591760835, |
| "grad_norm": 0.7341604750585072, |
| "learning_rate": 9.011873057518503e-06, |
| "loss": 1.699, |
| "step": 7750 |
| }, |
| { |
| "epoch": 2.2315834644542982, |
| "grad_norm": 0.8261410417964352, |
| "learning_rate": 8.999081733762568e-06, |
| "loss": 1.6993, |
| "step": 7800 |
| }, |
| { |
| "epoch": 2.2458875697325134, |
| "grad_norm": 0.7611803854476097, |
| "learning_rate": 8.986218500448598e-06, |
| "loss": 1.7069, |
| "step": 7850 |
| }, |
| { |
| "epoch": 2.260191675010728, |
| "grad_norm": 0.771284372878959, |
| "learning_rate": 8.973283621928644e-06, |
| "loss": 1.7018, |
| "step": 7900 |
| }, |
| { |
| "epoch": 2.2744957802889427, |
| "grad_norm": 0.6069248728630589, |
| "learning_rate": 8.96027736402713e-06, |
| "loss": 1.6894, |
| "step": 7950 |
| }, |
| { |
| "epoch": 2.288799885567158, |
| "grad_norm": 0.6935253796751046, |
| "learning_rate": 8.947199994035402e-06, |
| "loss": 1.6857, |
| "step": 8000 |
| }, |
| { |
| "epoch": 2.288799885567158, |
| "eval_loss": 1.709199070930481, |
| "eval_runtime": 13.105, |
| "eval_samples_per_second": 76.307, |
| "eval_steps_per_second": 2.442, |
| "step": 8000 |
| }, |
| { |
| "epoch": 2.3031039908453725, |
| "grad_norm": 0.6395079401863663, |
| "learning_rate": 8.934051780706226e-06, |
| "loss": 1.6867, |
| "step": 8050 |
| }, |
| { |
| "epoch": 2.3174080961235877, |
| "grad_norm": 0.8808561738320105, |
| "learning_rate": 8.920832994248268e-06, |
| "loss": 1.6947, |
| "step": 8100 |
| }, |
| { |
| "epoch": 2.3317122014018024, |
| "grad_norm": 0.6888071930360784, |
| "learning_rate": 8.907543906320542e-06, |
| "loss": 1.7005, |
| "step": 8150 |
| }, |
| { |
| "epoch": 2.346016306680017, |
| "grad_norm": 0.6515296032830782, |
| "learning_rate": 8.894184790026823e-06, |
| "loss": 1.686, |
| "step": 8200 |
| }, |
| { |
| "epoch": 2.360320411958232, |
| "grad_norm": 0.7401595322241383, |
| "learning_rate": 8.880755919910048e-06, |
| "loss": 1.6865, |
| "step": 8250 |
| }, |
| { |
| "epoch": 2.374624517236447, |
| "grad_norm": 0.6000159117604718, |
| "learning_rate": 8.867257571946646e-06, |
| "loss": 1.6996, |
| "step": 8300 |
| }, |
| { |
| "epoch": 2.3889286225146615, |
| "grad_norm": 0.6577366719572108, |
| "learning_rate": 8.853690023540898e-06, |
| "loss": 1.6929, |
| "step": 8350 |
| }, |
| { |
| "epoch": 2.4032327277928767, |
| "grad_norm": 0.7392515614712054, |
| "learning_rate": 8.840053553519216e-06, |
| "loss": 1.6848, |
| "step": 8400 |
| }, |
| { |
| "epoch": 2.4032327277928767, |
| "eval_loss": 1.7050341367721558, |
| "eval_runtime": 13.1691, |
| "eval_samples_per_second": 75.935, |
| "eval_steps_per_second": 2.43, |
| "step": 8400 |
| }, |
| { |
| "epoch": 2.4175368330710914, |
| "grad_norm": 0.9889920645015057, |
| "learning_rate": 8.82634844212442e-06, |
| "loss": 1.6893, |
| "step": 8450 |
| }, |
| { |
| "epoch": 2.431840938349306, |
| "grad_norm": 0.9590160412926145, |
| "learning_rate": 8.81257497100998e-06, |
| "loss": 1.6846, |
| "step": 8500 |
| }, |
| { |
| "epoch": 2.446145043627521, |
| "grad_norm": 0.8138219488107618, |
| "learning_rate": 8.79873342323422e-06, |
| "loss": 1.689, |
| "step": 8550 |
| }, |
| { |
| "epoch": 2.460449148905736, |
| "grad_norm": 0.6771189511098283, |
| "learning_rate": 8.78482408325451e-06, |
| "loss": 1.684, |
| "step": 8600 |
| }, |
| { |
| "epoch": 2.474753254183951, |
| "grad_norm": 0.637569459443737, |
| "learning_rate": 8.770847236921412e-06, |
| "loss": 1.6893, |
| "step": 8650 |
| }, |
| { |
| "epoch": 2.4890573594621657, |
| "grad_norm": 0.8520093583138391, |
| "learning_rate": 8.756803171472817e-06, |
| "loss": 1.6858, |
| "step": 8700 |
| }, |
| { |
| "epoch": 2.5033614647403803, |
| "grad_norm": 1.0853751747133211, |
| "learning_rate": 8.742692175528027e-06, |
| "loss": 1.6929, |
| "step": 8750 |
| }, |
| { |
| "epoch": 2.5176655700185955, |
| "grad_norm": 0.5845518675223795, |
| "learning_rate": 8.728514539081837e-06, |
| "loss": 1.6795, |
| "step": 8800 |
| }, |
| { |
| "epoch": 2.5176655700185955, |
| "eval_loss": 1.6996084451675415, |
| "eval_runtime": 13.1313, |
| "eval_samples_per_second": 76.154, |
| "eval_steps_per_second": 2.437, |
| "step": 8800 |
| }, |
| { |
| "epoch": 2.53196967529681, |
| "grad_norm": 0.6987736172220224, |
| "learning_rate": 8.714270553498567e-06, |
| "loss": 1.6786, |
| "step": 8850 |
| }, |
| { |
| "epoch": 2.5462737805750253, |
| "grad_norm": 0.6845251902273746, |
| "learning_rate": 8.699960511506077e-06, |
| "loss": 1.6915, |
| "step": 8900 |
| }, |
| { |
| "epoch": 2.56057788585324, |
| "grad_norm": 0.5813491957603439, |
| "learning_rate": 8.685584707189749e-06, |
| "loss": 1.6864, |
| "step": 8950 |
| }, |
| { |
| "epoch": 2.5748819911314547, |
| "grad_norm": 0.7758297556506881, |
| "learning_rate": 8.671143435986447e-06, |
| "loss": 1.6853, |
| "step": 9000 |
| }, |
| { |
| "epoch": 2.5891860964096693, |
| "grad_norm": 0.6480784306311093, |
| "learning_rate": 8.656636994678447e-06, |
| "loss": 1.6945, |
| "step": 9050 |
| }, |
| { |
| "epoch": 2.6034902016878845, |
| "grad_norm": 0.752842227645809, |
| "learning_rate": 8.642065681387329e-06, |
| "loss": 1.6856, |
| "step": 9100 |
| }, |
| { |
| "epoch": 2.617794306966099, |
| "grad_norm": 0.8563670508754716, |
| "learning_rate": 8.627429795567858e-06, |
| "loss": 1.6813, |
| "step": 9150 |
| }, |
| { |
| "epoch": 2.6320984122443143, |
| "grad_norm": 0.649858543338678, |
| "learning_rate": 8.61272963800183e-06, |
| "loss": 1.6807, |
| "step": 9200 |
| }, |
| { |
| "epoch": 2.6320984122443143, |
| "eval_loss": 1.6953411102294922, |
| "eval_runtime": 13.1331, |
| "eval_samples_per_second": 76.143, |
| "eval_steps_per_second": 2.437, |
| "step": 9200 |
| }, |
| { |
| "epoch": 2.646402517522529, |
| "grad_norm": 0.6485355884487186, |
| "learning_rate": 8.597965510791883e-06, |
| "loss": 1.691, |
| "step": 9250 |
| }, |
| { |
| "epoch": 2.6607066228007437, |
| "grad_norm": 0.8331706908409144, |
| "learning_rate": 8.5831377173553e-06, |
| "loss": 1.6757, |
| "step": 9300 |
| }, |
| { |
| "epoch": 2.675010728078959, |
| "grad_norm": 0.7237339174259747, |
| "learning_rate": 8.568246562417762e-06, |
| "loss": 1.6838, |
| "step": 9350 |
| }, |
| { |
| "epoch": 2.6893148333571735, |
| "grad_norm": 0.6551466282285732, |
| "learning_rate": 8.553292352007096e-06, |
| "loss": 1.6815, |
| "step": 9400 |
| }, |
| { |
| "epoch": 2.7036189386353886, |
| "grad_norm": 0.6601355008192104, |
| "learning_rate": 8.538275393446976e-06, |
| "loss": 1.6857, |
| "step": 9450 |
| }, |
| { |
| "epoch": 2.7179230439136033, |
| "grad_norm": 0.6614316555641063, |
| "learning_rate": 8.523195995350613e-06, |
| "loss": 1.6823, |
| "step": 9500 |
| }, |
| { |
| "epoch": 2.732227149191818, |
| "grad_norm": 0.9043872683705563, |
| "learning_rate": 8.508054467614417e-06, |
| "loss": 1.6807, |
| "step": 9550 |
| }, |
| { |
| "epoch": 2.7465312544700327, |
| "grad_norm": 0.6024321505835114, |
| "learning_rate": 8.492851121411614e-06, |
| "loss": 1.6775, |
| "step": 9600 |
| }, |
| { |
| "epoch": 2.7465312544700327, |
| "eval_loss": 1.6916736364364624, |
| "eval_runtime": 13.1667, |
| "eval_samples_per_second": 75.949, |
| "eval_steps_per_second": 2.43, |
| "step": 9600 |
| }, |
| { |
| "epoch": 2.760835359748248, |
| "grad_norm": 0.5885102755974863, |
| "learning_rate": 8.477586269185868e-06, |
| "loss": 1.6783, |
| "step": 9650 |
| }, |
| { |
| "epoch": 2.7751394650264625, |
| "grad_norm": 0.5923048948697456, |
| "learning_rate": 8.462260224644848e-06, |
| "loss": 1.6754, |
| "step": 9700 |
| }, |
| { |
| "epoch": 2.7894435703046776, |
| "grad_norm": 0.658689052183787, |
| "learning_rate": 8.446873302753783e-06, |
| "loss": 1.6879, |
| "step": 9750 |
| }, |
| { |
| "epoch": 2.8037476755828923, |
| "grad_norm": 0.6839558742874721, |
| "learning_rate": 8.431425819728998e-06, |
| "loss": 1.6833, |
| "step": 9800 |
| }, |
| { |
| "epoch": 2.818051780861107, |
| "grad_norm": 0.563076108100303, |
| "learning_rate": 8.415918093031403e-06, |
| "loss": 1.6746, |
| "step": 9850 |
| }, |
| { |
| "epoch": 2.832355886139322, |
| "grad_norm": 0.6412513978425523, |
| "learning_rate": 8.400350441359976e-06, |
| "loss": 1.6673, |
| "step": 9900 |
| }, |
| { |
| "epoch": 2.846659991417537, |
| "grad_norm": 0.6509884557319101, |
| "learning_rate": 8.384723184645211e-06, |
| "loss": 1.6736, |
| "step": 9950 |
| }, |
| { |
| "epoch": 2.860964096695752, |
| "grad_norm": 0.6795285605676114, |
| "learning_rate": 8.369036644042546e-06, |
| "loss": 1.6602, |
| "step": 10000 |
| }, |
| { |
| "epoch": 2.860964096695752, |
| "eval_loss": 1.6880682706832886, |
| "eval_runtime": 13.1264, |
| "eval_samples_per_second": 76.182, |
| "eval_steps_per_second": 2.438, |
| "step": 10000 |
| }, |
| { |
| "epoch": 2.8752682019739666, |
| "grad_norm": 0.6331270907140948, |
| "learning_rate": 8.353291141925763e-06, |
| "loss": 1.6678, |
| "step": 10050 |
| }, |
| { |
| "epoch": 2.8895723072521813, |
| "grad_norm": 0.6165703754864105, |
| "learning_rate": 8.337487001880353e-06, |
| "loss": 1.6779, |
| "step": 10100 |
| }, |
| { |
| "epoch": 2.903876412530396, |
| "grad_norm": 0.9180760124515276, |
| "learning_rate": 8.32162454869688e-06, |
| "loss": 1.6709, |
| "step": 10150 |
| }, |
| { |
| "epoch": 2.918180517808611, |
| "grad_norm": 0.6067014553628142, |
| "learning_rate": 8.305704108364301e-06, |
| "loss": 1.6748, |
| "step": 10200 |
| }, |
| { |
| "epoch": 2.932484623086826, |
| "grad_norm": 0.6822167283880958, |
| "learning_rate": 8.289726008063265e-06, |
| "loss": 1.6734, |
| "step": 10250 |
| }, |
| { |
| "epoch": 2.946788728365041, |
| "grad_norm": 0.5414334397184379, |
| "learning_rate": 8.273690576159383e-06, |
| "loss": 1.674, |
| "step": 10300 |
| }, |
| { |
| "epoch": 2.9610928336432556, |
| "grad_norm": 0.6960756922615432, |
| "learning_rate": 8.257598142196496e-06, |
| "loss": 1.6611, |
| "step": 10350 |
| }, |
| { |
| "epoch": 2.9753969389214703, |
| "grad_norm": 0.6604617349648914, |
| "learning_rate": 8.241449036889892e-06, |
| "loss": 1.6676, |
| "step": 10400 |
| }, |
| { |
| "epoch": 2.9753969389214703, |
| "eval_loss": 1.6850733757019043, |
| "eval_runtime": 13.109, |
| "eval_samples_per_second": 76.284, |
| "eval_steps_per_second": 2.441, |
| "step": 10400 |
| }, |
| { |
| "epoch": 2.9897010441996854, |
| "grad_norm": 0.803147857087257, |
| "learning_rate": 8.225243592119501e-06, |
| "loss": 1.6742, |
| "step": 10450 |
| }, |
| { |
| "epoch": 3.0040051494779, |
| "grad_norm": 0.7661508794366779, |
| "learning_rate": 8.208982140923095e-06, |
| "loss": 1.6643, |
| "step": 10500 |
| }, |
| { |
| "epoch": 3.0183092547561152, |
| "grad_norm": 0.9234438361457962, |
| "learning_rate": 8.192665017489431e-06, |
| "loss": 1.6358, |
| "step": 10550 |
| }, |
| { |
| "epoch": 3.03261336003433, |
| "grad_norm": 0.6961614825939386, |
| "learning_rate": 8.17629255715138e-06, |
| "loss": 1.6545, |
| "step": 10600 |
| }, |
| { |
| "epoch": 3.0469174653125446, |
| "grad_norm": 0.6895369107938377, |
| "learning_rate": 8.159865096379046e-06, |
| "loss": 1.6321, |
| "step": 10650 |
| }, |
| { |
| "epoch": 3.0612215705907597, |
| "grad_norm": 0.7078335937537477, |
| "learning_rate": 8.14338297277284e-06, |
| "loss": 1.6349, |
| "step": 10700 |
| }, |
| { |
| "epoch": 3.0755256758689744, |
| "grad_norm": 0.6076483839068002, |
| "learning_rate": 8.126846525056555e-06, |
| "loss": 1.6365, |
| "step": 10750 |
| }, |
| { |
| "epoch": 3.089829781147189, |
| "grad_norm": 0.6898999480835771, |
| "learning_rate": 8.110256093070393e-06, |
| "loss": 1.6546, |
| "step": 10800 |
| }, |
| { |
| "epoch": 3.089829781147189, |
| "eval_loss": 1.6832627058029175, |
| "eval_runtime": 13.1347, |
| "eval_samples_per_second": 76.134, |
| "eval_steps_per_second": 2.436, |
| "step": 10800 |
| }, |
| { |
| "epoch": 3.104133886425404, |
| "grad_norm": 0.7013516335683451, |
| "learning_rate": 8.093612017763986e-06, |
| "loss": 1.639, |
| "step": 10850 |
| }, |
| { |
| "epoch": 3.118437991703619, |
| "grad_norm": 0.6585546257998172, |
| "learning_rate": 8.076914641189388e-06, |
| "loss": 1.649, |
| "step": 10900 |
| }, |
| { |
| "epoch": 3.1327420969818336, |
| "grad_norm": 0.6634919864983012, |
| "learning_rate": 8.060164306494052e-06, |
| "loss": 1.6349, |
| "step": 10950 |
| }, |
| { |
| "epoch": 3.1470462022600487, |
| "grad_norm": 0.6535125345700776, |
| "learning_rate": 8.043361357913763e-06, |
| "loss": 1.6354, |
| "step": 11000 |
| }, |
| { |
| "epoch": 3.1613503075382634, |
| "grad_norm": 0.6365757392209358, |
| "learning_rate": 8.026506140765581e-06, |
| "loss": 1.6411, |
| "step": 11050 |
| }, |
| { |
| "epoch": 3.1756544128164785, |
| "grad_norm": 0.7613496780923112, |
| "learning_rate": 8.009599001440733e-06, |
| "loss": 1.6443, |
| "step": 11100 |
| }, |
| { |
| "epoch": 3.189958518094693, |
| "grad_norm": 0.6788371997973981, |
| "learning_rate": 7.992640287397498e-06, |
| "loss": 1.6406, |
| "step": 11150 |
| }, |
| { |
| "epoch": 3.204262623372908, |
| "grad_norm": 0.7267161462233287, |
| "learning_rate": 7.975630347154062e-06, |
| "loss": 1.6416, |
| "step": 11200 |
| }, |
| { |
| "epoch": 3.204262623372908, |
| "eval_loss": 1.6800661087036133, |
| "eval_runtime": 13.1106, |
| "eval_samples_per_second": 76.274, |
| "eval_steps_per_second": 2.441, |
| "step": 11200 |
| }, |
| { |
| "epoch": 3.218566728651123, |
| "grad_norm": 0.6610572852986926, |
| "learning_rate": 7.958569530281369e-06, |
| "loss": 1.637, |
| "step": 11250 |
| }, |
| { |
| "epoch": 3.2328708339293377, |
| "grad_norm": 0.9622885989079347, |
| "learning_rate": 7.941458187395918e-06, |
| "loss": 1.6356, |
| "step": 11300 |
| }, |
| { |
| "epoch": 3.2471749392075524, |
| "grad_norm": 0.6376730332996383, |
| "learning_rate": 7.924296670152573e-06, |
| "loss": 1.6353, |
| "step": 11350 |
| }, |
| { |
| "epoch": 3.2614790444857675, |
| "grad_norm": 0.7223900847461898, |
| "learning_rate": 7.907085331237328e-06, |
| "loss": 1.6484, |
| "step": 11400 |
| }, |
| { |
| "epoch": 3.275783149763982, |
| "grad_norm": 0.5989178515182939, |
| "learning_rate": 7.889824524360058e-06, |
| "loss": 1.6451, |
| "step": 11450 |
| }, |
| { |
| "epoch": 3.290087255042197, |
| "grad_norm": 0.6779901779346886, |
| "learning_rate": 7.872514604247261e-06, |
| "loss": 1.6453, |
| "step": 11500 |
| }, |
| { |
| "epoch": 3.304391360320412, |
| "grad_norm": 0.7005880085935425, |
| "learning_rate": 7.855155926634755e-06, |
| "loss": 1.6472, |
| "step": 11550 |
| }, |
| { |
| "epoch": 3.3186954655986267, |
| "grad_norm": 0.7609081934908348, |
| "learning_rate": 7.837748848260372e-06, |
| "loss": 1.6465, |
| "step": 11600 |
| }, |
| { |
| "epoch": 3.3186954655986267, |
| "eval_loss": 1.6766809225082397, |
| "eval_runtime": 13.1249, |
| "eval_samples_per_second": 76.191, |
| "eval_steps_per_second": 2.438, |
| "step": 11600 |
| }, |
| { |
| "epoch": 3.332999570876842, |
| "grad_norm": 0.6431999289309001, |
| "learning_rate": 7.820293726856625e-06, |
| "loss": 1.6363, |
| "step": 11650 |
| }, |
| { |
| "epoch": 3.3473036761550565, |
| "grad_norm": 0.7971053094613834, |
| "learning_rate": 7.802790921143367e-06, |
| "loss": 1.642, |
| "step": 11700 |
| }, |
| { |
| "epoch": 3.361607781433271, |
| "grad_norm": 0.8315194233152097, |
| "learning_rate": 7.785240790820403e-06, |
| "loss": 1.6356, |
| "step": 11750 |
| }, |
| { |
| "epoch": 3.3759118867114863, |
| "grad_norm": 0.7233317425806471, |
| "learning_rate": 7.767643696560103e-06, |
| "loss": 1.625, |
| "step": 11800 |
| }, |
| { |
| "epoch": 3.390215991989701, |
| "grad_norm": 0.7043201889466395, |
| "learning_rate": 7.75e-06, |
| "loss": 1.6454, |
| "step": 11850 |
| }, |
| { |
| "epoch": 3.4045200972679157, |
| "grad_norm": 0.8514114238482182, |
| "learning_rate": 7.732310063735346e-06, |
| "loss": 1.6283, |
| "step": 11900 |
| }, |
| { |
| "epoch": 3.418824202546131, |
| "grad_norm": 0.8115066197624821, |
| "learning_rate": 7.71457425131166e-06, |
| "loss": 1.6235, |
| "step": 11950 |
| }, |
| { |
| "epoch": 3.4331283078243455, |
| "grad_norm": 0.6131161672110997, |
| "learning_rate": 7.696792927217266e-06, |
| "loss": 1.6332, |
| "step": 12000 |
| }, |
| { |
| "epoch": 3.4331283078243455, |
| "eval_loss": 1.6723066568374634, |
| "eval_runtime": 13.1167, |
| "eval_samples_per_second": 76.238, |
| "eval_steps_per_second": 2.44, |
| "step": 12000 |
| }, |
| { |
| "epoch": 3.44743241310256, |
| "grad_norm": 0.5768237911834109, |
| "learning_rate": 7.6789664568758e-06, |
| "loss": 1.6232, |
| "step": 12050 |
| }, |
| { |
| "epoch": 3.4617365183807753, |
| "grad_norm": 0.6357616712470612, |
| "learning_rate": 7.661095206638688e-06, |
| "loss": 1.6384, |
| "step": 12100 |
| }, |
| { |
| "epoch": 3.47604062365899, |
| "grad_norm": 0.6064381660716113, |
| "learning_rate": 7.643179543777636e-06, |
| "loss": 1.6423, |
| "step": 12150 |
| }, |
| { |
| "epoch": 3.490344728937205, |
| "grad_norm": 0.6129739275970794, |
| "learning_rate": 7.625219836477073e-06, |
| "loss": 1.6376, |
| "step": 12200 |
| }, |
| { |
| "epoch": 3.50464883421542, |
| "grad_norm": 0.6413234223207633, |
| "learning_rate": 7.607216453826575e-06, |
| "loss": 1.6342, |
| "step": 12250 |
| }, |
| { |
| "epoch": 3.5189529394936345, |
| "grad_norm": 0.649929569744414, |
| "learning_rate": 7.589169765813298e-06, |
| "loss": 1.6418, |
| "step": 12300 |
| }, |
| { |
| "epoch": 3.5332570447718497, |
| "grad_norm": 0.5694768483225782, |
| "learning_rate": 7.571080143314362e-06, |
| "loss": 1.6242, |
| "step": 12350 |
| }, |
| { |
| "epoch": 3.5475611500500643, |
| "grad_norm": 0.7350828307091747, |
| "learning_rate": 7.552947958089234e-06, |
| "loss": 1.6217, |
| "step": 12400 |
| }, |
| { |
| "epoch": 3.5475611500500643, |
| "eval_loss": 1.6681584119796753, |
| "eval_runtime": 13.1321, |
| "eval_samples_per_second": 76.149, |
| "eval_steps_per_second": 2.437, |
| "step": 12400 |
| }, |
| { |
| "epoch": 3.5618652553282795, |
| "grad_norm": 0.8136490657582578, |
| "learning_rate": 7.534773582772087e-06, |
| "loss": 1.6303, |
| "step": 12450 |
| }, |
| { |
| "epoch": 3.576169360606494, |
| "grad_norm": 0.7180717788221607, |
| "learning_rate": 7.51655739086414e-06, |
| "loss": 1.6361, |
| "step": 12500 |
| }, |
| { |
| "epoch": 3.590473465884709, |
| "grad_norm": 0.7947804203928884, |
| "learning_rate": 7.498299756725984e-06, |
| "loss": 1.6387, |
| "step": 12550 |
| }, |
| { |
| "epoch": 3.6047775711629235, |
| "grad_norm": 0.560362351616168, |
| "learning_rate": 7.480001055569892e-06, |
| "loss": 1.6309, |
| "step": 12600 |
| }, |
| { |
| "epoch": 3.6190816764411387, |
| "grad_norm": 0.6426607371565076, |
| "learning_rate": 7.4616616634521e-06, |
| "loss": 1.633, |
| "step": 12650 |
| }, |
| { |
| "epoch": 3.6333857817193533, |
| "grad_norm": 0.6171738461490535, |
| "learning_rate": 7.443281957265086e-06, |
| "loss": 1.6281, |
| "step": 12700 |
| }, |
| { |
| "epoch": 3.6476898869975685, |
| "grad_norm": 0.6363040797185718, |
| "learning_rate": 7.424862314729819e-06, |
| "loss": 1.6311, |
| "step": 12750 |
| }, |
| { |
| "epoch": 3.661993992275783, |
| "grad_norm": 0.7596986784676188, |
| "learning_rate": 7.406403114388003e-06, |
| "loss": 1.6292, |
| "step": 12800 |
| }, |
| { |
| "epoch": 3.661993992275783, |
| "eval_loss": 1.6635466814041138, |
| "eval_runtime": 13.15, |
| "eval_samples_per_second": 76.046, |
| "eval_steps_per_second": 2.433, |
| "step": 12800 |
| }, |
| { |
| "epoch": 3.676298097553998, |
| "grad_norm": 0.7293094685670296, |
| "learning_rate": 7.387904735594291e-06, |
| "loss": 1.6287, |
| "step": 12850 |
| }, |
| { |
| "epoch": 3.690602202832213, |
| "grad_norm": 0.6158921819942098, |
| "learning_rate": 7.36936755850849e-06, |
| "loss": 1.6385, |
| "step": 12900 |
| }, |
| { |
| "epoch": 3.7049063081104276, |
| "grad_norm": 0.5784944167280377, |
| "learning_rate": 7.3507919640877535e-06, |
| "loss": 1.6328, |
| "step": 12950 |
| }, |
| { |
| "epoch": 3.7192104133886428, |
| "grad_norm": 0.6275769116215365, |
| "learning_rate": 7.332178334078746e-06, |
| "loss": 1.6346, |
| "step": 13000 |
| }, |
| { |
| "epoch": 3.7335145186668575, |
| "grad_norm": 0.7049335766852525, |
| "learning_rate": 7.313527051009803e-06, |
| "loss": 1.6314, |
| "step": 13050 |
| }, |
| { |
| "epoch": 3.747818623945072, |
| "grad_norm": 0.5885503989540276, |
| "learning_rate": 7.2948384981830655e-06, |
| "loss": 1.6276, |
| "step": 13100 |
| }, |
| { |
| "epoch": 3.762122729223287, |
| "grad_norm": 0.6144133824628834, |
| "learning_rate": 7.2761130596666045e-06, |
| "loss": 1.6194, |
| "step": 13150 |
| }, |
| { |
| "epoch": 3.776426834501502, |
| "grad_norm": 0.6880469614329583, |
| "learning_rate": 7.25735112028653e-06, |
| "loss": 1.6263, |
| "step": 13200 |
| }, |
| { |
| "epoch": 3.776426834501502, |
| "eval_loss": 1.6603440046310425, |
| "eval_runtime": 13.1562, |
| "eval_samples_per_second": 76.01, |
| "eval_steps_per_second": 2.432, |
| "step": 13200 |
| }, |
| { |
| "epoch": 3.7907309397797166, |
| "grad_norm": 0.6695854088226235, |
| "learning_rate": 7.2385530656190785e-06, |
| "loss": 1.6163, |
| "step": 13250 |
| }, |
| { |
| "epoch": 3.8050350450579318, |
| "grad_norm": 0.7403722736764125, |
| "learning_rate": 7.219719281982694e-06, |
| "loss": 1.6223, |
| "step": 13300 |
| }, |
| { |
| "epoch": 3.8193391503361465, |
| "grad_norm": 0.6035992360261074, |
| "learning_rate": 7.20085015643008e-06, |
| "loss": 1.6177, |
| "step": 13350 |
| }, |
| { |
| "epoch": 3.833643255614361, |
| "grad_norm": 0.5706930216009009, |
| "learning_rate": 7.181946076740257e-06, |
| "loss": 1.6177, |
| "step": 13400 |
| }, |
| { |
| "epoch": 3.8479473608925763, |
| "grad_norm": 0.6806370138912894, |
| "learning_rate": 7.163007431410583e-06, |
| "loss": 1.6262, |
| "step": 13450 |
| }, |
| { |
| "epoch": 3.862251466170791, |
| "grad_norm": 0.6234185420830436, |
| "learning_rate": 7.144034609648779e-06, |
| "loss": 1.6331, |
| "step": 13500 |
| }, |
| { |
| "epoch": 3.876555571449006, |
| "grad_norm": 0.7677470872346758, |
| "learning_rate": 7.125028001364918e-06, |
| "loss": 1.6197, |
| "step": 13550 |
| }, |
| { |
| "epoch": 3.8908596767272208, |
| "grad_norm": 0.6034269149098646, |
| "learning_rate": 7.105987997163424e-06, |
| "loss": 1.6266, |
| "step": 13600 |
| }, |
| { |
| "epoch": 3.8908596767272208, |
| "eval_loss": 1.6566662788391113, |
| "eval_runtime": 13.1267, |
| "eval_samples_per_second": 76.181, |
| "eval_steps_per_second": 2.438, |
| "step": 13600 |
| }, |
| { |
| "epoch": 3.9051637820054355, |
| "grad_norm": 0.8302705314191118, |
| "learning_rate": 7.086914988335039e-06, |
| "loss": 1.6269, |
| "step": 13650 |
| }, |
| { |
| "epoch": 3.91946788728365, |
| "grad_norm": 0.6389864823326697, |
| "learning_rate": 7.0678093668487836e-06, |
| "loss": 1.6323, |
| "step": 13700 |
| }, |
| { |
| "epoch": 3.9337719925618653, |
| "grad_norm": 0.7866068542682738, |
| "learning_rate": 7.048671525343898e-06, |
| "loss": 1.6106, |
| "step": 13750 |
| }, |
| { |
| "epoch": 3.94807609784008, |
| "grad_norm": 0.7727876700796928, |
| "learning_rate": 7.029501857121776e-06, |
| "loss": 1.6173, |
| "step": 13800 |
| }, |
| { |
| "epoch": 3.962380203118295, |
| "grad_norm": 0.6170837296024364, |
| "learning_rate": 7.010300756137882e-06, |
| "loss": 1.6109, |
| "step": 13850 |
| }, |
| { |
| "epoch": 3.9766843083965098, |
| "grad_norm": 0.6847504318409611, |
| "learning_rate": 6.991068616993655e-06, |
| "loss": 1.6208, |
| "step": 13900 |
| }, |
| { |
| "epoch": 3.9909884136747245, |
| "grad_norm": 0.7026084916595097, |
| "learning_rate": 6.971805834928399e-06, |
| "loss": 1.6147, |
| "step": 13950 |
| }, |
| { |
| "epoch": 4.005292518952939, |
| "grad_norm": 0.5565822000879449, |
| "learning_rate": 6.952512805811156e-06, |
| "loss": 1.6238, |
| "step": 14000 |
| }, |
| { |
| "epoch": 4.005292518952939, |
| "eval_loss": 1.6554406881332397, |
| "eval_runtime": 13.1672, |
| "eval_samples_per_second": 75.947, |
| "eval_steps_per_second": 2.43, |
| "step": 14000 |
| }, |
| { |
| "epoch": 4.019596624231155, |
| "grad_norm": 0.7906915205693054, |
| "learning_rate": 6.933189926132581e-06, |
| "loss": 1.5887, |
| "step": 14050 |
| }, |
| { |
| "epoch": 4.033900729509369, |
| "grad_norm": 0.6014668104450175, |
| "learning_rate": 6.913837592996783e-06, |
| "loss": 1.5824, |
| "step": 14100 |
| }, |
| { |
| "epoch": 4.048204834787584, |
| "grad_norm": 0.6349783814707618, |
| "learning_rate": 6.894456204113167e-06, |
| "loss": 1.5993, |
| "step": 14150 |
| }, |
| { |
| "epoch": 4.062508940065799, |
| "grad_norm": 0.7004788731024523, |
| "learning_rate": 6.875046157788267e-06, |
| "loss": 1.5856, |
| "step": 14200 |
| }, |
| { |
| "epoch": 4.0768130453440135, |
| "grad_norm": 0.5890601619700472, |
| "learning_rate": 6.855607852917555e-06, |
| "loss": 1.5992, |
| "step": 14250 |
| }, |
| { |
| "epoch": 4.091117150622228, |
| "grad_norm": 0.7378157690547319, |
| "learning_rate": 6.836141688977238e-06, |
| "loss": 1.5933, |
| "step": 14300 |
| }, |
| { |
| "epoch": 4.105421255900444, |
| "grad_norm": 0.5985785926592089, |
| "learning_rate": 6.816648066016059e-06, |
| "loss": 1.6059, |
| "step": 14350 |
| }, |
| { |
| "epoch": 4.119725361178658, |
| "grad_norm": 0.7254510710383179, |
| "learning_rate": 6.7971273846470696e-06, |
| "loss": 1.5755, |
| "step": 14400 |
| }, |
| { |
| "epoch": 4.119725361178658, |
| "eval_loss": 1.652251124382019, |
| "eval_runtime": 13.1229, |
| "eval_samples_per_second": 76.203, |
| "eval_steps_per_second": 2.438, |
| "step": 14400 |
| }, |
| { |
| "epoch": 4.134029466456873, |
| "grad_norm": 0.701638747657545, |
| "learning_rate": 6.777580046039399e-06, |
| "loss": 1.5878, |
| "step": 14450 |
| }, |
| { |
| "epoch": 4.148333571735088, |
| "grad_norm": 0.6156043257629281, |
| "learning_rate": 6.758006451910008e-06, |
| "loss": 1.5888, |
| "step": 14500 |
| }, |
| { |
| "epoch": 4.1626376770133025, |
| "grad_norm": 0.6311348402263556, |
| "learning_rate": 6.738407004515434e-06, |
| "loss": 1.5906, |
| "step": 14550 |
| }, |
| { |
| "epoch": 4.176941782291518, |
| "grad_norm": 0.7457367222382035, |
| "learning_rate": 6.718782106643524e-06, |
| "loss": 1.5935, |
| "step": 14600 |
| }, |
| { |
| "epoch": 4.191245887569733, |
| "grad_norm": 0.6497146588667823, |
| "learning_rate": 6.699132161605158e-06, |
| "loss": 1.5944, |
| "step": 14650 |
| }, |
| { |
| "epoch": 4.205549992847947, |
| "grad_norm": 0.6021866479049087, |
| "learning_rate": 6.679457573225961e-06, |
| "loss": 1.5872, |
| "step": 14700 |
| }, |
| { |
| "epoch": 4.219854098126162, |
| "grad_norm": 0.606706719077781, |
| "learning_rate": 6.659758745837998e-06, |
| "loss": 1.5892, |
| "step": 14750 |
| }, |
| { |
| "epoch": 4.234158203404377, |
| "grad_norm": 0.796088041146077, |
| "learning_rate": 6.640036084271477e-06, |
| "loss": 1.5778, |
| "step": 14800 |
| }, |
| { |
| "epoch": 4.234158203404377, |
| "eval_loss": 1.652685523033142, |
| "eval_runtime": 13.11, |
| "eval_samples_per_second": 76.278, |
| "eval_steps_per_second": 2.441, |
| "step": 14800 |
| }, |
| { |
| "epoch": 4.248462308682592, |
| "grad_norm": 0.5720799505387223, |
| "learning_rate": 6.620289993846416e-06, |
| "loss": 1.5853, |
| "step": 14850 |
| }, |
| { |
| "epoch": 4.262766413960807, |
| "grad_norm": 0.638641651597387, |
| "learning_rate": 6.600520880364318e-06, |
| "loss": 1.5821, |
| "step": 14900 |
| }, |
| { |
| "epoch": 4.277070519239022, |
| "grad_norm": 0.6826954900918857, |
| "learning_rate": 6.5807291500998385e-06, |
| "loss": 1.5926, |
| "step": 14950 |
| }, |
| { |
| "epoch": 4.291374624517236, |
| "grad_norm": 0.7142773516133535, |
| "learning_rate": 6.560915209792424e-06, |
| "loss": 1.5756, |
| "step": 15000 |
| }, |
| { |
| "epoch": 4.305678729795451, |
| "grad_norm": 0.6893665096075864, |
| "learning_rate": 6.541079466637962e-06, |
| "loss": 1.5898, |
| "step": 15050 |
| }, |
| { |
| "epoch": 4.319982835073666, |
| "grad_norm": 0.6934025133503401, |
| "learning_rate": 6.52122232828041e-06, |
| "loss": 1.5906, |
| "step": 15100 |
| }, |
| { |
| "epoch": 4.334286940351881, |
| "grad_norm": 0.5946314963781922, |
| "learning_rate": 6.501344202803415e-06, |
| "loss": 1.5876, |
| "step": 15150 |
| }, |
| { |
| "epoch": 4.348591045630096, |
| "grad_norm": 0.6765765096436505, |
| "learning_rate": 6.4814454987219355e-06, |
| "loss": 1.5876, |
| "step": 15200 |
| }, |
| { |
| "epoch": 4.348591045630096, |
| "eval_loss": 1.6479697227478027, |
| "eval_runtime": 13.1313, |
| "eval_samples_per_second": 76.154, |
| "eval_steps_per_second": 2.437, |
| "step": 15200 |
| }, |
| { |
| "epoch": 4.362895150908311, |
| "grad_norm": 0.6577195429173401, |
| "learning_rate": 6.461526624973836e-06, |
| "loss": 1.5899, |
| "step": 15250 |
| }, |
| { |
| "epoch": 4.377199256186525, |
| "grad_norm": 0.8277675923235659, |
| "learning_rate": 6.441587990911489e-06, |
| "loss": 1.5887, |
| "step": 15300 |
| }, |
| { |
| "epoch": 4.39150336146474, |
| "grad_norm": 0.7292088895911856, |
| "learning_rate": 6.421630006293359e-06, |
| "loss": 1.5728, |
| "step": 15350 |
| }, |
| { |
| "epoch": 4.405807466742955, |
| "grad_norm": 0.7886180995867164, |
| "learning_rate": 6.401653081275586e-06, |
| "loss": 1.5765, |
| "step": 15400 |
| }, |
| { |
| "epoch": 4.42011157202117, |
| "grad_norm": 0.6450755047638138, |
| "learning_rate": 6.38165762640355e-06, |
| "loss": 1.5831, |
| "step": 15450 |
| }, |
| { |
| "epoch": 4.434415677299385, |
| "grad_norm": 0.6698809351304791, |
| "learning_rate": 6.361644052603445e-06, |
| "loss": 1.5795, |
| "step": 15500 |
| }, |
| { |
| "epoch": 4.4487197825776, |
| "grad_norm": 0.7804327125701724, |
| "learning_rate": 6.341612771173817e-06, |
| "loss": 1.5896, |
| "step": 15550 |
| }, |
| { |
| "epoch": 4.463023887855814, |
| "grad_norm": 0.5570965590960153, |
| "learning_rate": 6.321564193777129e-06, |
| "loss": 1.5778, |
| "step": 15600 |
| }, |
| { |
| "epoch": 4.463023887855814, |
| "eval_loss": 1.6438684463500977, |
| "eval_runtime": 13.1716, |
| "eval_samples_per_second": 75.921, |
| "eval_steps_per_second": 2.429, |
| "step": 15600 |
| }, |
| { |
| "epoch": 4.477327993134029, |
| "grad_norm": 0.81125064044786, |
| "learning_rate": 6.301498732431287e-06, |
| "loss": 1.5898, |
| "step": 15650 |
| }, |
| { |
| "epoch": 4.491632098412245, |
| "grad_norm": 0.7306624425115567, |
| "learning_rate": 6.281416799501188e-06, |
| "loss": 1.5775, |
| "step": 15700 |
| }, |
| { |
| "epoch": 4.505936203690459, |
| "grad_norm": 0.7181992700870559, |
| "learning_rate": 6.261318807690223e-06, |
| "loss": 1.5844, |
| "step": 15750 |
| }, |
| { |
| "epoch": 4.520240308968674, |
| "grad_norm": 0.7748503653032427, |
| "learning_rate": 6.24120517003182e-06, |
| "loss": 1.5807, |
| "step": 15800 |
| }, |
| { |
| "epoch": 4.534544414246889, |
| "grad_norm": 0.692890389924739, |
| "learning_rate": 6.221076299880939e-06, |
| "loss": 1.5779, |
| "step": 15850 |
| }, |
| { |
| "epoch": 4.548848519525103, |
| "grad_norm": 0.750802300795554, |
| "learning_rate": 6.200932610905584e-06, |
| "loss": 1.5914, |
| "step": 15900 |
| }, |
| { |
| "epoch": 4.563152624803319, |
| "grad_norm": 0.6119826848690993, |
| "learning_rate": 6.180774517078301e-06, |
| "loss": 1.5745, |
| "step": 15950 |
| }, |
| { |
| "epoch": 4.577456730081534, |
| "grad_norm": 0.6472801196876264, |
| "learning_rate": 6.160602432667668e-06, |
| "loss": 1.5737, |
| "step": 16000 |
| }, |
| { |
| "epoch": 4.577456730081534, |
| "eval_loss": 1.6402463912963867, |
| "eval_runtime": 13.0995, |
| "eval_samples_per_second": 76.339, |
| "eval_steps_per_second": 2.443, |
| "step": 16000 |
| }, |
| { |
| "epoch": 4.591760835359748, |
| "grad_norm": 0.6082312742368647, |
| "learning_rate": 6.140416772229785e-06, |
| "loss": 1.5843, |
| "step": 16050 |
| }, |
| { |
| "epoch": 4.606064940637963, |
| "grad_norm": 0.6109383078063736, |
| "learning_rate": 6.12021795059975e-06, |
| "loss": 1.5795, |
| "step": 16100 |
| }, |
| { |
| "epoch": 4.620369045916178, |
| "grad_norm": 0.5808974596081403, |
| "learning_rate": 6.10000638288314e-06, |
| "loss": 1.5834, |
| "step": 16150 |
| }, |
| { |
| "epoch": 4.634673151194393, |
| "grad_norm": 0.6471974483578133, |
| "learning_rate": 6.079782484447475e-06, |
| "loss": 1.5685, |
| "step": 16200 |
| }, |
| { |
| "epoch": 4.648977256472608, |
| "grad_norm": 0.6706099898660832, |
| "learning_rate": 6.059546670913684e-06, |
| "loss": 1.582, |
| "step": 16250 |
| }, |
| { |
| "epoch": 4.663281361750823, |
| "grad_norm": 0.634560651660765, |
| "learning_rate": 6.03929935814756e-06, |
| "loss": 1.572, |
| "step": 16300 |
| }, |
| { |
| "epoch": 4.677585467029037, |
| "grad_norm": 0.6380102332948996, |
| "learning_rate": 6.01904096225122e-06, |
| "loss": 1.5715, |
| "step": 16350 |
| }, |
| { |
| "epoch": 4.691889572307252, |
| "grad_norm": 0.6355595383367377, |
| "learning_rate": 5.998771899554551e-06, |
| "loss": 1.5724, |
| "step": 16400 |
| }, |
| { |
| "epoch": 4.691889572307252, |
| "eval_loss": 1.6363039016723633, |
| "eval_runtime": 13.141, |
| "eval_samples_per_second": 76.098, |
| "eval_steps_per_second": 2.435, |
| "step": 16400 |
| }, |
| { |
| "epoch": 4.706193677585467, |
| "grad_norm": 0.6132759482920135, |
| "learning_rate": 5.978492586606647e-06, |
| "loss": 1.5769, |
| "step": 16450 |
| }, |
| { |
| "epoch": 4.720497782863681, |
| "grad_norm": 0.6639573383492796, |
| "learning_rate": 5.958203440167261e-06, |
| "loss": 1.5671, |
| "step": 16500 |
| }, |
| { |
| "epoch": 4.734801888141897, |
| "grad_norm": 0.5417322612862875, |
| "learning_rate": 5.93790487719823e-06, |
| "loss": 1.5755, |
| "step": 16550 |
| }, |
| { |
| "epoch": 4.749105993420112, |
| "grad_norm": 0.6771963599708853, |
| "learning_rate": 5.917597314854914e-06, |
| "loss": 1.569, |
| "step": 16600 |
| }, |
| { |
| "epoch": 4.763410098698326, |
| "grad_norm": 0.7200514277275499, |
| "learning_rate": 5.897281170477614e-06, |
| "loss": 1.577, |
| "step": 16650 |
| }, |
| { |
| "epoch": 4.777714203976541, |
| "grad_norm": 0.5811950543492076, |
| "learning_rate": 5.876956861583007e-06, |
| "loss": 1.5622, |
| "step": 16700 |
| }, |
| { |
| "epoch": 4.792018309254756, |
| "grad_norm": 0.7104533054634696, |
| "learning_rate": 5.856624805855548e-06, |
| "loss": 1.579, |
| "step": 16750 |
| }, |
| { |
| "epoch": 4.806322414532971, |
| "grad_norm": 0.655887536338841, |
| "learning_rate": 5.83628542113891e-06, |
| "loss": 1.5857, |
| "step": 16800 |
| }, |
| { |
| "epoch": 4.806322414532971, |
| "eval_loss": 1.633513331413269, |
| "eval_runtime": 13.1532, |
| "eval_samples_per_second": 76.027, |
| "eval_steps_per_second": 2.433, |
| "step": 16800 |
| }, |
| { |
| "epoch": 4.820626519811186, |
| "grad_norm": 0.6455948477676112, |
| "learning_rate": 5.815939125427373e-06, |
| "loss": 1.5707, |
| "step": 16850 |
| }, |
| { |
| "epoch": 4.834930625089401, |
| "grad_norm": 0.6271361263073431, |
| "learning_rate": 5.795586336857253e-06, |
| "loss": 1.563, |
| "step": 16900 |
| }, |
| { |
| "epoch": 4.849234730367615, |
| "grad_norm": 0.7674869603862556, |
| "learning_rate": 5.775227473698294e-06, |
| "loss": 1.5779, |
| "step": 16950 |
| }, |
| { |
| "epoch": 4.86353883564583, |
| "grad_norm": 0.637190923985461, |
| "learning_rate": 5.754862954345081e-06, |
| "loss": 1.5713, |
| "step": 17000 |
| }, |
| { |
| "epoch": 4.877842940924046, |
| "grad_norm": 0.6651093545639066, |
| "learning_rate": 5.734493197308442e-06, |
| "loss": 1.5752, |
| "step": 17050 |
| }, |
| { |
| "epoch": 4.89214704620226, |
| "grad_norm": 0.7868806203083217, |
| "learning_rate": 5.714118621206843e-06, |
| "loss": 1.5692, |
| "step": 17100 |
| }, |
| { |
| "epoch": 4.906451151480475, |
| "grad_norm": 0.8067206283766785, |
| "learning_rate": 5.693739644757781e-06, |
| "loss": 1.5824, |
| "step": 17150 |
| }, |
| { |
| "epoch": 4.92075525675869, |
| "grad_norm": 0.6156907735488538, |
| "learning_rate": 5.673356686769194e-06, |
| "loss": 1.5582, |
| "step": 17200 |
| }, |
| { |
| "epoch": 4.92075525675869, |
| "eval_loss": 1.6318423748016357, |
| "eval_runtime": 13.1193, |
| "eval_samples_per_second": 76.224, |
| "eval_steps_per_second": 2.439, |
| "step": 17200 |
| }, |
| { |
| "epoch": 4.935059362036904, |
| "grad_norm": 0.7876895181186726, |
| "learning_rate": 5.6529701661308415e-06, |
| "loss": 1.5689, |
| "step": 17250 |
| }, |
| { |
| "epoch": 4.94936346731512, |
| "grad_norm": 0.6292685628972012, |
| "learning_rate": 5.632580501805692e-06, |
| "loss": 1.5739, |
| "step": 17300 |
| }, |
| { |
| "epoch": 4.963667572593335, |
| "grad_norm": 0.7149239853712192, |
| "learning_rate": 5.612188112821328e-06, |
| "loss": 1.5797, |
| "step": 17350 |
| }, |
| { |
| "epoch": 4.977971677871549, |
| "grad_norm": 0.834110019775529, |
| "learning_rate": 5.591793418261326e-06, |
| "loss": 1.5763, |
| "step": 17400 |
| }, |
| { |
| "epoch": 4.992275783149764, |
| "grad_norm": 0.7309326453697327, |
| "learning_rate": 5.571396837256637e-06, |
| "loss": 1.5726, |
| "step": 17450 |
| }, |
| { |
| "epoch": 5.006579888427979, |
| "grad_norm": 0.6697280370753163, |
| "learning_rate": 5.550998788976988e-06, |
| "loss": 1.5634, |
| "step": 17500 |
| }, |
| { |
| "epoch": 5.020883993706193, |
| "grad_norm": 0.6174166960990144, |
| "learning_rate": 5.530599692622257e-06, |
| "loss": 1.5467, |
| "step": 17550 |
| }, |
| { |
| "epoch": 5.035188098984409, |
| "grad_norm": 0.6241514857747369, |
| "learning_rate": 5.510199967413862e-06, |
| "loss": 1.5508, |
| "step": 17600 |
| }, |
| { |
| "epoch": 5.035188098984409, |
| "eval_loss": 1.6312223672866821, |
| "eval_runtime": 13.1239, |
| "eval_samples_per_second": 76.197, |
| "eval_steps_per_second": 2.438, |
| "step": 17600 |
| }, |
| { |
| "epoch": 5.049492204262624, |
| "grad_norm": 0.607902431110512, |
| "learning_rate": 5.489800032586141e-06, |
| "loss": 1.5467, |
| "step": 17650 |
| }, |
| { |
| "epoch": 5.063796309540838, |
| "grad_norm": 0.6361444457465594, |
| "learning_rate": 5.4694003073777446e-06, |
| "loss": 1.5323, |
| "step": 17700 |
| }, |
| { |
| "epoch": 5.078100414819053, |
| "grad_norm": 0.7502195043466207, |
| "learning_rate": 5.449001211023014e-06, |
| "loss": 1.5372, |
| "step": 17750 |
| }, |
| { |
| "epoch": 5.092404520097268, |
| "grad_norm": 0.6959223099372485, |
| "learning_rate": 5.428603162743365e-06, |
| "loss": 1.5507, |
| "step": 17800 |
| }, |
| { |
| "epoch": 5.106708625375482, |
| "grad_norm": 0.8282524586776218, |
| "learning_rate": 5.408206581738677e-06, |
| "loss": 1.5511, |
| "step": 17850 |
| }, |
| { |
| "epoch": 5.121012730653698, |
| "grad_norm": 0.6152478153542872, |
| "learning_rate": 5.387811887178673e-06, |
| "loss": 1.5454, |
| "step": 17900 |
| }, |
| { |
| "epoch": 5.135316835931913, |
| "grad_norm": 0.7178652735808533, |
| "learning_rate": 5.367419498194309e-06, |
| "loss": 1.5328, |
| "step": 17950 |
| }, |
| { |
| "epoch": 5.149620941210127, |
| "grad_norm": 0.6199866059077553, |
| "learning_rate": 5.347029833869161e-06, |
| "loss": 1.5323, |
| "step": 18000 |
| }, |
| { |
| "epoch": 5.149620941210127, |
| "eval_loss": 1.6294597387313843, |
| "eval_runtime": 13.0999, |
| "eval_samples_per_second": 76.336, |
| "eval_steps_per_second": 2.443, |
| "step": 18000 |
| }, |
| { |
| "epoch": 5.163925046488342, |
| "grad_norm": 0.5994719638332001, |
| "learning_rate": 5.326643313230806e-06, |
| "loss": 1.544, |
| "step": 18050 |
| }, |
| { |
| "epoch": 5.178229151766557, |
| "grad_norm": 0.6101948560685385, |
| "learning_rate": 5.306260355242221e-06, |
| "loss": 1.5393, |
| "step": 18100 |
| }, |
| { |
| "epoch": 5.192533257044772, |
| "grad_norm": 0.6358629359192572, |
| "learning_rate": 5.2858813787931605e-06, |
| "loss": 1.5418, |
| "step": 18150 |
| }, |
| { |
| "epoch": 5.206837362322987, |
| "grad_norm": 0.5992154701681206, |
| "learning_rate": 5.26550680269156e-06, |
| "loss": 1.5399, |
| "step": 18200 |
| }, |
| { |
| "epoch": 5.221141467601202, |
| "grad_norm": 0.7386890859479155, |
| "learning_rate": 5.24513704565492e-06, |
| "loss": 1.5541, |
| "step": 18250 |
| }, |
| { |
| "epoch": 5.235445572879416, |
| "grad_norm": 0.6887954344111628, |
| "learning_rate": 5.224772526301709e-06, |
| "loss": 1.5342, |
| "step": 18300 |
| }, |
| { |
| "epoch": 5.249749678157631, |
| "grad_norm": 0.6961019703931222, |
| "learning_rate": 5.20441366314275e-06, |
| "loss": 1.5463, |
| "step": 18350 |
| }, |
| { |
| "epoch": 5.2640537834358465, |
| "grad_norm": 0.7142803422247822, |
| "learning_rate": 5.184060874572628e-06, |
| "loss": 1.5415, |
| "step": 18400 |
| }, |
| { |
| "epoch": 5.2640537834358465, |
| "eval_loss": 1.6250876188278198, |
| "eval_runtime": 13.1209, |
| "eval_samples_per_second": 76.214, |
| "eval_steps_per_second": 2.439, |
| "step": 18400 |
| }, |
| { |
| "epoch": 5.278357888714061, |
| "grad_norm": 0.7209586773593375, |
| "learning_rate": 5.163714578861091e-06, |
| "loss": 1.5432, |
| "step": 18450 |
| }, |
| { |
| "epoch": 5.292661993992276, |
| "grad_norm": 0.8558312401002982, |
| "learning_rate": 5.143375194144452e-06, |
| "loss": 1.5396, |
| "step": 18500 |
| }, |
| { |
| "epoch": 5.306966099270491, |
| "grad_norm": 0.6140496298664848, |
| "learning_rate": 5.123043138416996e-06, |
| "loss": 1.5347, |
| "step": 18550 |
| }, |
| { |
| "epoch": 5.321270204548705, |
| "grad_norm": 0.584034537933172, |
| "learning_rate": 5.102718829522387e-06, |
| "loss": 1.5455, |
| "step": 18600 |
| }, |
| { |
| "epoch": 5.33557430982692, |
| "grad_norm": 0.7567323256836789, |
| "learning_rate": 5.082402685145088e-06, |
| "loss": 1.5373, |
| "step": 18650 |
| }, |
| { |
| "epoch": 5.3498784151051355, |
| "grad_norm": 0.6914309741643679, |
| "learning_rate": 5.062095122801771e-06, |
| "loss": 1.5353, |
| "step": 18700 |
| }, |
| { |
| "epoch": 5.36418252038335, |
| "grad_norm": 0.5912773883840152, |
| "learning_rate": 5.041796559832742e-06, |
| "loss": 1.5339, |
| "step": 18750 |
| }, |
| { |
| "epoch": 5.378486625661565, |
| "grad_norm": 0.6263016315822758, |
| "learning_rate": 5.021507413393355e-06, |
| "loss": 1.5409, |
| "step": 18800 |
| }, |
| { |
| "epoch": 5.378486625661565, |
| "eval_loss": 1.6224277019500732, |
| "eval_runtime": 13.1785, |
| "eval_samples_per_second": 75.881, |
| "eval_steps_per_second": 2.428, |
| "step": 18800 |
| }, |
| { |
| "epoch": 5.39279073093978, |
| "grad_norm": 0.5697288065062571, |
| "learning_rate": 5.001228100445451e-06, |
| "loss": 1.5362, |
| "step": 18850 |
| }, |
| { |
| "epoch": 5.407094836217994, |
| "grad_norm": 0.67253268999925, |
| "learning_rate": 4.9809590377487795e-06, |
| "loss": 1.5245, |
| "step": 18900 |
| }, |
| { |
| "epoch": 5.42139894149621, |
| "grad_norm": 0.6736080205591396, |
| "learning_rate": 4.960700641852442e-06, |
| "loss": 1.5263, |
| "step": 18950 |
| }, |
| { |
| "epoch": 5.4357030467744245, |
| "grad_norm": 0.7729780534104512, |
| "learning_rate": 4.9404533290863186e-06, |
| "loss": 1.5297, |
| "step": 19000 |
| }, |
| { |
| "epoch": 5.450007152052639, |
| "grad_norm": 0.6214293118686126, |
| "learning_rate": 4.920217515552526e-06, |
| "loss": 1.5397, |
| "step": 19050 |
| }, |
| { |
| "epoch": 5.464311257330854, |
| "grad_norm": 0.6343711973214149, |
| "learning_rate": 4.8999936171168615e-06, |
| "loss": 1.5439, |
| "step": 19100 |
| }, |
| { |
| "epoch": 5.478615362609069, |
| "grad_norm": 0.7291330555122011, |
| "learning_rate": 4.879782049400251e-06, |
| "loss": 1.5336, |
| "step": 19150 |
| }, |
| { |
| "epoch": 5.492919467887283, |
| "grad_norm": 0.6712685546736542, |
| "learning_rate": 4.8595832277702175e-06, |
| "loss": 1.5329, |
| "step": 19200 |
| }, |
| { |
| "epoch": 5.492919467887283, |
| "eval_loss": 1.622052788734436, |
| "eval_runtime": 13.144, |
| "eval_samples_per_second": 76.08, |
| "eval_steps_per_second": 2.435, |
| "step": 19200 |
| }, |
| { |
| "epoch": 5.507223573165499, |
| "grad_norm": 0.5709944496814116, |
| "learning_rate": 4.839397567332334e-06, |
| "loss": 1.5404, |
| "step": 19250 |
| }, |
| { |
| "epoch": 5.5215276784437135, |
| "grad_norm": 0.683232233789078, |
| "learning_rate": 4.8192254829217e-06, |
| "loss": 1.5422, |
| "step": 19300 |
| }, |
| { |
| "epoch": 5.535831783721928, |
| "grad_norm": 0.6192373192804173, |
| "learning_rate": 4.799067389094416e-06, |
| "loss": 1.5443, |
| "step": 19350 |
| }, |
| { |
| "epoch": 5.550135889000143, |
| "grad_norm": 0.7502591192271298, |
| "learning_rate": 4.7789237001190624e-06, |
| "loss": 1.5296, |
| "step": 19400 |
| }, |
| { |
| "epoch": 5.564439994278358, |
| "grad_norm": 0.6351610614349129, |
| "learning_rate": 4.758794829968181e-06, |
| "loss": 1.5348, |
| "step": 19450 |
| }, |
| { |
| "epoch": 5.578744099556573, |
| "grad_norm": 0.6108826056364483, |
| "learning_rate": 4.738681192309778e-06, |
| "loss": 1.5279, |
| "step": 19500 |
| }, |
| { |
| "epoch": 5.593048204834788, |
| "grad_norm": 0.6437182525691378, |
| "learning_rate": 4.718583200498814e-06, |
| "loss": 1.5366, |
| "step": 19550 |
| }, |
| { |
| "epoch": 5.6073523101130025, |
| "grad_norm": 0.639815861886649, |
| "learning_rate": 4.698501267568713e-06, |
| "loss": 1.5357, |
| "step": 19600 |
| }, |
| { |
| "epoch": 5.6073523101130025, |
| "eval_loss": 1.6180142164230347, |
| "eval_runtime": 13.085, |
| "eval_samples_per_second": 76.423, |
| "eval_steps_per_second": 2.446, |
| "step": 19600 |
| }, |
| { |
| "epoch": 5.621656415391217, |
| "grad_norm": 0.7200835319066389, |
| "learning_rate": 4.678435806222873e-06, |
| "loss": 1.5321, |
| "step": 19650 |
| }, |
| { |
| "epoch": 5.635960520669432, |
| "grad_norm": 0.783772141557461, |
| "learning_rate": 4.658387228826185e-06, |
| "loss": 1.5222, |
| "step": 19700 |
| }, |
| { |
| "epoch": 5.6502646259476474, |
| "grad_norm": 0.6132596760513117, |
| "learning_rate": 4.638355947396557e-06, |
| "loss": 1.5297, |
| "step": 19750 |
| }, |
| { |
| "epoch": 5.664568731225862, |
| "grad_norm": 0.6283378075518518, |
| "learning_rate": 4.61834237359645e-06, |
| "loss": 1.5295, |
| "step": 19800 |
| }, |
| { |
| "epoch": 5.678872836504077, |
| "grad_norm": 0.6326692578607431, |
| "learning_rate": 4.598346918724417e-06, |
| "loss": 1.5351, |
| "step": 19850 |
| }, |
| { |
| "epoch": 5.6931769417822915, |
| "grad_norm": 0.5940063235489941, |
| "learning_rate": 4.578369993706643e-06, |
| "loss": 1.5308, |
| "step": 19900 |
| }, |
| { |
| "epoch": 5.707481047060506, |
| "grad_norm": 0.6349874592624309, |
| "learning_rate": 4.5584120090885125e-06, |
| "loss": 1.5214, |
| "step": 19950 |
| }, |
| { |
| "epoch": 5.721785152338721, |
| "grad_norm": 0.7828872685863243, |
| "learning_rate": 4.538473375026164e-06, |
| "loss": 1.5428, |
| "step": 20000 |
| }, |
| { |
| "epoch": 5.721785152338721, |
| "eval_loss": 1.6172399520874023, |
| "eval_runtime": 13.1288, |
| "eval_samples_per_second": 76.168, |
| "eval_steps_per_second": 2.437, |
| "step": 20000 |
| }, |
| { |
| "epoch": 5.736089257616936, |
| "grad_norm": 0.6299009394488376, |
| "learning_rate": 4.518554501278064e-06, |
| "loss": 1.5295, |
| "step": 20050 |
| }, |
| { |
| "epoch": 5.750393362895151, |
| "grad_norm": 0.6258430789114957, |
| "learning_rate": 4.498655797196586e-06, |
| "loss": 1.5329, |
| "step": 20100 |
| }, |
| { |
| "epoch": 5.764697468173366, |
| "grad_norm": 0.696293581732556, |
| "learning_rate": 4.478777671719593e-06, |
| "loss": 1.5413, |
| "step": 20150 |
| }, |
| { |
| "epoch": 5.7790015734515805, |
| "grad_norm": 0.7124246186685519, |
| "learning_rate": 4.458920533362039e-06, |
| "loss": 1.5228, |
| "step": 20200 |
| }, |
| { |
| "epoch": 5.793305678729795, |
| "grad_norm": 0.5760090289969003, |
| "learning_rate": 4.439084790207577e-06, |
| "loss": 1.5353, |
| "step": 20250 |
| }, |
| { |
| "epoch": 5.80760978400801, |
| "grad_norm": 0.6449974415534122, |
| "learning_rate": 4.419270849900164e-06, |
| "loss": 1.5357, |
| "step": 20300 |
| }, |
| { |
| "epoch": 5.821913889286225, |
| "grad_norm": 0.6282375399916217, |
| "learning_rate": 4.399479119635683e-06, |
| "loss": 1.5321, |
| "step": 20350 |
| }, |
| { |
| "epoch": 5.83621799456444, |
| "grad_norm": 0.5919862101266721, |
| "learning_rate": 4.3797100061535856e-06, |
| "loss": 1.526, |
| "step": 20400 |
| }, |
| { |
| "epoch": 5.83621799456444, |
| "eval_loss": 1.6131237745285034, |
| "eval_runtime": 13.1139, |
| "eval_samples_per_second": 76.255, |
| "eval_steps_per_second": 2.44, |
| "step": 20400 |
| }, |
| { |
| "epoch": 5.850522099842655, |
| "grad_norm": 0.6029658262074238, |
| "learning_rate": 4.359963915728523e-06, |
| "loss": 1.5247, |
| "step": 20450 |
| }, |
| { |
| "epoch": 5.8648262051208695, |
| "grad_norm": 0.670869434044021, |
| "learning_rate": 4.340241254162004e-06, |
| "loss": 1.525, |
| "step": 20500 |
| }, |
| { |
| "epoch": 5.879130310399084, |
| "grad_norm": 0.6367180654906258, |
| "learning_rate": 4.320542426774042e-06, |
| "loss": 1.5269, |
| "step": 20550 |
| }, |
| { |
| "epoch": 5.8934344156773, |
| "grad_norm": 0.5920409715922154, |
| "learning_rate": 4.300867838394843e-06, |
| "loss": 1.5241, |
| "step": 20600 |
| }, |
| { |
| "epoch": 5.907738520955514, |
| "grad_norm": 0.5951037902934461, |
| "learning_rate": 4.281217893356478e-06, |
| "loss": 1.5404, |
| "step": 20650 |
| }, |
| { |
| "epoch": 5.922042626233729, |
| "grad_norm": 0.7114294289873735, |
| "learning_rate": 4.261592995484567e-06, |
| "loss": 1.5329, |
| "step": 20700 |
| }, |
| { |
| "epoch": 5.936346731511944, |
| "grad_norm": 0.5783828576596121, |
| "learning_rate": 4.241993548089994e-06, |
| "loss": 1.533, |
| "step": 20750 |
| }, |
| { |
| "epoch": 5.9506508367901585, |
| "grad_norm": 0.6407329854514554, |
| "learning_rate": 4.2224199539606026e-06, |
| "loss": 1.5145, |
| "step": 20800 |
| }, |
| { |
| "epoch": 5.9506508367901585, |
| "eval_loss": 1.6121162176132202, |
| "eval_runtime": 13.1147, |
| "eval_samples_per_second": 76.25, |
| "eval_steps_per_second": 2.44, |
| "step": 20800 |
| }, |
| { |
| "epoch": 5.964954942068374, |
| "grad_norm": 0.634888254231562, |
| "learning_rate": 4.202872615352932e-06, |
| "loss": 1.5212, |
| "step": 20850 |
| }, |
| { |
| "epoch": 5.979259047346589, |
| "grad_norm": 0.6229293024525429, |
| "learning_rate": 4.183351933983942e-06, |
| "loss": 1.5226, |
| "step": 20900 |
| }, |
| { |
| "epoch": 5.993563152624803, |
| "grad_norm": 0.8314381666700691, |
| "learning_rate": 4.163858311022765e-06, |
| "loss": 1.5387, |
| "step": 20950 |
| }, |
| { |
| "epoch": 6.007867257903018, |
| "grad_norm": 0.59693291322645, |
| "learning_rate": 4.144392147082447e-06, |
| "loss": 1.5043, |
| "step": 21000 |
| }, |
| { |
| "epoch": 6.022171363181233, |
| "grad_norm": 0.665194604166561, |
| "learning_rate": 4.124953842211733e-06, |
| "loss": 1.5002, |
| "step": 21050 |
| }, |
| { |
| "epoch": 6.0364754684594475, |
| "grad_norm": 0.6456678113415395, |
| "learning_rate": 4.105543795886834e-06, |
| "loss": 1.5004, |
| "step": 21100 |
| }, |
| { |
| "epoch": 6.050779573737663, |
| "grad_norm": 0.642020830639423, |
| "learning_rate": 4.086162407003219e-06, |
| "loss": 1.499, |
| "step": 21150 |
| }, |
| { |
| "epoch": 6.065083679015878, |
| "grad_norm": 0.6641186183494755, |
| "learning_rate": 4.066810073867421e-06, |
| "loss": 1.4973, |
| "step": 21200 |
| }, |
| { |
| "epoch": 6.065083679015878, |
| "eval_loss": 1.6121020317077637, |
| "eval_runtime": 13.0899, |
| "eval_samples_per_second": 76.395, |
| "eval_steps_per_second": 2.445, |
| "step": 21200 |
| }, |
| { |
| "epoch": 6.079387784294092, |
| "grad_norm": 0.634272316306304, |
| "learning_rate": 4.047487194188846e-06, |
| "loss": 1.5132, |
| "step": 21250 |
| }, |
| { |
| "epoch": 6.093691889572307, |
| "grad_norm": 0.6748754827417714, |
| "learning_rate": 4.028194165071603e-06, |
| "loss": 1.5064, |
| "step": 21300 |
| }, |
| { |
| "epoch": 6.107995994850522, |
| "grad_norm": 0.7061607465519819, |
| "learning_rate": 4.008931383006345e-06, |
| "loss": 1.505, |
| "step": 21350 |
| }, |
| { |
| "epoch": 6.122300100128737, |
| "grad_norm": 0.662154780004609, |
| "learning_rate": 3.9896992438621195e-06, |
| "loss": 1.5005, |
| "step": 21400 |
| }, |
| { |
| "epoch": 6.136604205406952, |
| "grad_norm": 0.6380113889018068, |
| "learning_rate": 3.9704981428782255e-06, |
| "loss": 1.5061, |
| "step": 21450 |
| }, |
| { |
| "epoch": 6.150908310685167, |
| "grad_norm": 0.6382768302952283, |
| "learning_rate": 3.951328474656103e-06, |
| "loss": 1.4931, |
| "step": 21500 |
| }, |
| { |
| "epoch": 6.165212415963381, |
| "grad_norm": 0.6767629241771801, |
| "learning_rate": 3.932190633151217e-06, |
| "loss": 1.5037, |
| "step": 21550 |
| }, |
| { |
| "epoch": 6.179516521241596, |
| "grad_norm": 0.6282216484213683, |
| "learning_rate": 3.913085011664961e-06, |
| "loss": 1.5036, |
| "step": 21600 |
| }, |
| { |
| "epoch": 6.179516521241596, |
| "eval_loss": 1.6118288040161133, |
| "eval_runtime": 13.1481, |
| "eval_samples_per_second": 76.057, |
| "eval_steps_per_second": 2.434, |
| "step": 21600 |
| }, |
| { |
| "epoch": 6.193820626519811, |
| "grad_norm": 0.6004617238206564, |
| "learning_rate": 3.894012002836578e-06, |
| "loss": 1.5117, |
| "step": 21650 |
| }, |
| { |
| "epoch": 6.208124731798026, |
| "grad_norm": 0.7154680728784966, |
| "learning_rate": 3.874971998635083e-06, |
| "loss": 1.5095, |
| "step": 21700 |
| }, |
| { |
| "epoch": 6.222428837076241, |
| "grad_norm": 0.6725566004485558, |
| "learning_rate": 3.855965390351222e-06, |
| "loss": 1.4909, |
| "step": 21750 |
| }, |
| { |
| "epoch": 6.236732942354456, |
| "grad_norm": 0.5965877047430266, |
| "learning_rate": 3.836992568589417e-06, |
| "loss": 1.4975, |
| "step": 21800 |
| }, |
| { |
| "epoch": 6.25103704763267, |
| "grad_norm": 0.6027834656468215, |
| "learning_rate": 3.818053923259745e-06, |
| "loss": 1.5007, |
| "step": 21850 |
| }, |
| { |
| "epoch": 6.265341152910885, |
| "grad_norm": 0.6663203163783424, |
| "learning_rate": 3.7991498435699213e-06, |
| "loss": 1.5051, |
| "step": 21900 |
| }, |
| { |
| "epoch": 6.279645258189101, |
| "grad_norm": 0.7125353969852749, |
| "learning_rate": 3.780280718017307e-06, |
| "loss": 1.5012, |
| "step": 21950 |
| }, |
| { |
| "epoch": 6.293949363467315, |
| "grad_norm": 0.6482680068197426, |
| "learning_rate": 3.7614469343809208e-06, |
| "loss": 1.5049, |
| "step": 22000 |
| }, |
| { |
| "epoch": 6.293949363467315, |
| "eval_loss": 1.6089048385620117, |
| "eval_runtime": 13.1237, |
| "eval_samples_per_second": 76.198, |
| "eval_steps_per_second": 2.438, |
| "step": 22000 |
| }, |
| { |
| "epoch": 6.30825346874553, |
| "grad_norm": 0.7881816467240778, |
| "learning_rate": 3.7426488797134706e-06, |
| "loss": 1.5095, |
| "step": 22050 |
| }, |
| { |
| "epoch": 6.322557574023745, |
| "grad_norm": 0.6541787463398775, |
| "learning_rate": 3.7238869403333965e-06, |
| "loss": 1.504, |
| "step": 22100 |
| }, |
| { |
| "epoch": 6.336861679301959, |
| "grad_norm": 0.664079330402415, |
| "learning_rate": 3.7051615018169363e-06, |
| "loss": 1.5066, |
| "step": 22150 |
| }, |
| { |
| "epoch": 6.351165784580174, |
| "grad_norm": 0.5916075057540389, |
| "learning_rate": 3.686472948990198e-06, |
| "loss": 1.5086, |
| "step": 22200 |
| }, |
| { |
| "epoch": 6.36546988985839, |
| "grad_norm": 0.7325464747774915, |
| "learning_rate": 3.6678216659212546e-06, |
| "loss": 1.4992, |
| "step": 22250 |
| }, |
| { |
| "epoch": 6.379773995136604, |
| "grad_norm": 0.7007083689793716, |
| "learning_rate": 3.6492080359122487e-06, |
| "loss": 1.5, |
| "step": 22300 |
| }, |
| { |
| "epoch": 6.394078100414819, |
| "grad_norm": 0.6739401831383736, |
| "learning_rate": 3.630632441491512e-06, |
| "loss": 1.4995, |
| "step": 22350 |
| }, |
| { |
| "epoch": 6.408382205693034, |
| "grad_norm": 0.6863089101260594, |
| "learning_rate": 3.6120952644057104e-06, |
| "loss": 1.5017, |
| "step": 22400 |
| }, |
| { |
| "epoch": 6.408382205693034, |
| "eval_loss": 1.606228232383728, |
| "eval_runtime": 13.1397, |
| "eval_samples_per_second": 76.105, |
| "eval_steps_per_second": 2.435, |
| "step": 22400 |
| }, |
| { |
| "epoch": 6.422686310971248, |
| "grad_norm": 0.6147754073523626, |
| "learning_rate": 3.593596885611997e-06, |
| "loss": 1.4926, |
| "step": 22450 |
| }, |
| { |
| "epoch": 6.436990416249464, |
| "grad_norm": 0.5871499975242569, |
| "learning_rate": 3.575137685270182e-06, |
| "loss": 1.4983, |
| "step": 22500 |
| }, |
| { |
| "epoch": 6.451294521527679, |
| "grad_norm": 0.5786188568209956, |
| "learning_rate": 3.556718042734915e-06, |
| "loss": 1.4949, |
| "step": 22550 |
| }, |
| { |
| "epoch": 6.465598626805893, |
| "grad_norm": 0.6071462043583132, |
| "learning_rate": 3.538338336547902e-06, |
| "loss": 1.4987, |
| "step": 22600 |
| }, |
| { |
| "epoch": 6.479902732084108, |
| "grad_norm": 0.5785217797393283, |
| "learning_rate": 3.5199989444301082e-06, |
| "loss": 1.4931, |
| "step": 22650 |
| }, |
| { |
| "epoch": 6.494206837362323, |
| "grad_norm": 0.623076022970078, |
| "learning_rate": 3.5017002432740164e-06, |
| "loss": 1.4981, |
| "step": 22700 |
| }, |
| { |
| "epoch": 6.508510942640537, |
| "grad_norm": 0.73033655866744, |
| "learning_rate": 3.483442609135862e-06, |
| "loss": 1.5008, |
| "step": 22750 |
| }, |
| { |
| "epoch": 6.522815047918753, |
| "grad_norm": 0.5934926417919788, |
| "learning_rate": 3.4652264172279153e-06, |
| "loss": 1.4881, |
| "step": 22800 |
| }, |
| { |
| "epoch": 6.522815047918753, |
| "eval_loss": 1.604162335395813, |
| "eval_runtime": 13.1475, |
| "eval_samples_per_second": 76.06, |
| "eval_steps_per_second": 2.434, |
| "step": 22800 |
| }, |
| { |
| "epoch": 6.537119153196968, |
| "grad_norm": 0.6069986350354387, |
| "learning_rate": 3.4470520419107667e-06, |
| "loss": 1.4879, |
| "step": 22850 |
| }, |
| { |
| "epoch": 6.551423258475182, |
| "grad_norm": 0.7399652835966712, |
| "learning_rate": 3.4289198566856396e-06, |
| "loss": 1.4942, |
| "step": 22900 |
| }, |
| { |
| "epoch": 6.565727363753397, |
| "grad_norm": 0.6114058208337956, |
| "learning_rate": 3.4108302341867045e-06, |
| "loss": 1.4961, |
| "step": 22950 |
| }, |
| { |
| "epoch": 6.580031469031612, |
| "grad_norm": 0.6007896515222266, |
| "learning_rate": 3.392783546173426e-06, |
| "loss": 1.4929, |
| "step": 23000 |
| }, |
| { |
| "epoch": 6.594335574309827, |
| "grad_norm": 0.7491177037600624, |
| "learning_rate": 3.374780163522929e-06, |
| "loss": 1.4955, |
| "step": 23050 |
| }, |
| { |
| "epoch": 6.608639679588042, |
| "grad_norm": 0.6917907740940776, |
| "learning_rate": 3.356820456222362e-06, |
| "loss": 1.5053, |
| "step": 23100 |
| }, |
| { |
| "epoch": 6.622943784866257, |
| "grad_norm": 0.5797836988455244, |
| "learning_rate": 3.338904793361314e-06, |
| "loss": 1.4974, |
| "step": 23150 |
| }, |
| { |
| "epoch": 6.637247890144471, |
| "grad_norm": 0.6589102327372058, |
| "learning_rate": 3.321033543124202e-06, |
| "loss": 1.4969, |
| "step": 23200 |
| }, |
| { |
| "epoch": 6.637247890144471, |
| "eval_loss": 1.6024819612503052, |
| "eval_runtime": 13.1262, |
| "eval_samples_per_second": 76.183, |
| "eval_steps_per_second": 2.438, |
| "step": 23200 |
| }, |
| { |
| "epoch": 6.651551995422686, |
| "grad_norm": 0.67042895269073, |
| "learning_rate": 3.3032070727827353e-06, |
| "loss": 1.4931, |
| "step": 23250 |
| }, |
| { |
| "epoch": 6.665856100700902, |
| "grad_norm": 0.6465762171859113, |
| "learning_rate": 3.285425748688341e-06, |
| "loss": 1.5047, |
| "step": 23300 |
| }, |
| { |
| "epoch": 6.680160205979116, |
| "grad_norm": 0.5928906211207966, |
| "learning_rate": 3.2676899362646563e-06, |
| "loss": 1.5009, |
| "step": 23350 |
| }, |
| { |
| "epoch": 6.694464311257331, |
| "grad_norm": 0.6204319354437976, |
| "learning_rate": 3.2500000000000015e-06, |
| "loss": 1.4904, |
| "step": 23400 |
| }, |
| { |
| "epoch": 6.708768416535546, |
| "grad_norm": 0.6370248234668529, |
| "learning_rate": 3.232356303439897e-06, |
| "loss": 1.4962, |
| "step": 23450 |
| }, |
| { |
| "epoch": 6.72307252181376, |
| "grad_norm": 0.5797494709987538, |
| "learning_rate": 3.2147592091795992e-06, |
| "loss": 1.4938, |
| "step": 23500 |
| }, |
| { |
| "epoch": 6.737376627091975, |
| "grad_norm": 0.5800297263686035, |
| "learning_rate": 3.1972090788566325e-06, |
| "loss": 1.4852, |
| "step": 23550 |
| }, |
| { |
| "epoch": 6.751680732370191, |
| "grad_norm": 0.6400211162161145, |
| "learning_rate": 3.1797062731433764e-06, |
| "loss": 1.4978, |
| "step": 23600 |
| }, |
| { |
| "epoch": 6.751680732370191, |
| "eval_loss": 1.6009289026260376, |
| "eval_runtime": 13.1205, |
| "eval_samples_per_second": 76.216, |
| "eval_steps_per_second": 2.439, |
| "step": 23600 |
| }, |
| { |
| "epoch": 6.765984837648405, |
| "grad_norm": 0.6047950853143369, |
| "learning_rate": 3.1622511517396297e-06, |
| "loss": 1.4931, |
| "step": 23650 |
| }, |
| { |
| "epoch": 6.78028894292662, |
| "grad_norm": 0.6976178519904986, |
| "learning_rate": 3.144844073365247e-06, |
| "loss": 1.4996, |
| "step": 23700 |
| }, |
| { |
| "epoch": 6.794593048204835, |
| "grad_norm": 0.6200936341605483, |
| "learning_rate": 3.127485395752739e-06, |
| "loss": 1.4874, |
| "step": 23750 |
| }, |
| { |
| "epoch": 6.808897153483049, |
| "grad_norm": 0.6678212868779091, |
| "learning_rate": 3.1101754756399427e-06, |
| "loss": 1.4951, |
| "step": 23800 |
| }, |
| { |
| "epoch": 6.823201258761264, |
| "grad_norm": 0.6777775649765228, |
| "learning_rate": 3.0929146687626756e-06, |
| "loss": 1.4936, |
| "step": 23850 |
| }, |
| { |
| "epoch": 6.83750536403948, |
| "grad_norm": 0.5932024233832688, |
| "learning_rate": 3.0757033298474274e-06, |
| "loss": 1.4908, |
| "step": 23900 |
| }, |
| { |
| "epoch": 6.851809469317694, |
| "grad_norm": 0.6344747435348402, |
| "learning_rate": 3.0585418126040832e-06, |
| "loss": 1.5015, |
| "step": 23950 |
| }, |
| { |
| "epoch": 6.866113574595909, |
| "grad_norm": 0.629407558903598, |
| "learning_rate": 3.0414304697186313e-06, |
| "loss": 1.4797, |
| "step": 24000 |
| }, |
| { |
| "epoch": 6.866113574595909, |
| "eval_loss": 1.5976924896240234, |
| "eval_runtime": 13.1566, |
| "eval_samples_per_second": 76.008, |
| "eval_steps_per_second": 2.432, |
| "step": 24000 |
| }, |
| { |
| "epoch": 6.880417679874124, |
| "grad_norm": 0.6197330636896008, |
| "learning_rate": 3.0243696528459392e-06, |
| "loss": 1.485, |
| "step": 24050 |
| }, |
| { |
| "epoch": 6.894721785152338, |
| "grad_norm": 0.6690327062688416, |
| "learning_rate": 3.007359712602504e-06, |
| "loss": 1.4944, |
| "step": 24100 |
| }, |
| { |
| "epoch": 6.909025890430554, |
| "grad_norm": 0.6435521457194968, |
| "learning_rate": 2.9904009985592685e-06, |
| "loss": 1.4966, |
| "step": 24150 |
| }, |
| { |
| "epoch": 6.923329995708769, |
| "grad_norm": 0.6127641972228168, |
| "learning_rate": 2.9734938592344188e-06, |
| "loss": 1.512, |
| "step": 24200 |
| }, |
| { |
| "epoch": 6.937634100986983, |
| "grad_norm": 0.6164050556423923, |
| "learning_rate": 2.9566386420862394e-06, |
| "loss": 1.488, |
| "step": 24250 |
| }, |
| { |
| "epoch": 6.951938206265198, |
| "grad_norm": 0.5843127773701906, |
| "learning_rate": 2.93983569350595e-06, |
| "loss": 1.494, |
| "step": 24300 |
| }, |
| { |
| "epoch": 6.966242311543413, |
| "grad_norm": 0.7288018217616302, |
| "learning_rate": 2.9230853588106135e-06, |
| "loss": 1.492, |
| "step": 24350 |
| }, |
| { |
| "epoch": 6.980546416821628, |
| "grad_norm": 0.6357813759666795, |
| "learning_rate": 2.9063879822360156e-06, |
| "loss": 1.4788, |
| "step": 24400 |
| }, |
| { |
| "epoch": 6.980546416821628, |
| "eval_loss": 1.597045660018921, |
| "eval_runtime": 13.1213, |
| "eval_samples_per_second": 76.212, |
| "eval_steps_per_second": 2.439, |
| "step": 24400 |
| }, |
| { |
| "epoch": 6.994850522099843, |
| "grad_norm": 0.5763643913962628, |
| "learning_rate": 2.889743906929609e-06, |
| "loss": 1.4907, |
| "step": 24450 |
| }, |
| { |
| "epoch": 7.009154627378058, |
| "grad_norm": 0.7160836057677523, |
| "learning_rate": 2.873153474943447e-06, |
| "loss": 1.4845, |
| "step": 24500 |
| }, |
| { |
| "epoch": 7.023458732656272, |
| "grad_norm": 0.6189813708548647, |
| "learning_rate": 2.85661702722716e-06, |
| "loss": 1.4738, |
| "step": 24550 |
| }, |
| { |
| "epoch": 7.037762837934487, |
| "grad_norm": 0.6666504346776301, |
| "learning_rate": 2.8401349036209563e-06, |
| "loss": 1.4773, |
| "step": 24600 |
| }, |
| { |
| "epoch": 7.052066943212702, |
| "grad_norm": 0.6423857792649676, |
| "learning_rate": 2.82370744284862e-06, |
| "loss": 1.4765, |
| "step": 24650 |
| }, |
| { |
| "epoch": 7.066371048490917, |
| "grad_norm": 0.6197018673973147, |
| "learning_rate": 2.807334982510572e-06, |
| "loss": 1.4769, |
| "step": 24700 |
| }, |
| { |
| "epoch": 7.080675153769132, |
| "grad_norm": 0.6066963496391917, |
| "learning_rate": 2.7910178590769056e-06, |
| "loss": 1.4796, |
| "step": 24750 |
| }, |
| { |
| "epoch": 7.094979259047347, |
| "grad_norm": 0.6142726779110321, |
| "learning_rate": 2.774756407880501e-06, |
| "loss": 1.4657, |
| "step": 24800 |
| }, |
| { |
| "epoch": 7.094979259047347, |
| "eval_loss": 1.5973756313323975, |
| "eval_runtime": 13.1335, |
| "eval_samples_per_second": 76.141, |
| "eval_steps_per_second": 2.437, |
| "step": 24800 |
| }, |
| { |
| "epoch": 7.109283364325561, |
| "grad_norm": 0.7048459012957223, |
| "learning_rate": 2.7585509631101103e-06, |
| "loss": 1.474, |
| "step": 24850 |
| }, |
| { |
| "epoch": 7.123587469603776, |
| "grad_norm": 0.5827521156236716, |
| "learning_rate": 2.7424018578035044e-06, |
| "loss": 1.4764, |
| "step": 24900 |
| }, |
| { |
| "epoch": 7.1378915748819916, |
| "grad_norm": 0.5957200643125377, |
| "learning_rate": 2.726309423840619e-06, |
| "loss": 1.4737, |
| "step": 24950 |
| }, |
| { |
| "epoch": 7.152195680160206, |
| "grad_norm": 0.6084463118656662, |
| "learning_rate": 2.710273991936737e-06, |
| "loss": 1.4848, |
| "step": 25000 |
| }, |
| { |
| "epoch": 7.166499785438421, |
| "grad_norm": 0.6674457907135838, |
| "learning_rate": 2.6942958916356993e-06, |
| "loss": 1.4798, |
| "step": 25050 |
| }, |
| { |
| "epoch": 7.180803890716636, |
| "grad_norm": 0.5705496515873673, |
| "learning_rate": 2.6783754513031197e-06, |
| "loss": 1.4795, |
| "step": 25100 |
| }, |
| { |
| "epoch": 7.19510799599485, |
| "grad_norm": 0.6575731057899443, |
| "learning_rate": 2.66251299811965e-06, |
| "loss": 1.4724, |
| "step": 25150 |
| }, |
| { |
| "epoch": 7.209412101273065, |
| "grad_norm": 0.6126232462556399, |
| "learning_rate": 2.6467088580742396e-06, |
| "loss": 1.4649, |
| "step": 25200 |
| }, |
| { |
| "epoch": 7.209412101273065, |
| "eval_loss": 1.5954419374465942, |
| "eval_runtime": 13.1247, |
| "eval_samples_per_second": 76.192, |
| "eval_steps_per_second": 2.438, |
| "step": 25200 |
| }, |
| { |
| "epoch": 7.2237162065512805, |
| "grad_norm": 0.564582536044739, |
| "learning_rate": 2.6309633559574547e-06, |
| "loss": 1.4781, |
| "step": 25250 |
| }, |
| { |
| "epoch": 7.238020311829495, |
| "grad_norm": 0.600751823096888, |
| "learning_rate": 2.6152768153547895e-06, |
| "loss": 1.4799, |
| "step": 25300 |
| }, |
| { |
| "epoch": 7.25232441710771, |
| "grad_norm": 0.6221919435914511, |
| "learning_rate": 2.599649558640026e-06, |
| "loss": 1.474, |
| "step": 25350 |
| }, |
| { |
| "epoch": 7.266628522385925, |
| "grad_norm": 0.6376984326932055, |
| "learning_rate": 2.584081906968599e-06, |
| "loss": 1.4717, |
| "step": 25400 |
| }, |
| { |
| "epoch": 7.280932627664139, |
| "grad_norm": 0.6231457394688249, |
| "learning_rate": 2.5685741802710025e-06, |
| "loss": 1.4777, |
| "step": 25450 |
| }, |
| { |
| "epoch": 7.295236732942355, |
| "grad_norm": 0.6912656538113927, |
| "learning_rate": 2.5531266972462176e-06, |
| "loss": 1.467, |
| "step": 25500 |
| }, |
| { |
| "epoch": 7.3095408382205695, |
| "grad_norm": 0.575138973241968, |
| "learning_rate": 2.537739775355153e-06, |
| "loss": 1.4708, |
| "step": 25550 |
| }, |
| { |
| "epoch": 7.323844943498784, |
| "grad_norm": 0.6138825273521185, |
| "learning_rate": 2.522413730814134e-06, |
| "loss": 1.4782, |
| "step": 25600 |
| }, |
| { |
| "epoch": 7.323844943498784, |
| "eval_loss": 1.594567894935608, |
| "eval_runtime": 13.1256, |
| "eval_samples_per_second": 76.187, |
| "eval_steps_per_second": 2.438, |
| "step": 25600 |
| }, |
| { |
| "epoch": 7.338149048776999, |
| "grad_norm": 0.6597269360462739, |
| "learning_rate": 2.507148878588387e-06, |
| "loss": 1.4745, |
| "step": 25650 |
| }, |
| { |
| "epoch": 7.352453154055214, |
| "grad_norm": 0.662162283465333, |
| "learning_rate": 2.4919455323855844e-06, |
| "loss": 1.472, |
| "step": 25700 |
| }, |
| { |
| "epoch": 7.366757259333428, |
| "grad_norm": 0.6105206167854126, |
| "learning_rate": 2.476804004649387e-06, |
| "loss": 1.4783, |
| "step": 25750 |
| }, |
| { |
| "epoch": 7.381061364611644, |
| "grad_norm": 0.6501751645549608, |
| "learning_rate": 2.461724606553027e-06, |
| "loss": 1.4612, |
| "step": 25800 |
| }, |
| { |
| "epoch": 7.3953654698898585, |
| "grad_norm": 0.6581591107567734, |
| "learning_rate": 2.4467076479929057e-06, |
| "loss": 1.4592, |
| "step": 25850 |
| }, |
| { |
| "epoch": 7.409669575168073, |
| "grad_norm": 0.6128493099219342, |
| "learning_rate": 2.43175343758224e-06, |
| "loss": 1.4709, |
| "step": 25900 |
| }, |
| { |
| "epoch": 7.423973680446288, |
| "grad_norm": 0.6202369817681859, |
| "learning_rate": 2.416862282644701e-06, |
| "loss": 1.4728, |
| "step": 25950 |
| }, |
| { |
| "epoch": 7.438277785724503, |
| "grad_norm": 0.6121747649704826, |
| "learning_rate": 2.4020344892081182e-06, |
| "loss": 1.4827, |
| "step": 26000 |
| }, |
| { |
| "epoch": 7.438277785724503, |
| "eval_loss": 1.593306064605713, |
| "eval_runtime": 13.1334, |
| "eval_samples_per_second": 76.142, |
| "eval_steps_per_second": 2.437, |
| "step": 26000 |
| }, |
| { |
| "epoch": 7.452581891002718, |
| "grad_norm": 0.6059881035687115, |
| "learning_rate": 2.3872703619981724e-06, |
| "loss": 1.4706, |
| "step": 26050 |
| }, |
| { |
| "epoch": 7.466885996280933, |
| "grad_norm": 0.6284120118138133, |
| "learning_rate": 2.3725702044321425e-06, |
| "loss": 1.4664, |
| "step": 26100 |
| }, |
| { |
| "epoch": 7.4811901015591475, |
| "grad_norm": 0.6757179221130658, |
| "learning_rate": 2.357934318612673e-06, |
| "loss": 1.4678, |
| "step": 26150 |
| }, |
| { |
| "epoch": 7.495494206837362, |
| "grad_norm": 0.6351885578913182, |
| "learning_rate": 2.343363005321554e-06, |
| "loss": 1.4616, |
| "step": 26200 |
| }, |
| { |
| "epoch": 7.509798312115577, |
| "grad_norm": 0.6372246562702591, |
| "learning_rate": 2.3288565640135553e-06, |
| "loss": 1.4691, |
| "step": 26250 |
| }, |
| { |
| "epoch": 7.524102417393792, |
| "grad_norm": 0.6099371593074536, |
| "learning_rate": 2.3144152928102537e-06, |
| "loss": 1.4694, |
| "step": 26300 |
| }, |
| { |
| "epoch": 7.538406522672007, |
| "grad_norm": 0.596517259235042, |
| "learning_rate": 2.3000394884939266e-06, |
| "loss": 1.4582, |
| "step": 26350 |
| }, |
| { |
| "epoch": 7.552710627950222, |
| "grad_norm": 0.6473760742475557, |
| "learning_rate": 2.285729446501434e-06, |
| "loss": 1.476, |
| "step": 26400 |
| }, |
| { |
| "epoch": 7.552710627950222, |
| "eval_loss": 1.5922271013259888, |
| "eval_runtime": 13.1523, |
| "eval_samples_per_second": 76.032, |
| "eval_steps_per_second": 2.433, |
| "step": 26400 |
| }, |
| { |
| "epoch": 7.5670147332284365, |
| "grad_norm": 0.727053136714405, |
| "learning_rate": 2.2714854609181638e-06, |
| "loss": 1.464, |
| "step": 26450 |
| }, |
| { |
| "epoch": 7.581318838506651, |
| "grad_norm": 0.6656286605140307, |
| "learning_rate": 2.2573078244719744e-06, |
| "loss": 1.4528, |
| "step": 26500 |
| }, |
| { |
| "epoch": 7.595622943784866, |
| "grad_norm": 0.6156105694109003, |
| "learning_rate": 2.2431968285271843e-06, |
| "loss": 1.4573, |
| "step": 26550 |
| }, |
| { |
| "epoch": 7.6099270490630815, |
| "grad_norm": 0.5943824923661235, |
| "learning_rate": 2.2291527630785895e-06, |
| "loss": 1.4644, |
| "step": 26600 |
| }, |
| { |
| "epoch": 7.624231154341296, |
| "grad_norm": 0.6737657950634879, |
| "learning_rate": 2.2151759167454918e-06, |
| "loss": 1.4592, |
| "step": 26650 |
| }, |
| { |
| "epoch": 7.638535259619511, |
| "grad_norm": 0.6052726703684476, |
| "learning_rate": 2.2012665767657825e-06, |
| "loss": 1.4735, |
| "step": 26700 |
| }, |
| { |
| "epoch": 7.6528393648977255, |
| "grad_norm": 0.634699066703415, |
| "learning_rate": 2.1874250289900216e-06, |
| "loss": 1.4721, |
| "step": 26750 |
| }, |
| { |
| "epoch": 7.66714347017594, |
| "grad_norm": 0.6282958976359639, |
| "learning_rate": 2.1736515578755813e-06, |
| "loss": 1.4719, |
| "step": 26800 |
| }, |
| { |
| "epoch": 7.66714347017594, |
| "eval_loss": 1.589935541152954, |
| "eval_runtime": 13.1215, |
| "eval_samples_per_second": 76.211, |
| "eval_steps_per_second": 2.439, |
| "step": 26800 |
| }, |
| { |
| "epoch": 7.681447575454156, |
| "grad_norm": 0.635973478905671, |
| "learning_rate": 2.159946446480785e-06, |
| "loss": 1.4667, |
| "step": 26850 |
| }, |
| { |
| "epoch": 7.6957516807323705, |
| "grad_norm": 0.6300462107788244, |
| "learning_rate": 2.146309976459104e-06, |
| "loss": 1.4677, |
| "step": 26900 |
| }, |
| { |
| "epoch": 7.710055786010585, |
| "grad_norm": 0.6136610995344491, |
| "learning_rate": 2.132742428053357e-06, |
| "loss": 1.4725, |
| "step": 26950 |
| }, |
| { |
| "epoch": 7.7243598912888, |
| "grad_norm": 0.5811410926679371, |
| "learning_rate": 2.1192440800899543e-06, |
| "loss": 1.461, |
| "step": 27000 |
| }, |
| { |
| "epoch": 7.7386639965670145, |
| "grad_norm": 0.5666126298419032, |
| "learning_rate": 2.105815209973176e-06, |
| "loss": 1.4561, |
| "step": 27050 |
| }, |
| { |
| "epoch": 7.752968101845229, |
| "grad_norm": 0.720102181736154, |
| "learning_rate": 2.0924560936794586e-06, |
| "loss": 1.4715, |
| "step": 27100 |
| }, |
| { |
| "epoch": 7.767272207123445, |
| "grad_norm": 0.6405735149181649, |
| "learning_rate": 2.0791670057517345e-06, |
| "loss": 1.4643, |
| "step": 27150 |
| }, |
| { |
| "epoch": 7.7815763124016595, |
| "grad_norm": 0.6229647710776449, |
| "learning_rate": 2.0659482192937757e-06, |
| "loss": 1.4609, |
| "step": 27200 |
| }, |
| { |
| "epoch": 7.7815763124016595, |
| "eval_loss": 1.5890839099884033, |
| "eval_runtime": 13.0892, |
| "eval_samples_per_second": 76.399, |
| "eval_steps_per_second": 2.445, |
| "step": 27200 |
| }, |
| { |
| "epoch": 7.795880417679874, |
| "grad_norm": 0.6101983306699219, |
| "learning_rate": 2.0528000059646e-06, |
| "loss": 1.4685, |
| "step": 27250 |
| }, |
| { |
| "epoch": 7.810184522958089, |
| "grad_norm": 0.6123847790160437, |
| "learning_rate": 2.0397226359728705e-06, |
| "loss": 1.4721, |
| "step": 27300 |
| }, |
| { |
| "epoch": 7.8244886282363035, |
| "grad_norm": 0.5890693272303438, |
| "learning_rate": 2.026716378071358e-06, |
| "loss": 1.4583, |
| "step": 27350 |
| }, |
| { |
| "epoch": 7.838792733514518, |
| "grad_norm": 0.6676445531960887, |
| "learning_rate": 2.0137814995514036e-06, |
| "loss": 1.472, |
| "step": 27400 |
| }, |
| { |
| "epoch": 7.853096838792734, |
| "grad_norm": 0.6850665159607517, |
| "learning_rate": 2.0009182662374332e-06, |
| "loss": 1.4635, |
| "step": 27450 |
| }, |
| { |
| "epoch": 7.8674009440709485, |
| "grad_norm": 0.6649159573122772, |
| "learning_rate": 1.9881269424814978e-06, |
| "loss": 1.4608, |
| "step": 27500 |
| }, |
| { |
| "epoch": 7.881705049349163, |
| "grad_norm": 0.5827299751362036, |
| "learning_rate": 1.9754077911578308e-06, |
| "loss": 1.4684, |
| "step": 27550 |
| }, |
| { |
| "epoch": 7.896009154627378, |
| "grad_norm": 0.6349322275652337, |
| "learning_rate": 1.9627610736574575e-06, |
| "loss": 1.4635, |
| "step": 27600 |
| }, |
| { |
| "epoch": 7.896009154627378, |
| "eval_loss": 1.588295340538025, |
| "eval_runtime": 13.3701, |
| "eval_samples_per_second": 74.794, |
| "eval_steps_per_second": 2.393, |
| "step": 27600 |
| }, |
| { |
| "epoch": 7.9103132599055925, |
| "grad_norm": 0.6758209768171158, |
| "learning_rate": 1.950187049882809e-06, |
| "loss": 1.4718, |
| "step": 27650 |
| }, |
| { |
| "epoch": 7.924617365183808, |
| "grad_norm": 0.5969675030716447, |
| "learning_rate": 1.937685978242398e-06, |
| "loss": 1.4634, |
| "step": 27700 |
| }, |
| { |
| "epoch": 7.938921470462023, |
| "grad_norm": 0.615432263769909, |
| "learning_rate": 1.925258115645493e-06, |
| "loss": 1.4759, |
| "step": 27750 |
| }, |
| { |
| "epoch": 7.9532255757402375, |
| "grad_norm": 0.6053585361833029, |
| "learning_rate": 1.9129037174968502e-06, |
| "loss": 1.4682, |
| "step": 27800 |
| }, |
| { |
| "epoch": 7.967529681018452, |
| "grad_norm": 0.6088696167731431, |
| "learning_rate": 1.900623037691452e-06, |
| "loss": 1.4604, |
| "step": 27850 |
| }, |
| { |
| "epoch": 7.981833786296667, |
| "grad_norm": 0.5809473580219169, |
| "learning_rate": 1.888416328609304e-06, |
| "loss": 1.4616, |
| "step": 27900 |
| }, |
| { |
| "epoch": 7.996137891574882, |
| "grad_norm": 0.5936177817067919, |
| "learning_rate": 1.8762838411102363e-06, |
| "loss": 1.4598, |
| "step": 27950 |
| }, |
| { |
| "epoch": 8.010441996853096, |
| "grad_norm": 0.6361295961404093, |
| "learning_rate": 1.8642258245287565e-06, |
| "loss": 1.45, |
| "step": 28000 |
| }, |
| { |
| "epoch": 8.010441996853096, |
| "eval_loss": 1.5894309282302856, |
| "eval_runtime": 13.1448, |
| "eval_samples_per_second": 76.076, |
| "eval_steps_per_second": 2.434, |
| "step": 28000 |
| }, |
| { |
| "epoch": 8.024746102131312, |
| "grad_norm": 0.6265798017363853, |
| "learning_rate": 1.8522425266689188e-06, |
| "loss": 1.4425, |
| "step": 28050 |
| }, |
| { |
| "epoch": 8.039050207409527, |
| "grad_norm": 0.6956974297968518, |
| "learning_rate": 1.8403341937992347e-06, |
| "loss": 1.4448, |
| "step": 28100 |
| }, |
| { |
| "epoch": 8.053354312687741, |
| "grad_norm": 0.627083132895609, |
| "learning_rate": 1.8285010706476137e-06, |
| "loss": 1.4438, |
| "step": 28150 |
| }, |
| { |
| "epoch": 8.067658417965957, |
| "grad_norm": 0.5941490143645667, |
| "learning_rate": 1.816743400396329e-06, |
| "loss": 1.4466, |
| "step": 28200 |
| }, |
| { |
| "epoch": 8.08196252324417, |
| "grad_norm": 0.6550383056050889, |
| "learning_rate": 1.8050614246770276e-06, |
| "loss": 1.4496, |
| "step": 28250 |
| }, |
| { |
| "epoch": 8.096266628522386, |
| "grad_norm": 0.5964983287141283, |
| "learning_rate": 1.793455383565753e-06, |
| "loss": 1.4663, |
| "step": 28300 |
| }, |
| { |
| "epoch": 8.1105707338006, |
| "grad_norm": 0.5934997527841852, |
| "learning_rate": 1.7819255155780241e-06, |
| "loss": 1.45, |
| "step": 28350 |
| }, |
| { |
| "epoch": 8.124874839078815, |
| "grad_norm": 0.5802546396980475, |
| "learning_rate": 1.7704720576639222e-06, |
| "loss": 1.4489, |
| "step": 28400 |
| }, |
| { |
| "epoch": 8.124874839078815, |
| "eval_loss": 1.5889995098114014, |
| "eval_runtime": 13.1217, |
| "eval_samples_per_second": 76.209, |
| "eval_steps_per_second": 2.439, |
| "step": 28400 |
| }, |
| { |
| "epoch": 8.139178944357031, |
| "grad_norm": 0.7065233633446718, |
| "learning_rate": 1.7590952452032333e-06, |
| "loss": 1.444, |
| "step": 28450 |
| }, |
| { |
| "epoch": 8.153483049635245, |
| "grad_norm": 0.6003608522086217, |
| "learning_rate": 1.7477953120005984e-06, |
| "loss": 1.4426, |
| "step": 28500 |
| }, |
| { |
| "epoch": 8.16778715491346, |
| "grad_norm": 0.6131368204140788, |
| "learning_rate": 1.7365724902807154e-06, |
| "loss": 1.4576, |
| "step": 28550 |
| }, |
| { |
| "epoch": 8.182091260191674, |
| "grad_norm": 0.6074281361277954, |
| "learning_rate": 1.7254270106835693e-06, |
| "loss": 1.4492, |
| "step": 28600 |
| }, |
| { |
| "epoch": 8.19639536546989, |
| "grad_norm": 0.6007813773273907, |
| "learning_rate": 1.7143591022596841e-06, |
| "loss": 1.4474, |
| "step": 28650 |
| }, |
| { |
| "epoch": 8.210699470748105, |
| "grad_norm": 0.6385077568660746, |
| "learning_rate": 1.7033689924654254e-06, |
| "loss": 1.4529, |
| "step": 28700 |
| }, |
| { |
| "epoch": 8.22500357602632, |
| "grad_norm": 0.6689966105698311, |
| "learning_rate": 1.692456907158313e-06, |
| "loss": 1.4412, |
| "step": 28750 |
| }, |
| { |
| "epoch": 8.239307681304535, |
| "grad_norm": 0.6146056122956874, |
| "learning_rate": 1.681623070592395e-06, |
| "loss": 1.4457, |
| "step": 28800 |
| }, |
| { |
| "epoch": 8.239307681304535, |
| "eval_loss": 1.587373971939087, |
| "eval_runtime": 13.1314, |
| "eval_samples_per_second": 76.154, |
| "eval_steps_per_second": 2.437, |
| "step": 28800 |
| }, |
| { |
| "epoch": 8.253611786582749, |
| "grad_norm": 0.6073017812921564, |
| "learning_rate": 1.6708677054136257e-06, |
| "loss": 1.4501, |
| "step": 28850 |
| }, |
| { |
| "epoch": 8.267915891860964, |
| "grad_norm": 0.6063223190404314, |
| "learning_rate": 1.6601910326553e-06, |
| "loss": 1.4507, |
| "step": 28900 |
| }, |
| { |
| "epoch": 8.28221999713918, |
| "grad_norm": 0.7473201451100423, |
| "learning_rate": 1.6495932717335023e-06, |
| "loss": 1.4567, |
| "step": 28950 |
| }, |
| { |
| "epoch": 8.296524102417393, |
| "grad_norm": 0.6223223319003798, |
| "learning_rate": 1.6390746404426033e-06, |
| "loss": 1.4492, |
| "step": 29000 |
| }, |
| { |
| "epoch": 8.310828207695609, |
| "grad_norm": 0.5404620482222038, |
| "learning_rate": 1.6286353549507834e-06, |
| "loss": 1.4523, |
| "step": 29050 |
| }, |
| { |
| "epoch": 8.325132312973823, |
| "grad_norm": 0.6204257838737108, |
| "learning_rate": 1.6182756297955865e-06, |
| "loss": 1.4623, |
| "step": 29100 |
| }, |
| { |
| "epoch": 8.339436418252038, |
| "grad_norm": 0.6659053743652089, |
| "learning_rate": 1.6079956778795175e-06, |
| "loss": 1.4484, |
| "step": 29150 |
| }, |
| { |
| "epoch": 8.353740523530254, |
| "grad_norm": 0.6506237144901641, |
| "learning_rate": 1.5977957104656564e-06, |
| "loss": 1.4474, |
| "step": 29200 |
| }, |
| { |
| "epoch": 8.353740523530254, |
| "eval_loss": 1.5863673686981201, |
| "eval_runtime": 13.1687, |
| "eval_samples_per_second": 75.938, |
| "eval_steps_per_second": 2.43, |
| "step": 29200 |
| }, |
| { |
| "epoch": 8.368044628808468, |
| "grad_norm": 0.643759788851684, |
| "learning_rate": 1.587675937173331e-06, |
| "loss": 1.4567, |
| "step": 29250 |
| }, |
| { |
| "epoch": 8.382348734086683, |
| "grad_norm": 0.6199745990986755, |
| "learning_rate": 1.5776365659737958e-06, |
| "loss": 1.4537, |
| "step": 29300 |
| }, |
| { |
| "epoch": 8.396652839364897, |
| "grad_norm": 0.5797274774601332, |
| "learning_rate": 1.5676778031859679e-06, |
| "loss": 1.4489, |
| "step": 29350 |
| }, |
| { |
| "epoch": 8.410956944643113, |
| "grad_norm": 0.5810915517535926, |
| "learning_rate": 1.5577998534721782e-06, |
| "loss": 1.4689, |
| "step": 29400 |
| }, |
| { |
| "epoch": 8.425261049921328, |
| "grad_norm": 0.6444744706356923, |
| "learning_rate": 1.548002919833971e-06, |
| "loss": 1.4421, |
| "step": 29450 |
| }, |
| { |
| "epoch": 8.439565155199542, |
| "grad_norm": 0.6515556286790462, |
| "learning_rate": 1.5382872036079344e-06, |
| "loss": 1.44, |
| "step": 29500 |
| }, |
| { |
| "epoch": 8.453869260477758, |
| "grad_norm": 0.5984018344351992, |
| "learning_rate": 1.528652904461555e-06, |
| "loss": 1.4447, |
| "step": 29550 |
| }, |
| { |
| "epoch": 8.468173365755971, |
| "grad_norm": 0.5768878455118189, |
| "learning_rate": 1.5191002203891188e-06, |
| "loss": 1.4539, |
| "step": 29600 |
| }, |
| { |
| "epoch": 8.468173365755971, |
| "eval_loss": 1.5855129957199097, |
| "eval_runtime": 13.1303, |
| "eval_samples_per_second": 76.16, |
| "eval_steps_per_second": 2.437, |
| "step": 29600 |
| }, |
| { |
| "epoch": 8.482477471034187, |
| "grad_norm": 0.608395053183915, |
| "learning_rate": 1.5096293477076443e-06, |
| "loss": 1.4449, |
| "step": 29650 |
| }, |
| { |
| "epoch": 8.4967815763124, |
| "grad_norm": 0.5908157542584366, |
| "learning_rate": 1.5002404810528447e-06, |
| "loss": 1.4453, |
| "step": 29700 |
| }, |
| { |
| "epoch": 8.511085681590616, |
| "grad_norm": 0.6155716861205135, |
| "learning_rate": 1.490933813375128e-06, |
| "loss": 1.4443, |
| "step": 29750 |
| }, |
| { |
| "epoch": 8.525389786868832, |
| "grad_norm": 0.6143329440860966, |
| "learning_rate": 1.4817095359356365e-06, |
| "loss": 1.4445, |
| "step": 29800 |
| }, |
| { |
| "epoch": 8.539693892147046, |
| "grad_norm": 0.6368160686195383, |
| "learning_rate": 1.4725678383023057e-06, |
| "loss": 1.4588, |
| "step": 29850 |
| }, |
| { |
| "epoch": 8.553997997425261, |
| "grad_norm": 0.608560533258368, |
| "learning_rate": 1.463508908345982e-06, |
| "loss": 1.4528, |
| "step": 29900 |
| }, |
| { |
| "epoch": 8.568302102703475, |
| "grad_norm": 0.6075645930538407, |
| "learning_rate": 1.4545329322365499e-06, |
| "loss": 1.4452, |
| "step": 29950 |
| }, |
| { |
| "epoch": 8.58260620798169, |
| "grad_norm": 0.6018775613036687, |
| "learning_rate": 1.4456400944391143e-06, |
| "loss": 1.4527, |
| "step": 30000 |
| }, |
| { |
| "epoch": 8.58260620798169, |
| "eval_loss": 1.5849485397338867, |
| "eval_runtime": 13.1197, |
| "eval_samples_per_second": 76.222, |
| "eval_steps_per_second": 2.439, |
| "step": 30000 |
| }, |
| { |
| "epoch": 8.596910313259906, |
| "grad_norm": 0.6041249971976146, |
| "learning_rate": 1.4368305777102054e-06, |
| "loss": 1.4478, |
| "step": 30050 |
| }, |
| { |
| "epoch": 8.61121441853812, |
| "grad_norm": 0.6621291172764414, |
| "learning_rate": 1.4281045630940216e-06, |
| "loss": 1.4589, |
| "step": 30100 |
| }, |
| { |
| "epoch": 8.625518523816336, |
| "grad_norm": 0.5815306829701986, |
| "learning_rate": 1.4194622299187155e-06, |
| "loss": 1.4507, |
| "step": 30150 |
| }, |
| { |
| "epoch": 8.63982262909455, |
| "grad_norm": 0.5783871548745824, |
| "learning_rate": 1.4109037557926999e-06, |
| "loss": 1.4578, |
| "step": 30200 |
| }, |
| { |
| "epoch": 8.654126734372765, |
| "grad_norm": 0.6297453226331882, |
| "learning_rate": 1.402429316601006e-06, |
| "loss": 1.4511, |
| "step": 30250 |
| }, |
| { |
| "epoch": 8.66843083965098, |
| "grad_norm": 0.6123759982689222, |
| "learning_rate": 1.3940390865016589e-06, |
| "loss": 1.4584, |
| "step": 30300 |
| }, |
| { |
| "epoch": 8.682734944929194, |
| "grad_norm": 0.5803997784485998, |
| "learning_rate": 1.385733237922111e-06, |
| "loss": 1.4543, |
| "step": 30350 |
| }, |
| { |
| "epoch": 8.69703905020741, |
| "grad_norm": 0.6530263919571504, |
| "learning_rate": 1.377511941555687e-06, |
| "loss": 1.436, |
| "step": 30400 |
| }, |
| { |
| "epoch": 8.69703905020741, |
| "eval_loss": 1.5838772058486938, |
| "eval_runtime": 13.1282, |
| "eval_samples_per_second": 76.172, |
| "eval_steps_per_second": 2.437, |
| "step": 30400 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 34950, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 10, |
| "save_steps": 800, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2.703424305810637e+16, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|