jekunz's picture
Upload folder using huggingface_hub
74d9d28 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 6.0,
"eval_steps": 500,
"global_step": 3084,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.19455252918287938,
"grad_norm": 5.050010681152344,
"learning_rate": 1.926070038910506e-05,
"loss": 8.5926,
"step": 100
},
{
"epoch": 0.38910505836575876,
"grad_norm": 1.6030714511871338,
"learning_rate": 3.8715953307393e-05,
"loss": 6.6577,
"step": 200
},
{
"epoch": 0.5836575875486382,
"grad_norm": 1.002219557762146,
"learning_rate": 5.8171206225680944e-05,
"loss": 5.4884,
"step": 300
},
{
"epoch": 0.7782101167315175,
"grad_norm": 1.1130906343460083,
"learning_rate": 7.762645914396888e-05,
"loss": 5.1208,
"step": 400
},
{
"epoch": 0.9727626459143969,
"grad_norm": 1.2561383247375488,
"learning_rate": 9.708171206225682e-05,
"loss": 5.0439,
"step": 500
},
{
"epoch": 1.0,
"eval_loss": 4.548050403594971,
"eval_runtime": 0.386,
"eval_samples_per_second": 593.336,
"eval_steps_per_second": 75.139,
"step": 514
},
{
"epoch": 1.1673151750972763,
"grad_norm": 1.0363208055496216,
"learning_rate": 9.816255944660613e-05,
"loss": 4.8297,
"step": 600
},
{
"epoch": 1.3618677042801557,
"grad_norm": 1.2391234636306763,
"learning_rate": 9.600086467790748e-05,
"loss": 4.6257,
"step": 700
},
{
"epoch": 1.556420233463035,
"grad_norm": 1.268858551979065,
"learning_rate": 9.383916990920882e-05,
"loss": 4.5661,
"step": 800
},
{
"epoch": 1.7509727626459144,
"grad_norm": 1.4311838150024414,
"learning_rate": 9.167747514051016e-05,
"loss": 4.5078,
"step": 900
},
{
"epoch": 1.9455252918287937,
"grad_norm": 1.9428343772888184,
"learning_rate": 8.951578037181151e-05,
"loss": 4.4511,
"step": 1000
},
{
"epoch": 2.0,
"eval_loss": 4.148034572601318,
"eval_runtime": 0.3881,
"eval_samples_per_second": 590.008,
"eval_steps_per_second": 74.717,
"step": 1028
},
{
"epoch": 2.140077821011673,
"grad_norm": 1.367798924446106,
"learning_rate": 8.735408560311285e-05,
"loss": 4.2833,
"step": 1100
},
{
"epoch": 2.3346303501945527,
"grad_norm": 1.3669586181640625,
"learning_rate": 8.519239083441418e-05,
"loss": 4.4121,
"step": 1200
},
{
"epoch": 2.529182879377432,
"grad_norm": 1.2630293369293213,
"learning_rate": 8.303069606571553e-05,
"loss": 4.189,
"step": 1300
},
{
"epoch": 2.7237354085603114,
"grad_norm": 1.7325519323349,
"learning_rate": 8.086900129701687e-05,
"loss": 4.2001,
"step": 1400
},
{
"epoch": 2.9182879377431905,
"grad_norm": 1.5118368864059448,
"learning_rate": 7.870730652831821e-05,
"loss": 4.1732,
"step": 1500
},
{
"epoch": 3.0,
"eval_loss": 3.9841933250427246,
"eval_runtime": 0.3918,
"eval_samples_per_second": 584.494,
"eval_steps_per_second": 74.019,
"step": 1542
},
{
"epoch": 3.11284046692607,
"grad_norm": 1.5012787580490112,
"learning_rate": 7.654561175961954e-05,
"loss": 4.236,
"step": 1600
},
{
"epoch": 3.307392996108949,
"grad_norm": 1.6979159116744995,
"learning_rate": 7.438391699092088e-05,
"loss": 4.0165,
"step": 1700
},
{
"epoch": 3.501945525291829,
"grad_norm": 2.076409339904785,
"learning_rate": 7.222222222222222e-05,
"loss": 4.0474,
"step": 1800
},
{
"epoch": 3.6964980544747084,
"grad_norm": 1.3077492713928223,
"learning_rate": 7.006052745352356e-05,
"loss": 3.9559,
"step": 1900
},
{
"epoch": 3.8910505836575875,
"grad_norm": 1.3804798126220703,
"learning_rate": 6.78988326848249e-05,
"loss": 4.0048,
"step": 2000
},
{
"epoch": 4.0,
"eval_loss": 3.8890349864959717,
"eval_runtime": 0.3928,
"eval_samples_per_second": 583.009,
"eval_steps_per_second": 73.831,
"step": 2056
},
{
"epoch": 4.085603112840467,
"grad_norm": 1.4003207683563232,
"learning_rate": 6.573713791612624e-05,
"loss": 3.9389,
"step": 2100
},
{
"epoch": 4.280155642023346,
"grad_norm": 1.5252618789672852,
"learning_rate": 6.357544314742758e-05,
"loss": 3.7979,
"step": 2200
},
{
"epoch": 4.474708171206226,
"grad_norm": 1.5208264589309692,
"learning_rate": 6.141374837872893e-05,
"loss": 3.9687,
"step": 2300
},
{
"epoch": 4.669260700389105,
"grad_norm": 1.471234679222107,
"learning_rate": 5.925205361003027e-05,
"loss": 3.8509,
"step": 2400
},
{
"epoch": 4.863813229571985,
"grad_norm": 1.653549313545227,
"learning_rate": 5.7090358841331606e-05,
"loss": 3.8276,
"step": 2500
},
{
"epoch": 5.0,
"eval_loss": 3.829970359802246,
"eval_runtime": 0.3885,
"eval_samples_per_second": 589.446,
"eval_steps_per_second": 74.646,
"step": 2570
},
{
"epoch": 5.058365758754864,
"grad_norm": 1.6126755475997925,
"learning_rate": 5.492866407263295e-05,
"loss": 3.9095,
"step": 2600
},
{
"epoch": 5.252918287937743,
"grad_norm": 1.7814581394195557,
"learning_rate": 5.2766969303934287e-05,
"loss": 3.7747,
"step": 2700
},
{
"epoch": 5.447470817120623,
"grad_norm": 1.5165146589279175,
"learning_rate": 5.0605274535235624e-05,
"loss": 3.6927,
"step": 2800
},
{
"epoch": 5.642023346303502,
"grad_norm": 1.8930598497390747,
"learning_rate": 4.844357976653697e-05,
"loss": 3.7231,
"step": 2900
},
{
"epoch": 5.836575875486381,
"grad_norm": 1.8033453226089478,
"learning_rate": 4.628188499783831e-05,
"loss": 3.7334,
"step": 3000
},
{
"epoch": 6.0,
"eval_loss": 3.7915844917297363,
"eval_runtime": 0.3882,
"eval_samples_per_second": 589.903,
"eval_steps_per_second": 74.704,
"step": 3084
}
],
"logging_steps": 100,
"max_steps": 5140,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 462648759828480.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}