| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 78, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.01282051282051282, |
| "grad_norm": 23.535282117342813, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": 0.9389, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.02564102564102564, |
| "grad_norm": 30.95165111144086, |
| "learning_rate": 1.3333333333333333e-05, |
| "loss": 1.1153, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.038461538461538464, |
| "grad_norm": 21.427477034083225, |
| "learning_rate": 2e-05, |
| "loss": 0.908, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.05128205128205128, |
| "grad_norm": 17.928895689942703, |
| "learning_rate": 1.9991228300988586e-05, |
| "loss": 0.7206, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.0641025641025641, |
| "grad_norm": 20.46581420059138, |
| "learning_rate": 1.9964928592495046e-05, |
| "loss": 0.8048, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.07692307692307693, |
| "grad_norm": 9.533694744269097, |
| "learning_rate": 1.9921147013144782e-05, |
| "loss": 0.6909, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.08974358974358974, |
| "grad_norm": 7.154372229051495, |
| "learning_rate": 1.985996037070505e-05, |
| "loss": 0.7208, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.10256410256410256, |
| "grad_norm": 6.050595599777209, |
| "learning_rate": 1.9781476007338058e-05, |
| "loss": 0.5383, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.11538461538461539, |
| "grad_norm": 4.81529735005149, |
| "learning_rate": 1.9685831611286312e-05, |
| "loss": 0.5185, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.1282051282051282, |
| "grad_norm": 5.904610892485658, |
| "learning_rate": 1.9573194975320672e-05, |
| "loss": 0.5649, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.14102564102564102, |
| "grad_norm": 5.012548893301235, |
| "learning_rate": 1.944376370237481e-05, |
| "loss": 0.6025, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.15384615384615385, |
| "grad_norm": 3.4664366084608105, |
| "learning_rate": 1.9297764858882516e-05, |
| "loss": 0.4299, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.16666666666666666, |
| "grad_norm": 4.713790985225752, |
| "learning_rate": 1.913545457642601e-05, |
| "loss": 0.6949, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.1794871794871795, |
| "grad_norm": 4.777031980778873, |
| "learning_rate": 1.895711760239413e-05, |
| "loss": 0.5172, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.19230769230769232, |
| "grad_norm": 3.3607607535520345, |
| "learning_rate": 1.8763066800438638e-05, |
| "loss": 0.4786, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.20512820512820512, |
| "grad_norm": 4.861766646485932, |
| "learning_rate": 1.855364260160507e-05, |
| "loss": 0.4538, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.21794871794871795, |
| "grad_norm": 4.173446943768888, |
| "learning_rate": 1.8329212407100996e-05, |
| "loss": 0.6317, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.23076923076923078, |
| "grad_norm": 3.857794925631418, |
| "learning_rate": 1.8090169943749477e-05, |
| "loss": 0.5543, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.24358974358974358, |
| "grad_norm": 3.589848386643151, |
| "learning_rate": 1.78369345732584e-05, |
| "loss": 0.5853, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.2564102564102564, |
| "grad_norm": 5.135379268595713, |
| "learning_rate": 1.7569950556517566e-05, |
| "loss": 0.5254, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.2692307692307692, |
| "grad_norm": 4.251176222586967, |
| "learning_rate": 1.7289686274214116e-05, |
| "loss": 0.5944, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.28205128205128205, |
| "grad_norm": 3.870818396954739, |
| "learning_rate": 1.6996633405133656e-05, |
| "loss": 0.5656, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.2948717948717949, |
| "grad_norm": 3.342745967846155, |
| "learning_rate": 1.6691306063588583e-05, |
| "loss": 0.5327, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.3076923076923077, |
| "grad_norm": 4.502112335496065, |
| "learning_rate": 1.63742398974869e-05, |
| "loss": 0.5252, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.32051282051282054, |
| "grad_norm": 4.61897270876649, |
| "learning_rate": 1.6045991148623752e-05, |
| "loss": 0.7316, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.3333333333333333, |
| "grad_norm": 4.630378520341254, |
| "learning_rate": 1.570713567684432e-05, |
| "loss": 0.5961, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.34615384615384615, |
| "grad_norm": 3.4435955161971847, |
| "learning_rate": 1.5358267949789968e-05, |
| "loss": 0.3898, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.358974358974359, |
| "grad_norm": 3.6296714817244355, |
| "learning_rate": 1.5000000000000002e-05, |
| "loss": 0.5917, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.3717948717948718, |
| "grad_norm": 3.553076230658495, |
| "learning_rate": 1.463296035119862e-05, |
| "loss": 0.4686, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.38461538461538464, |
| "grad_norm": 3.3205165192761865, |
| "learning_rate": 1.4257792915650728e-05, |
| "loss": 0.5327, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.3974358974358974, |
| "grad_norm": 4.258971760384096, |
| "learning_rate": 1.3875155864521031e-05, |
| "loss": 0.5969, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.41025641025641024, |
| "grad_norm": 4.160699064174649, |
| "learning_rate": 1.3485720473218153e-05, |
| "loss": 0.5726, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.4230769230769231, |
| "grad_norm": 3.8858738820549443, |
| "learning_rate": 1.3090169943749475e-05, |
| "loss": 0.481, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.4358974358974359, |
| "grad_norm": 3.5923086165234364, |
| "learning_rate": 1.2689198206152657e-05, |
| "loss": 0.4831, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.44871794871794873, |
| "grad_norm": 3.040207060233123, |
| "learning_rate": 1.2283508701106559e-05, |
| "loss": 0.4379, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.46153846153846156, |
| "grad_norm": 4.835462956087315, |
| "learning_rate": 1.187381314585725e-05, |
| "loss": 0.5842, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.47435897435897434, |
| "grad_norm": 3.430545038253428, |
| "learning_rate": 1.1460830285624119e-05, |
| "loss": 0.4721, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.48717948717948717, |
| "grad_norm": 3.550867514464785, |
| "learning_rate": 1.1045284632676535e-05, |
| "loss": 0.5327, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.5, |
| "grad_norm": 4.25197469639109, |
| "learning_rate": 1.0627905195293135e-05, |
| "loss": 0.5851, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.5128205128205128, |
| "grad_norm": 5.301567871121325, |
| "learning_rate": 1.0209424198833571e-05, |
| "loss": 0.6567, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.5256410256410257, |
| "grad_norm": 4.76666007720682, |
| "learning_rate": 9.790575801166432e-06, |
| "loss": 0.5344, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.5384615384615384, |
| "grad_norm": 4.628624999659393, |
| "learning_rate": 9.372094804706867e-06, |
| "loss": 0.5901, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.5512820512820513, |
| "grad_norm": 2.7785526596724655, |
| "learning_rate": 8.954715367323468e-06, |
| "loss": 0.4797, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.5641025641025641, |
| "grad_norm": 3.793194400617578, |
| "learning_rate": 8.539169714375885e-06, |
| "loss": 0.5728, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.5769230769230769, |
| "grad_norm": 3.016128749634024, |
| "learning_rate": 8.126186854142752e-06, |
| "loss": 0.4808, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.5897435897435898, |
| "grad_norm": 3.992077099503845, |
| "learning_rate": 7.716491298893443e-06, |
| "loss": 0.5197, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.6025641025641025, |
| "grad_norm": 4.490369000424712, |
| "learning_rate": 7.310801793847344e-06, |
| "loss": 0.4184, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.6153846153846154, |
| "grad_norm": 2.33159901322508, |
| "learning_rate": 6.909830056250527e-06, |
| "loss": 0.3481, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.6282051282051282, |
| "grad_norm": 2.550092727870004, |
| "learning_rate": 6.5142795267818505e-06, |
| "loss": 0.3653, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.6410256410256411, |
| "grad_norm": 2.6572708529729487, |
| "learning_rate": 6.124844135478971e-06, |
| "loss": 0.4036, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.6538461538461539, |
| "grad_norm": 4.037851736171114, |
| "learning_rate": 5.742207084349274e-06, |
| "loss": 0.5075, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.6666666666666666, |
| "grad_norm": 2.9926759956064877, |
| "learning_rate": 5.367039648801386e-06, |
| "loss": 0.4551, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.6794871794871795, |
| "grad_norm": 2.8400263733822477, |
| "learning_rate": 5.000000000000003e-06, |
| "loss": 0.515, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.6923076923076923, |
| "grad_norm": 3.8124291528622973, |
| "learning_rate": 4.641732050210032e-06, |
| "loss": 0.5086, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.7051282051282052, |
| "grad_norm": 2.841356103484977, |
| "learning_rate": 4.292864323155684e-06, |
| "loss": 0.4376, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.717948717948718, |
| "grad_norm": 3.142335016274588, |
| "learning_rate": 3.954008851376252e-06, |
| "loss": 0.4039, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.7307692307692307, |
| "grad_norm": 2.852087372917465, |
| "learning_rate": 3.625760102513103e-06, |
| "loss": 0.5008, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.7435897435897436, |
| "grad_norm": 3.324712819809194, |
| "learning_rate": 3.308693936411421e-06, |
| "loss": 0.4739, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.7564102564102564, |
| "grad_norm": 2.9649459892785535, |
| "learning_rate": 3.003366594866345e-06, |
| "loss": 0.5892, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.7692307692307693, |
| "grad_norm": 3.044018836047704, |
| "learning_rate": 2.7103137257858867e-06, |
| "loss": 0.4852, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.782051282051282, |
| "grad_norm": 4.050018523402415, |
| "learning_rate": 2.4300494434824373e-06, |
| "loss": 0.591, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.7948717948717948, |
| "grad_norm": 2.3281157099809398, |
| "learning_rate": 2.163065426741603e-06, |
| "loss": 0.4416, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.8076923076923077, |
| "grad_norm": 2.9262958648242883, |
| "learning_rate": 1.9098300562505266e-06, |
| "loss": 0.4393, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.8205128205128205, |
| "grad_norm": 3.884834370223603, |
| "learning_rate": 1.6707875928990059e-06, |
| "loss": 0.5305, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.8333333333333334, |
| "grad_norm": 3.5229544410053544, |
| "learning_rate": 1.446357398394934e-06, |
| "loss": 0.5226, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.8461538461538461, |
| "grad_norm": 3.550541933958789, |
| "learning_rate": 1.2369331995613664e-06, |
| "loss": 0.6602, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.8589743589743589, |
| "grad_norm": 3.077396885216065, |
| "learning_rate": 1.042882397605871e-06, |
| "loss": 0.4248, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.8717948717948718, |
| "grad_norm": 3.1902818565830784, |
| "learning_rate": 8.645454235739903e-07, |
| "loss": 0.4756, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.8846153846153846, |
| "grad_norm": 2.9790818845625346, |
| "learning_rate": 7.022351411174866e-07, |
| "loss": 0.4006, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.8974358974358975, |
| "grad_norm": 4.103022950661447, |
| "learning_rate": 5.562362976251901e-07, |
| "loss": 0.5541, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.9102564102564102, |
| "grad_norm": 3.414968428161163, |
| "learning_rate": 4.268050246793276e-07, |
| "loss": 0.5879, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.9230769230769231, |
| "grad_norm": 2.992490474364374, |
| "learning_rate": 3.1416838871368925e-07, |
| "loss": 0.435, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.9358974358974359, |
| "grad_norm": 3.207632291423965, |
| "learning_rate": 2.1852399266194312e-07, |
| "loss": 0.4987, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.9487179487179487, |
| "grad_norm": 2.9139113123937044, |
| "learning_rate": 1.400396292949513e-07, |
| "loss": 0.4156, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.9615384615384616, |
| "grad_norm": 2.536721514944869, |
| "learning_rate": 7.885298685522235e-08, |
| "loss": 0.4533, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.9743589743589743, |
| "grad_norm": 2.586955803538133, |
| "learning_rate": 3.50714075049563e-08, |
| "loss": 0.4356, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.9871794871794872, |
| "grad_norm": 2.8019180608719463, |
| "learning_rate": 8.771699011416169e-09, |
| "loss": 0.3865, |
| "step": 77 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 2.8737311595523893, |
| "learning_rate": 0.0, |
| "loss": 0.4296, |
| "step": 78 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 78, |
| "total_flos": 112793145835520.0, |
| "train_loss": 0.5409963910396283, |
| "train_runtime": 7617.3016, |
| "train_samples_per_second": 1.147, |
| "train_steps_per_second": 0.01 |
| } |
| ], |
| "logging_steps": 1.0, |
| "max_steps": 78, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 10, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 112793145835520.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|