EleanorZzz commited on
Commit
f2f6845
·
verified ·
1 Parent(s): 7afb401

End of training

Browse files
Files changed (5) hide show
  1. README.md +1 -1
  2. all_results.json +6 -6
  3. train_results.json +6 -6
  4. trainer_state.json +242 -81
  5. training_loss.png +0 -0
README.md CHANGED
@@ -16,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # persuasion_simulation_tulu3_8b_sft_sft_w_promp_10epochs
18
 
19
- This model is a fine-tuned version of [allenai/Llama-3.1-Tulu-3-8B-SFT](https://huggingface.co/allenai/Llama-3.1-Tulu-3-8B-SFT) on an unknown dataset.
20
 
21
  ## Model description
22
 
 
16
 
17
  # persuasion_simulation_tulu3_8b_sft_sft_w_promp_10epochs
18
 
19
+ This model is a fine-tuned version of [allenai/Llama-3.1-Tulu-3-8B-SFT](https://huggingface.co/allenai/Llama-3.1-Tulu-3-8B-SFT) on the persuasion_simulation dataset.
20
 
21
  ## Model description
22
 
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 4.908108108108108,
3
- "total_flos": 6842883440640.0,
4
- "train_loss": 0.6322610969128816,
5
- "train_runtime": 1920.4432,
6
- "train_samples_per_second": 1.924,
7
- "train_steps_per_second": 0.12
8
  }
 
1
  {
2
+ "epoch": 9.8,
3
+ "total_flos": 13673089597440.0,
4
+ "train_loss": 0.34487654119729994,
5
+ "train_runtime": 4380.6043,
6
+ "train_samples_per_second": 1.687,
7
+ "train_steps_per_second": 0.105
8
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "epoch": 4.908108108108108,
3
- "total_flos": 6842883440640.0,
4
- "train_loss": 0.6322610969128816,
5
- "train_runtime": 1920.4432,
6
- "train_samples_per_second": 1.924,
7
- "train_steps_per_second": 0.12
8
  }
 
1
  {
2
+ "epoch": 9.8,
3
+ "total_flos": 13673089597440.0,
4
+ "train_loss": 0.34487654119729994,
5
+ "train_runtime": 4380.6043,
6
+ "train_samples_per_second": 1.687,
7
+ "train_steps_per_second": 0.105
8
  }
trainer_state.json CHANGED
@@ -2,188 +2,349 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 4.908108108108108,
6
  "eval_steps": 500,
7
- "global_step": 230,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 0.21621621621621623,
14
- "grad_norm": 7.521789519465365,
15
- "learning_rate": 4.347826086956522e-06,
16
- "loss": 2.5607,
17
  "step": 10
18
  },
19
  {
20
  "epoch": 0.43243243243243246,
21
- "grad_norm": 3.213619656849043,
22
- "learning_rate": 8.695652173913044e-06,
23
- "loss": 1.1719,
24
  "step": 20
25
  },
26
  {
27
  "epoch": 0.6486486486486487,
28
- "grad_norm": 1.9437444652365807,
29
- "learning_rate": 9.971810547786794e-06,
30
- "loss": 1.0216,
31
  "step": 30
32
  },
33
  {
34
  "epoch": 0.8648648648648649,
35
- "grad_norm": 1.869691703158474,
36
- "learning_rate": 9.834504404631032e-06,
37
- "loss": 0.9938,
38
  "step": 40
39
  },
40
  {
41
  "epoch": 1.0648648648648649,
42
- "grad_norm": 1.8384505363173806,
43
- "learning_rate": 9.586056507527266e-06,
44
- "loss": 0.9102,
45
  "step": 50
46
  },
47
  {
48
  "epoch": 1.281081081081081,
49
- "grad_norm": 1.7699518643234728,
50
- "learning_rate": 9.232178493644006e-06,
51
- "loss": 0.7888,
52
  "step": 60
53
  },
54
  {
55
  "epoch": 1.4972972972972973,
56
- "grad_norm": 1.9014403364367916,
57
- "learning_rate": 8.781005762156593e-06,
58
- "loss": 0.7679,
59
  "step": 70
60
  },
61
  {
62
  "epoch": 1.7135135135135136,
63
- "grad_norm": 1.6683999198252772,
64
- "learning_rate": 8.24291044731378e-06,
65
- "loss": 0.7813,
66
  "step": 80
67
  },
68
  {
69
  "epoch": 1.9297297297297298,
70
- "grad_norm": 1.758096343286875,
71
- "learning_rate": 7.630262970585355e-06,
72
- "loss": 0.7773,
73
  "step": 90
74
  },
75
  {
76
  "epoch": 2.1297297297297297,
77
- "grad_norm": 2.2320248203710222,
78
- "learning_rate": 6.957147653634198e-06,
79
- "loss": 0.6619,
80
  "step": 100
81
  },
82
  {
83
  "epoch": 2.345945945945946,
84
- "grad_norm": 2.1598362259959885,
85
- "learning_rate": 6.2390389299645e-06,
86
- "loss": 0.5109,
87
  "step": 110
88
  },
89
  {
90
  "epoch": 2.562162162162162,
91
- "grad_norm": 1.9698378181001832,
92
- "learning_rate": 5.492445598905843e-06,
93
- "loss": 0.5128,
94
  "step": 120
95
  },
96
  {
97
  "epoch": 2.7783783783783784,
98
- "grad_norm": 2.207004805490676,
99
- "learning_rate": 4.7345313002762545e-06,
100
- "loss": 0.5028,
101
  "step": 130
102
  },
103
  {
104
  "epoch": 2.9945945945945946,
105
- "grad_norm": 2.0954739967997753,
106
- "learning_rate": 3.982719934736832e-06,
107
- "loss": 0.498,
108
  "step": 140
109
  },
110
  {
111
  "epoch": 3.1945945945945944,
112
- "grad_norm": 4.1425174803645275,
113
- "learning_rate": 3.2542951009381584e-06,
114
- "loss": 0.3427,
115
  "step": 150
116
  },
117
  {
118
  "epoch": 3.410810810810811,
119
- "grad_norm": 2.9187774194524816,
120
- "learning_rate": 2.566002758108256e-06,
121
- "loss": 0.2867,
122
  "step": 160
123
  },
124
  {
125
  "epoch": 3.627027027027027,
126
- "grad_norm": 2.5178609130134593,
127
- "learning_rate": 1.933666248581418e-06,
128
- "loss": 0.2728,
129
  "step": 170
130
  },
131
  {
132
  "epoch": 3.8432432432432435,
133
- "grad_norm": 2.546624967772346,
134
- "learning_rate": 1.3718225306210049e-06,
135
- "loss": 0.2752,
136
  "step": 180
137
  },
138
  {
139
  "epoch": 4.043243243243243,
140
- "grad_norm": 3.0183024513477084,
141
- "learning_rate": 8.933879842801558e-07,
142
- "loss": 0.2473,
143
  "step": 190
144
  },
145
  {
146
  "epoch": 4.2594594594594595,
147
- "grad_norm": 3.05168507955595,
148
- "learning_rate": 5.0936147318152e-07,
149
- "loss": 0.1699,
150
  "step": 200
151
  },
152
  {
153
  "epoch": 4.475675675675676,
154
- "grad_norm": 2.293078858586022,
155
- "learning_rate": 2.2857148861060552e-07,
156
- "loss": 0.1666,
157
  "step": 210
158
  },
159
  {
160
  "epoch": 4.691891891891892,
161
- "grad_norm": 2.118854972987958,
162
- "learning_rate": 5.747318889684883e-08,
163
- "loss": 0.1614,
164
  "step": 220
165
  },
166
  {
167
  "epoch": 4.908108108108108,
168
- "grad_norm": 2.0202763257750007,
169
- "learning_rate": 0.0,
170
- "loss": 0.1594,
171
  "step": 230
172
  },
173
  {
174
- "epoch": 4.908108108108108,
175
- "step": 230,
176
- "total_flos": 6842883440640.0,
177
- "train_loss": 0.6322610969128816,
178
- "train_runtime": 1920.4432,
179
- "train_samples_per_second": 1.924,
180
- "train_steps_per_second": 0.12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
181
  }
182
  ],
183
  "logging_steps": 10,
184
- "max_steps": 230,
185
  "num_input_tokens_seen": 0,
186
- "num_train_epochs": 5,
187
  "save_steps": 500,
188
  "stateful_callbacks": {
189
  "TrainerControl": {
@@ -197,7 +358,7 @@
197
  "attributes": {}
198
  }
199
  },
200
- "total_flos": 6842883440640.0,
201
  "train_batch_size": 1,
202
  "trial_name": null,
203
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 9.8,
6
  "eval_steps": 500,
7
+ "global_step": 460,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 0.21621621621621623,
14
+ "grad_norm": 7.928413532238152,
15
+ "learning_rate": 2.173913043478261e-06,
16
+ "loss": 2.1422,
17
  "step": 10
18
  },
19
  {
20
  "epoch": 0.43243243243243246,
21
+ "grad_norm": 6.01372163805319,
22
+ "learning_rate": 4.347826086956522e-06,
23
+ "loss": 1.3261,
24
  "step": 20
25
  },
26
  {
27
  "epoch": 0.6486486486486487,
28
+ "grad_norm": 2.143165243967343,
29
+ "learning_rate": 6.521739130434783e-06,
30
+ "loss": 1.0565,
31
  "step": 30
32
  },
33
  {
34
  "epoch": 0.8648648648648649,
35
+ "grad_norm": 2.002981301105067,
36
+ "learning_rate": 8.695652173913044e-06,
37
+ "loss": 1.0008,
38
  "step": 40
39
  },
40
  {
41
  "epoch": 1.0648648648648649,
42
+ "grad_norm": 2.003768096340962,
43
+ "learning_rate": 9.997696831512027e-06,
44
+ "loss": 0.9263,
45
  "step": 50
46
  },
47
  {
48
  "epoch": 1.281081081081081,
49
+ "grad_norm": 1.9628397812309482,
50
+ "learning_rate": 9.971810547786794e-06,
51
+ "loss": 0.832,
52
  "step": 60
53
  },
54
  {
55
  "epoch": 1.4972972972972973,
56
+ "grad_norm": 1.9670767045650799,
57
+ "learning_rate": 9.917308508168712e-06,
58
+ "loss": 0.8092,
59
  "step": 70
60
  },
61
  {
62
  "epoch": 1.7135135135135136,
63
+ "grad_norm": 1.8258318637551825,
64
+ "learning_rate": 9.834504404631032e-06,
65
+ "loss": 0.8275,
66
  "step": 80
67
  },
68
  {
69
  "epoch": 1.9297297297297298,
70
+ "grad_norm": 1.853270348883972,
71
+ "learning_rate": 9.72387482452377e-06,
72
+ "loss": 0.8257,
73
  "step": 90
74
  },
75
  {
76
  "epoch": 2.1297297297297297,
77
+ "grad_norm": 2.181312155139493,
78
+ "learning_rate": 9.586056507527266e-06,
79
+ "loss": 0.7141,
80
  "step": 100
81
  },
82
  {
83
  "epoch": 2.345945945945946,
84
+ "grad_norm": 2.2635996413273576,
85
+ "learning_rate": 9.421842680832862e-06,
86
+ "loss": 0.5572,
87
  "step": 110
88
  },
89
  {
90
  "epoch": 2.562162162162162,
91
+ "grad_norm": 2.0267513780569555,
92
+ "learning_rate": 9.232178493644006e-06,
93
+ "loss": 0.5536,
94
  "step": 120
95
  },
96
  {
97
  "epoch": 2.7783783783783784,
98
+ "grad_norm": 2.2717141213154233,
99
+ "learning_rate": 9.018155577274891e-06,
100
+ "loss": 0.5567,
101
  "step": 130
102
  },
103
  {
104
  "epoch": 2.9945945945945946,
105
+ "grad_norm": 2.161425393067172,
106
+ "learning_rate": 8.781005762156593e-06,
107
+ "loss": 0.554,
108
  "step": 140
109
  },
110
  {
111
  "epoch": 3.1945945945945944,
112
+ "grad_norm": 2.6379003441929583,
113
+ "learning_rate": 8.522093987913063e-06,
114
+ "loss": 0.3631,
115
  "step": 150
116
  },
117
  {
118
  "epoch": 3.410810810810811,
119
+ "grad_norm": 2.248635452478229,
120
+ "learning_rate": 8.24291044731378e-06,
121
+ "loss": 0.302,
122
  "step": 160
123
  },
124
  {
125
  "epoch": 3.627027027027027,
126
+ "grad_norm": 2.3569737350184297,
127
+ "learning_rate": 7.94506200931932e-06,
128
+ "loss": 0.2989,
129
  "step": 170
130
  },
131
  {
132
  "epoch": 3.8432432432432435,
133
+ "grad_norm": 2.622041501021548,
134
+ "learning_rate": 7.630262970585355e-06,
135
+ "loss": 0.3074,
136
  "step": 180
137
  },
138
  {
139
  "epoch": 4.043243243243243,
140
+ "grad_norm": 2.274362887049295,
141
+ "learning_rate": 7.300325188655762e-06,
142
+ "loss": 0.2728,
143
  "step": 190
144
  },
145
  {
146
  "epoch": 4.2594594594594595,
147
+ "grad_norm": 2.1104698967330995,
148
+ "learning_rate": 6.957147653634198e-06,
149
+ "loss": 0.1462,
150
  "step": 200
151
  },
152
  {
153
  "epoch": 4.475675675675676,
154
+ "grad_norm": 2.1899462721461993,
155
+ "learning_rate": 6.6027055583554865e-06,
156
+ "loss": 0.1454,
157
  "step": 210
158
  },
159
  {
160
  "epoch": 4.691891891891892,
161
+ "grad_norm": 2.6349171406091316,
162
+ "learning_rate": 6.2390389299645e-06,
163
+ "loss": 0.1545,
164
  "step": 220
165
  },
166
  {
167
  "epoch": 4.908108108108108,
168
+ "grad_norm": 2.1896943214948026,
169
+ "learning_rate": 5.8682408883346535e-06,
170
+ "loss": 0.147,
171
  "step": 230
172
  },
173
  {
174
+ "epoch": 5.108108108108108,
175
+ "grad_norm": 1.8580244939014094,
176
+ "learning_rate": 5.492445598905843e-06,
177
+ "loss": 0.1175,
178
+ "step": 240
179
+ },
180
+ {
181
+ "epoch": 5.324324324324325,
182
+ "grad_norm": 1.7205360918150199,
183
+ "learning_rate": 5.113815989280528e-06,
184
+ "loss": 0.0828,
185
+ "step": 250
186
+ },
187
+ {
188
+ "epoch": 5.54054054054054,
189
+ "grad_norm": 2.041469867596687,
190
+ "learning_rate": 4.7345313002762545e-06,
191
+ "loss": 0.0814,
192
+ "step": 260
193
+ },
194
+ {
195
+ "epoch": 5.756756756756757,
196
+ "grad_norm": 1.731451961015919,
197
+ "learning_rate": 4.356774543085845e-06,
198
+ "loss": 0.083,
199
+ "step": 270
200
+ },
201
+ {
202
+ "epoch": 5.972972972972973,
203
+ "grad_norm": 1.640206077124201,
204
+ "learning_rate": 3.982719934736832e-06,
205
+ "loss": 0.0863,
206
+ "step": 280
207
+ },
208
+ {
209
+ "epoch": 6.172972972972973,
210
+ "grad_norm": 1.6295826391354875,
211
+ "learning_rate": 3.6145203841665577e-06,
212
+ "loss": 0.0574,
213
+ "step": 290
214
+ },
215
+ {
216
+ "epoch": 6.389189189189189,
217
+ "grad_norm": 1.440211343971061,
218
+ "learning_rate": 3.2542951009381584e-06,
219
+ "loss": 0.0536,
220
+ "step": 300
221
+ },
222
+ {
223
+ "epoch": 6.605405405405405,
224
+ "grad_norm": 1.4014358014924952,
225
+ "learning_rate": 2.9041173979166813e-06,
226
+ "loss": 0.0543,
227
+ "step": 310
228
+ },
229
+ {
230
+ "epoch": 6.821621621621622,
231
+ "grad_norm": 1.3463804839310973,
232
+ "learning_rate": 2.566002758108256e-06,
233
+ "loss": 0.0541,
234
+ "step": 320
235
+ },
236
+ {
237
+ "epoch": 7.021621621621621,
238
+ "grad_norm": 2.579253801667801,
239
+ "learning_rate": 2.241897234344864e-06,
240
+ "loss": 0.0507,
241
+ "step": 330
242
+ },
243
+ {
244
+ "epoch": 7.237837837837838,
245
+ "grad_norm": 1.1381118685241185,
246
+ "learning_rate": 1.933666248581418e-06,
247
+ "loss": 0.0337,
248
+ "step": 340
249
+ },
250
+ {
251
+ "epoch": 7.454054054054054,
252
+ "grad_norm": 1.1614358836368845,
253
+ "learning_rate": 1.6430838552720168e-06,
254
+ "loss": 0.0347,
255
+ "step": 350
256
+ },
257
+ {
258
+ "epoch": 7.6702702702702705,
259
+ "grad_norm": 0.9439702681790234,
260
+ "learning_rate": 1.3718225306210049e-06,
261
+ "loss": 0.0346,
262
+ "step": 360
263
+ },
264
+ {
265
+ "epoch": 7.886486486486486,
266
+ "grad_norm": 0.9350344061498097,
267
+ "learning_rate": 1.1214435464779006e-06,
268
+ "loss": 0.0345,
269
+ "step": 370
270
+ },
271
+ {
272
+ "epoch": 8.086486486486486,
273
+ "grad_norm": 0.6158081847960301,
274
+ "learning_rate": 8.933879842801558e-07,
275
+ "loss": 0.0292,
276
+ "step": 380
277
+ },
278
+ {
279
+ "epoch": 8.302702702702703,
280
+ "grad_norm": 0.8470578217546869,
281
+ "learning_rate": 6.889684407639324e-07,
282
+ "loss": 0.0219,
283
+ "step": 390
284
+ },
285
+ {
286
+ "epoch": 8.518918918918919,
287
+ "grad_norm": 0.7060993947541574,
288
+ "learning_rate": 5.0936147318152e-07,
289
+ "loss": 0.022,
290
+ "step": 400
291
+ },
292
+ {
293
+ "epoch": 8.735135135135135,
294
+ "grad_norm": 0.5972472978430013,
295
+ "learning_rate": 3.55600827507665e-07,
296
+ "loss": 0.0216,
297
+ "step": 410
298
+ },
299
+ {
300
+ "epoch": 8.951351351351352,
301
+ "grad_norm": 0.6814447950625864,
302
+ "learning_rate": 2.2857148861060552e-07,
303
+ "loss": 0.0217,
304
+ "step": 420
305
+ },
306
+ {
307
+ "epoch": 9.151351351351352,
308
+ "grad_norm": 0.45326924305283917,
309
+ "learning_rate": 1.2900458663260506e-07,
310
+ "loss": 0.0186,
311
+ "step": 430
312
+ },
313
+ {
314
+ "epoch": 9.367567567567567,
315
+ "grad_norm": 0.3976668377272393,
316
+ "learning_rate": 5.747318889684883e-08,
317
+ "loss": 0.0173,
318
+ "step": 440
319
+ },
320
+ {
321
+ "epoch": 9.583783783783783,
322
+ "grad_norm": 0.47000235405688096,
323
+ "learning_rate": 1.4389001560803917e-08,
324
+ "loss": 0.0176,
325
+ "step": 450
326
+ },
327
+ {
328
+ "epoch": 9.8,
329
+ "grad_norm": 0.4716229816777135,
330
+ "learning_rate": 0.0,
331
+ "loss": 0.0169,
332
+ "step": 460
333
+ },
334
+ {
335
+ "epoch": 9.8,
336
+ "step": 460,
337
+ "total_flos": 13673089597440.0,
338
+ "train_loss": 0.34487654119729994,
339
+ "train_runtime": 4380.6043,
340
+ "train_samples_per_second": 1.687,
341
+ "train_steps_per_second": 0.105
342
  }
343
  ],
344
  "logging_steps": 10,
345
+ "max_steps": 460,
346
  "num_input_tokens_seen": 0,
347
+ "num_train_epochs": 10,
348
  "save_steps": 500,
349
  "stateful_callbacks": {
350
  "TrainerControl": {
 
358
  "attributes": {}
359
  }
360
  },
361
+ "total_flos": 13673089597440.0,
362
  "train_batch_size": 1,
363
  "trial_name": null,
364
  "trial_params": null
training_loss.png CHANGED