File size: 12,864 Bytes
7bc029d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.9865470852017937,
  "eval_steps": 500,
  "global_step": 55,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.017937219730941704,
      "grad_norm": 0.6536183953285217,
      "learning_rate": 4.995922759815339e-05,
      "loss": 0.8371,
      "num_input_tokens_seen": 2097152,
      "step": 1
    },
    {
      "epoch": 0.03587443946188341,
      "grad_norm": 0.517680823802948,
      "learning_rate": 4.9837043383713753e-05,
      "loss": 0.7804,
      "num_input_tokens_seen": 4194304,
      "step": 2
    },
    {
      "epoch": 0.053811659192825115,
      "grad_norm": 0.4423481225967407,
      "learning_rate": 4.963384589619233e-05,
      "loss": 0.7695,
      "num_input_tokens_seen": 6291456,
      "step": 3
    },
    {
      "epoch": 0.07174887892376682,
      "grad_norm": 0.39828750491142273,
      "learning_rate": 4.935029792355834e-05,
      "loss": 0.7419,
      "num_input_tokens_seen": 8388608,
      "step": 4
    },
    {
      "epoch": 0.08968609865470852,
      "grad_norm": 0.31201115250587463,
      "learning_rate": 4.898732434036244e-05,
      "loss": 0.7166,
      "num_input_tokens_seen": 10485760,
      "step": 5
    },
    {
      "epoch": 0.10762331838565023,
      "grad_norm": 0.2536958158016205,
      "learning_rate": 4.854610909098812e-05,
      "loss": 0.7194,
      "num_input_tokens_seen": 12582912,
      "step": 6
    },
    {
      "epoch": 0.12556053811659193,
      "grad_norm": 0.2193588763475418,
      "learning_rate": 4.802809132787125e-05,
      "loss": 0.6975,
      "num_input_tokens_seen": 14680064,
      "step": 7
    },
    {
      "epoch": 0.14349775784753363,
      "grad_norm": 0.18916621804237366,
      "learning_rate": 4.743496071728396e-05,
      "loss": 0.7168,
      "num_input_tokens_seen": 16777216,
      "step": 8
    },
    {
      "epoch": 0.16143497757847533,
      "grad_norm": 0.1561172604560852,
      "learning_rate": 4.6768651927994434e-05,
      "loss": 0.6707,
      "num_input_tokens_seen": 18874368,
      "step": 9
    },
    {
      "epoch": 0.17937219730941703,
      "grad_norm": 0.12857139110565186,
      "learning_rate": 4.6031338320779534e-05,
      "loss": 0.6769,
      "num_input_tokens_seen": 20971520,
      "step": 10
    },
    {
      "epoch": 0.19730941704035873,
      "grad_norm": 0.11340289562940598,
      "learning_rate": 4.522542485937369e-05,
      "loss": 0.6873,
      "num_input_tokens_seen": 23068672,
      "step": 11
    },
    {
      "epoch": 0.21524663677130046,
      "grad_norm": 0.10658581554889679,
      "learning_rate": 4.4353540265977064e-05,
      "loss": 0.6643,
      "num_input_tokens_seen": 25165824,
      "step": 12
    },
    {
      "epoch": 0.23318385650224216,
      "grad_norm": 0.08937722444534302,
      "learning_rate": 4.341852844691012e-05,
      "loss": 0.6849,
      "num_input_tokens_seen": 27262976,
      "step": 13
    },
    {
      "epoch": 0.25112107623318386,
      "grad_norm": 0.07756289094686508,
      "learning_rate": 4.242343921638234e-05,
      "loss": 0.6461,
      "num_input_tokens_seen": 29360128,
      "step": 14
    },
    {
      "epoch": 0.26905829596412556,
      "grad_norm": 0.07581546157598495,
      "learning_rate": 4.137151834863213e-05,
      "loss": 0.6623,
      "num_input_tokens_seen": 31457280,
      "step": 15
    },
    {
      "epoch": 0.28699551569506726,
      "grad_norm": 0.07386067509651184,
      "learning_rate": 4.0266196990885955e-05,
      "loss": 0.6751,
      "num_input_tokens_seen": 33554432,
      "step": 16
    },
    {
      "epoch": 0.30493273542600896,
      "grad_norm": 0.06293580681085587,
      "learning_rate": 3.911108047166924e-05,
      "loss": 0.6472,
      "num_input_tokens_seen": 35651584,
      "step": 17
    },
    {
      "epoch": 0.32286995515695066,
      "grad_norm": 0.06199085712432861,
      "learning_rate": 3.790993654097405e-05,
      "loss": 0.6728,
      "num_input_tokens_seen": 37748736,
      "step": 18
    },
    {
      "epoch": 0.34080717488789236,
      "grad_norm": 0.060734592378139496,
      "learning_rate": 3.6666683080641846e-05,
      "loss": 0.7017,
      "num_input_tokens_seen": 39845888,
      "step": 19
    },
    {
      "epoch": 0.35874439461883406,
      "grad_norm": 0.05623164027929306,
      "learning_rate": 3.5385375325047166e-05,
      "loss": 0.6502,
      "num_input_tokens_seen": 41943040,
      "step": 20
    },
    {
      "epoch": 0.37668161434977576,
      "grad_norm": 0.0574677549302578,
      "learning_rate": 3.4070192633766025e-05,
      "loss": 0.6476,
      "num_input_tokens_seen": 44040192,
      "step": 21
    },
    {
      "epoch": 0.39461883408071746,
      "grad_norm": 0.05185185372829437,
      "learning_rate": 3.272542485937369e-05,
      "loss": 0.6411,
      "num_input_tokens_seen": 46137344,
      "step": 22
    },
    {
      "epoch": 0.4125560538116592,
      "grad_norm": 0.05139186978340149,
      "learning_rate": 3.135545835483718e-05,
      "loss": 0.6428,
      "num_input_tokens_seen": 48234496,
      "step": 23
    },
    {
      "epoch": 0.4304932735426009,
      "grad_norm": 0.050159115344285965,
      "learning_rate": 2.996476166614364e-05,
      "loss": 0.6661,
      "num_input_tokens_seen": 50331648,
      "step": 24
    },
    {
      "epoch": 0.4484304932735426,
      "grad_norm": 0.04851464927196503,
      "learning_rate": 2.8557870956832132e-05,
      "loss": 0.6378,
      "num_input_tokens_seen": 52428800,
      "step": 25
    },
    {
      "epoch": 0.4663677130044843,
      "grad_norm": 0.04896726831793785,
      "learning_rate": 2.7139375211970996e-05,
      "loss": 0.6532,
      "num_input_tokens_seen": 54525952,
      "step": 26
    },
    {
      "epoch": 0.484304932735426,
      "grad_norm": 0.04698600620031357,
      "learning_rate": 2.5713901269842404e-05,
      "loss": 0.6403,
      "num_input_tokens_seen": 56623104,
      "step": 27
    },
    {
      "epoch": 0.5022421524663677,
      "grad_norm": 0.048034097999334335,
      "learning_rate": 2.42860987301576e-05,
      "loss": 0.6248,
      "num_input_tokens_seen": 58720256,
      "step": 28
    },
    {
      "epoch": 0.5201793721973094,
      "grad_norm": 0.044828303158283234,
      "learning_rate": 2.2860624788029013e-05,
      "loss": 0.6583,
      "num_input_tokens_seen": 60817408,
      "step": 29
    },
    {
      "epoch": 0.5381165919282511,
      "grad_norm": 0.04563640430569649,
      "learning_rate": 2.1442129043167874e-05,
      "loss": 0.6579,
      "num_input_tokens_seen": 62914560,
      "step": 30
    },
    {
      "epoch": 0.5560538116591929,
      "grad_norm": 0.044318560510873795,
      "learning_rate": 2.003523833385637e-05,
      "loss": 0.6659,
      "num_input_tokens_seen": 65011712,
      "step": 31
    },
    {
      "epoch": 0.5739910313901345,
      "grad_norm": 0.04331167787313461,
      "learning_rate": 1.8644541645162834e-05,
      "loss": 0.6423,
      "num_input_tokens_seen": 67108864,
      "step": 32
    },
    {
      "epoch": 0.5919282511210763,
      "grad_norm": 0.04475367069244385,
      "learning_rate": 1.7274575140626318e-05,
      "loss": 0.6509,
      "num_input_tokens_seen": 69206016,
      "step": 33
    },
    {
      "epoch": 0.6098654708520179,
      "grad_norm": 0.045547887682914734,
      "learning_rate": 1.5929807366233977e-05,
      "loss": 0.6551,
      "num_input_tokens_seen": 71303168,
      "step": 34
    },
    {
      "epoch": 0.6278026905829597,
      "grad_norm": 0.043985530734062195,
      "learning_rate": 1.4614624674952842e-05,
      "loss": 0.6232,
      "num_input_tokens_seen": 73400320,
      "step": 35
    },
    {
      "epoch": 0.6457399103139013,
      "grad_norm": 0.0414094403386116,
      "learning_rate": 1.3333316919358157e-05,
      "loss": 0.6137,
      "num_input_tokens_seen": 75497472,
      "step": 36
    },
    {
      "epoch": 0.6636771300448431,
      "grad_norm": 0.041019294410943985,
      "learning_rate": 1.2090063459025955e-05,
      "loss": 0.6426,
      "num_input_tokens_seen": 77594624,
      "step": 37
    },
    {
      "epoch": 0.6816143497757847,
      "grad_norm": 0.04383592680096626,
      "learning_rate": 1.0888919528330777e-05,
      "loss": 0.6512,
      "num_input_tokens_seen": 79691776,
      "step": 38
    },
    {
      "epoch": 0.6995515695067265,
      "grad_norm": 0.040539514273405075,
      "learning_rate": 9.733803009114045e-06,
      "loss": 0.6269,
      "num_input_tokens_seen": 81788928,
      "step": 39
    },
    {
      "epoch": 0.7174887892376681,
      "grad_norm": 0.04238974675536156,
      "learning_rate": 8.628481651367876e-06,
      "loss": 0.6201,
      "num_input_tokens_seen": 83886080,
      "step": 40
    },
    {
      "epoch": 0.7354260089686099,
      "grad_norm": 0.04115669056773186,
      "learning_rate": 7.576560783617668e-06,
      "loss": 0.642,
      "num_input_tokens_seen": 85983232,
      "step": 41
    },
    {
      "epoch": 0.7533632286995515,
      "grad_norm": 0.04178008437156677,
      "learning_rate": 6.5814715530898745e-06,
      "loss": 0.648,
      "num_input_tokens_seen": 88080384,
      "step": 42
    },
    {
      "epoch": 0.7713004484304933,
      "grad_norm": 0.04329155012965202,
      "learning_rate": 5.646459734022938e-06,
      "loss": 0.6442,
      "num_input_tokens_seen": 90177536,
      "step": 43
    },
    {
      "epoch": 0.7892376681614349,
      "grad_norm": 0.043740272521972656,
      "learning_rate": 4.7745751406263165e-06,
      "loss": 0.6488,
      "num_input_tokens_seen": 92274688,
      "step": 44
    },
    {
      "epoch": 0.8071748878923767,
      "grad_norm": 0.04263562709093094,
      "learning_rate": 3.968661679220468e-06,
      "loss": 0.65,
      "num_input_tokens_seen": 94371840,
      "step": 45
    },
    {
      "epoch": 0.8251121076233184,
      "grad_norm": 0.041693028062582016,
      "learning_rate": 3.2313480720055745e-06,
      "loss": 0.6584,
      "num_input_tokens_seen": 96468992,
      "step": 46
    },
    {
      "epoch": 0.8430493273542601,
      "grad_norm": 0.04151754826307297,
      "learning_rate": 2.565039282716045e-06,
      "loss": 0.6392,
      "num_input_tokens_seen": 98566144,
      "step": 47
    },
    {
      "epoch": 0.8609865470852018,
      "grad_norm": 0.04260968416929245,
      "learning_rate": 1.97190867212875e-06,
      "loss": 0.6524,
      "num_input_tokens_seen": 100663296,
      "step": 48
    },
    {
      "epoch": 0.8789237668161435,
      "grad_norm": 0.04022514820098877,
      "learning_rate": 1.4538909090118846e-06,
      "loss": 0.6276,
      "num_input_tokens_seen": 102760448,
      "step": 49
    },
    {
      "epoch": 0.8968609865470852,
      "grad_norm": 0.039072513580322266,
      "learning_rate": 1.0126756596375686e-06,
      "loss": 0.6282,
      "num_input_tokens_seen": 104857600,
      "step": 50
    },
    {
      "epoch": 0.9147982062780269,
      "grad_norm": 0.03952722251415253,
      "learning_rate": 6.497020764416633e-07,
      "loss": 0.6344,
      "num_input_tokens_seen": 106954752,
      "step": 51
    },
    {
      "epoch": 0.9327354260089686,
      "grad_norm": 0.04045777767896652,
      "learning_rate": 3.6615410380767544e-07,
      "loss": 0.6464,
      "num_input_tokens_seen": 109051904,
      "step": 52
    },
    {
      "epoch": 0.9506726457399103,
      "grad_norm": 0.03984501212835312,
      "learning_rate": 1.6295661628624447e-07,
      "loss": 0.6253,
      "num_input_tokens_seen": 111149056,
      "step": 53
    },
    {
      "epoch": 0.968609865470852,
      "grad_norm": 0.040761884301900864,
      "learning_rate": 4.07724018466088e-08,
      "loss": 0.6375,
      "num_input_tokens_seen": 113246208,
      "step": 54
    },
    {
      "epoch": 0.9865470852017937,
      "grad_norm": 0.04142209142446518,
      "learning_rate": 0.0,
      "loss": 0.6419,
      "num_input_tokens_seen": 115343360,
      "step": 55
    },
    {
      "epoch": 0.9865470852017937,
      "num_input_tokens_seen": 115343360,
      "step": 55,
      "total_flos": 5.104238176512246e+18,
      "train_loss": 0.6637221011248502,
      "train_runtime": 9208.1472,
      "train_samples_per_second": 3.097,
      "train_steps_per_second": 0.006
    }
  ],
  "logging_steps": 1.0,
  "max_steps": 55,
  "num_input_tokens_seen": 115343360,
  "num_train_epochs": 1,
  "save_steps": 500,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 5.104238176512246e+18,
  "train_batch_size": 16,
  "trial_name": null,
  "trial_params": null
}