w11wo commited on
Commit
6477ec0
·
1 Parent(s): 96ce86a

End of training

Browse files
README.md CHANGED
@@ -20,11 +20,11 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [indobenchmark/indobert-base-p1](https://huggingface.co/indobenchmark/indobert-base-p1) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.0108
24
- - Accuracy: 0.8619
25
- - F1: 0.7176
26
- - Precision: 0.7344
27
- - Recall: 0.7015
28
 
29
  ## Model description
30
 
 
20
 
21
  This model is a fine-tuned version of [indobenchmark/indobert-base-p1](https://huggingface.co/indobenchmark/indobert-base-p1) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.8125
24
+ - Accuracy: 0.8662
25
+ - F1: 0.7273
26
+ - Precision: 0.7385
27
+ - Recall: 0.7164
28
 
29
  ## Model description
30
 
all_results.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 11.0,
3
+ "eval_accuracy": 0.8661710037174721,
4
+ "eval_f1": 0.7272727272727273,
5
+ "eval_loss": 0.8124564290046692,
6
+ "eval_precision": 0.7384615384615385,
7
+ "eval_recall": 0.7164179104477612,
8
+ "eval_runtime": 2.139,
9
+ "eval_samples": 268,
10
+ "eval_samples_per_second": 251.523,
11
+ "eval_steps_per_second": 4.208,
12
+ "train_loss": 0.11986877576079226,
13
+ "train_runtime": 294.5192,
14
+ "train_samples": 1878,
15
+ "train_samples_per_second": 637.649,
16
+ "train_steps_per_second": 20.033
17
+ }
eval_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 11.0,
3
+ "eval_accuracy": 0.8661710037174721,
4
+ "eval_f1": 0.7272727272727273,
5
+ "eval_loss": 0.8124564290046692,
6
+ "eval_precision": 0.7384615384615385,
7
+ "eval_recall": 0.7164179104477612,
8
+ "eval_runtime": 2.139,
9
+ "eval_samples": 268,
10
+ "eval_samples_per_second": 251.523,
11
+ "eval_steps_per_second": 4.208
12
+ }
predict_results.txt ADDED
@@ -0,0 +1,539 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ index prediction
2
+ 0 1
3
+ 1 0
4
+ 2 0
5
+ 3 0
6
+ 4 0
7
+ 5 0
8
+ 6 1
9
+ 7 1
10
+ 8 0
11
+ 9 0
12
+ 10 0
13
+ 11 0
14
+ 12 1
15
+ 13 0
16
+ 14 0
17
+ 15 1
18
+ 16 1
19
+ 17 0
20
+ 18 0
21
+ 19 0
22
+ 20 0
23
+ 21 0
24
+ 22 0
25
+ 23 1
26
+ 24 0
27
+ 25 0
28
+ 26 0
29
+ 27 1
30
+ 28 0
31
+ 29 0
32
+ 30 1
33
+ 31 0
34
+ 32 1
35
+ 33 1
36
+ 34 0
37
+ 35 0
38
+ 36 0
39
+ 37 0
40
+ 38 0
41
+ 39 0
42
+ 40 0
43
+ 41 0
44
+ 42 0
45
+ 43 0
46
+ 44 0
47
+ 45 0
48
+ 46 0
49
+ 47 1
50
+ 48 0
51
+ 49 0
52
+ 50 1
53
+ 51 0
54
+ 52 0
55
+ 53 0
56
+ 54 0
57
+ 55 0
58
+ 56 0
59
+ 57 0
60
+ 58 0
61
+ 59 0
62
+ 60 0
63
+ 61 0
64
+ 62 0
65
+ 63 0
66
+ 64 0
67
+ 65 0
68
+ 66 0
69
+ 67 0
70
+ 68 0
71
+ 69 0
72
+ 70 1
73
+ 71 1
74
+ 72 0
75
+ 73 1
76
+ 74 0
77
+ 75 1
78
+ 76 1
79
+ 77 0
80
+ 78 0
81
+ 79 0
82
+ 80 0
83
+ 81 0
84
+ 82 0
85
+ 83 1
86
+ 84 0
87
+ 85 0
88
+ 86 0
89
+ 87 0
90
+ 88 1
91
+ 89 0
92
+ 90 0
93
+ 91 0
94
+ 92 0
95
+ 93 0
96
+ 94 0
97
+ 95 0
98
+ 96 1
99
+ 97 0
100
+ 98 0
101
+ 99 0
102
+ 100 0
103
+ 101 0
104
+ 102 0
105
+ 103 0
106
+ 104 0
107
+ 105 1
108
+ 106 0
109
+ 107 0
110
+ 108 1
111
+ 109 0
112
+ 110 0
113
+ 111 1
114
+ 112 0
115
+ 113 0
116
+ 114 0
117
+ 115 0
118
+ 116 0
119
+ 117 0
120
+ 118 0
121
+ 119 1
122
+ 120 0
123
+ 121 0
124
+ 122 0
125
+ 123 0
126
+ 124 1
127
+ 125 0
128
+ 126 0
129
+ 127 0
130
+ 128 0
131
+ 129 0
132
+ 130 1
133
+ 131 0
134
+ 132 0
135
+ 133 0
136
+ 134 0
137
+ 135 0
138
+ 136 1
139
+ 137 0
140
+ 138 0
141
+ 139 0
142
+ 140 1
143
+ 141 0
144
+ 142 1
145
+ 143 0
146
+ 144 1
147
+ 145 0
148
+ 146 0
149
+ 147 0
150
+ 148 0
151
+ 149 0
152
+ 150 0
153
+ 151 0
154
+ 152 0
155
+ 153 0
156
+ 154 1
157
+ 155 0
158
+ 156 0
159
+ 157 0
160
+ 158 0
161
+ 159 0
162
+ 160 0
163
+ 161 0
164
+ 162 0
165
+ 163 1
166
+ 164 0
167
+ 165 0
168
+ 166 0
169
+ 167 1
170
+ 168 0
171
+ 169 0
172
+ 170 0
173
+ 171 1
174
+ 172 0
175
+ 173 0
176
+ 174 1
177
+ 175 0
178
+ 176 0
179
+ 177 0
180
+ 178 0
181
+ 179 0
182
+ 180 1
183
+ 181 0
184
+ 182 1
185
+ 183 1
186
+ 184 0
187
+ 185 1
188
+ 186 0
189
+ 187 1
190
+ 188 0
191
+ 189 0
192
+ 190 0
193
+ 191 0
194
+ 192 0
195
+ 193 0
196
+ 194 0
197
+ 195 0
198
+ 196 0
199
+ 197 0
200
+ 198 1
201
+ 199 0
202
+ 200 0
203
+ 201 0
204
+ 202 1
205
+ 203 1
206
+ 204 0
207
+ 205 1
208
+ 206 1
209
+ 207 0
210
+ 208 1
211
+ 209 0
212
+ 210 0
213
+ 211 0
214
+ 212 1
215
+ 213 0
216
+ 214 0
217
+ 215 0
218
+ 216 1
219
+ 217 0
220
+ 218 1
221
+ 219 0
222
+ 220 0
223
+ 221 0
224
+ 222 0
225
+ 223 1
226
+ 224 0
227
+ 225 0
228
+ 226 1
229
+ 227 0
230
+ 228 1
231
+ 229 0
232
+ 230 0
233
+ 231 1
234
+ 232 1
235
+ 233 0
236
+ 234 1
237
+ 235 0
238
+ 236 0
239
+ 237 1
240
+ 238 0
241
+ 239 0
242
+ 240 1
243
+ 241 0
244
+ 242 0
245
+ 243 0
246
+ 244 0
247
+ 245 0
248
+ 246 0
249
+ 247 0
250
+ 248 0
251
+ 249 1
252
+ 250 0
253
+ 251 0
254
+ 252 0
255
+ 253 1
256
+ 254 1
257
+ 255 0
258
+ 256 0
259
+ 257 1
260
+ 258 0
261
+ 259 1
262
+ 260 1
263
+ 261 0
264
+ 262 0
265
+ 263 0
266
+ 264 0
267
+ 265 1
268
+ 266 0
269
+ 267 0
270
+ 268 0
271
+ 269 0
272
+ 270 0
273
+ 271 1
274
+ 272 1
275
+ 273 0
276
+ 274 0
277
+ 275 0
278
+ 276 0
279
+ 277 0
280
+ 278 0
281
+ 279 1
282
+ 280 0
283
+ 281 0
284
+ 282 0
285
+ 283 0
286
+ 284 0
287
+ 285 0
288
+ 286 0
289
+ 287 0
290
+ 288 0
291
+ 289 0
292
+ 290 0
293
+ 291 0
294
+ 292 0
295
+ 293 0
296
+ 294 1
297
+ 295 0
298
+ 296 0
299
+ 297 0
300
+ 298 0
301
+ 299 1
302
+ 300 0
303
+ 301 0
304
+ 302 0
305
+ 303 1
306
+ 304 0
307
+ 305 0
308
+ 306 1
309
+ 307 0
310
+ 308 0
311
+ 309 0
312
+ 310 0
313
+ 311 0
314
+ 312 0
315
+ 313 1
316
+ 314 0
317
+ 315 0
318
+ 316 0
319
+ 317 0
320
+ 318 0
321
+ 319 0
322
+ 320 1
323
+ 321 1
324
+ 322 0
325
+ 323 0
326
+ 324 1
327
+ 325 1
328
+ 326 1
329
+ 327 0
330
+ 328 0
331
+ 329 0
332
+ 330 0
333
+ 331 1
334
+ 332 0
335
+ 333 0
336
+ 334 0
337
+ 335 0
338
+ 336 0
339
+ 337 0
340
+ 338 0
341
+ 339 0
342
+ 340 1
343
+ 341 1
344
+ 342 1
345
+ 343 0
346
+ 344 0
347
+ 345 0
348
+ 346 1
349
+ 347 1
350
+ 348 0
351
+ 349 0
352
+ 350 0
353
+ 351 0
354
+ 352 0
355
+ 353 1
356
+ 354 0
357
+ 355 0
358
+ 356 0
359
+ 357 0
360
+ 358 0
361
+ 359 0
362
+ 360 1
363
+ 361 0
364
+ 362 0
365
+ 363 0
366
+ 364 0
367
+ 365 0
368
+ 366 1
369
+ 367 1
370
+ 368 1
371
+ 369 0
372
+ 370 0
373
+ 371 1
374
+ 372 0
375
+ 373 1
376
+ 374 0
377
+ 375 0
378
+ 376 0
379
+ 377 0
380
+ 378 1
381
+ 379 0
382
+ 380 1
383
+ 381 0
384
+ 382 0
385
+ 383 0
386
+ 384 1
387
+ 385 0
388
+ 386 0
389
+ 387 0
390
+ 388 0
391
+ 389 0
392
+ 390 1
393
+ 391 0
394
+ 392 0
395
+ 393 0
396
+ 394 1
397
+ 395 0
398
+ 396 0
399
+ 397 0
400
+ 398 1
401
+ 399 0
402
+ 400 1
403
+ 401 0
404
+ 402 0
405
+ 403 1
406
+ 404 0
407
+ 405 1
408
+ 406 1
409
+ 407 1
410
+ 408 0
411
+ 409 0
412
+ 410 0
413
+ 411 1
414
+ 412 0
415
+ 413 0
416
+ 414 0
417
+ 415 0
418
+ 416 0
419
+ 417 1
420
+ 418 1
421
+ 419 1
422
+ 420 0
423
+ 421 0
424
+ 422 1
425
+ 423 0
426
+ 424 0
427
+ 425 0
428
+ 426 1
429
+ 427 0
430
+ 428 0
431
+ 429 1
432
+ 430 0
433
+ 431 0
434
+ 432 0
435
+ 433 1
436
+ 434 0
437
+ 435 0
438
+ 436 0
439
+ 437 0
440
+ 438 0
441
+ 439 0
442
+ 440 0
443
+ 441 1
444
+ 442 0
445
+ 443 0
446
+ 444 0
447
+ 445 0
448
+ 446 0
449
+ 447 0
450
+ 448 1
451
+ 449 1
452
+ 450 0
453
+ 451 0
454
+ 452 0
455
+ 453 0
456
+ 454 0
457
+ 455 0
458
+ 456 0
459
+ 457 0
460
+ 458 0
461
+ 459 0
462
+ 460 0
463
+ 461 0
464
+ 462 0
465
+ 463 1
466
+ 464 0
467
+ 465 0
468
+ 466 0
469
+ 467 1
470
+ 468 0
471
+ 469 0
472
+ 470 0
473
+ 471 1
474
+ 472 0
475
+ 473 1
476
+ 474 0
477
+ 475 1
478
+ 476 0
479
+ 477 0
480
+ 478 1
481
+ 479 0
482
+ 480 1
483
+ 481 0
484
+ 482 0
485
+ 483 0
486
+ 484 0
487
+ 485 0
488
+ 486 1
489
+ 487 1
490
+ 488 0
491
+ 489 1
492
+ 490 0
493
+ 491 0
494
+ 492 0
495
+ 493 0
496
+ 494 1
497
+ 495 1
498
+ 496 0
499
+ 497 0
500
+ 498 0
501
+ 499 0
502
+ 500 0
503
+ 501 0
504
+ 502 0
505
+ 503 0
506
+ 504 0
507
+ 505 0
508
+ 506 0
509
+ 507 0
510
+ 508 0
511
+ 509 0
512
+ 510 0
513
+ 511 0
514
+ 512 0
515
+ 513 0
516
+ 514 0
517
+ 515 0
518
+ 516 0
519
+ 517 1
520
+ 518 0
521
+ 519 1
522
+ 520 0
523
+ 521 0
524
+ 522 0
525
+ 523 0
526
+ 524 0
527
+ 525 0
528
+ 526 0
529
+ 527 0
530
+ 528 0
531
+ 529 0
532
+ 530 0
533
+ 531 0
534
+ 532 1
535
+ 533 0
536
+ 534 0
537
+ 535 0
538
+ 536 1
539
+ 537 1
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 11.0,
3
+ "train_loss": 0.11986877576079226,
4
+ "train_runtime": 294.5192,
5
+ "train_samples": 1878,
6
+ "train_samples_per_second": 637.649,
7
+ "train_steps_per_second": 20.033
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,228 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.7445255474452556,
3
+ "best_model_checkpoint": "outputs/indobert-base-p1-twitter-indonesia-sarcastic/checkpoint-472",
4
+ "epoch": 11.0,
5
+ "eval_steps": 500,
6
+ "global_step": 649,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "learning_rate": 9.997532801828659e-06,
14
+ "loss": 0.5101,
15
+ "step": 59
16
+ },
17
+ {
18
+ "epoch": 1.0,
19
+ "eval_accuracy": 0.7835820895522388,
20
+ "eval_f1": 0.608108108108108,
21
+ "eval_loss": 0.4220750629901886,
22
+ "eval_precision": 0.5555555555555556,
23
+ "eval_recall": 0.6716417910447762,
24
+ "eval_runtime": 1.0673,
25
+ "eval_samples_per_second": 251.108,
26
+ "eval_steps_per_second": 4.685,
27
+ "step": 59
28
+ },
29
+ {
30
+ "epoch": 2.0,
31
+ "learning_rate": 9.990133642141359e-06,
32
+ "loss": 0.3508,
33
+ "step": 118
34
+ },
35
+ {
36
+ "epoch": 2.0,
37
+ "eval_accuracy": 0.8246268656716418,
38
+ "eval_f1": 0.6713286713286712,
39
+ "eval_loss": 0.34792423248291016,
40
+ "eval_precision": 0.631578947368421,
41
+ "eval_recall": 0.7164179104477612,
42
+ "eval_runtime": 1.0685,
43
+ "eval_samples_per_second": 250.816,
44
+ "eval_steps_per_second": 4.679,
45
+ "step": 118
46
+ },
47
+ {
48
+ "epoch": 3.0,
49
+ "learning_rate": 9.9778098230154e-06,
50
+ "loss": 0.221,
51
+ "step": 177
52
+ },
53
+ {
54
+ "epoch": 3.0,
55
+ "eval_accuracy": 0.8582089552238806,
56
+ "eval_f1": 0.6935483870967741,
57
+ "eval_loss": 0.3511403799057007,
58
+ "eval_precision": 0.7543859649122807,
59
+ "eval_recall": 0.6417910447761194,
60
+ "eval_runtime": 1.0653,
61
+ "eval_samples_per_second": 251.58,
62
+ "eval_steps_per_second": 4.694,
63
+ "step": 177
64
+ },
65
+ {
66
+ "epoch": 4.0,
67
+ "learning_rate": 9.960573506572391e-06,
68
+ "loss": 0.1157,
69
+ "step": 236
70
+ },
71
+ {
72
+ "epoch": 4.0,
73
+ "eval_accuracy": 0.8395522388059702,
74
+ "eval_f1": 0.6861313868613139,
75
+ "eval_loss": 0.4352029263973236,
76
+ "eval_precision": 0.6714285714285714,
77
+ "eval_recall": 0.7014925373134329,
78
+ "eval_runtime": 1.0649,
79
+ "eval_samples_per_second": 251.673,
80
+ "eval_steps_per_second": 4.695,
81
+ "step": 236
82
+ },
83
+ {
84
+ "epoch": 5.0,
85
+ "learning_rate": 9.938441702975689e-06,
86
+ "loss": 0.0453,
87
+ "step": 295
88
+ },
89
+ {
90
+ "epoch": 5.0,
91
+ "eval_accuracy": 0.8582089552238806,
92
+ "eval_f1": 0.7076923076923077,
93
+ "eval_loss": 0.6923192739486694,
94
+ "eval_precision": 0.7301587301587301,
95
+ "eval_recall": 0.6865671641791045,
96
+ "eval_runtime": 1.064,
97
+ "eval_samples_per_second": 251.881,
98
+ "eval_steps_per_second": 4.699,
99
+ "step": 295
100
+ },
101
+ {
102
+ "epoch": 6.0,
103
+ "learning_rate": 9.911436253643445e-06,
104
+ "loss": 0.0192,
105
+ "step": 354
106
+ },
107
+ {
108
+ "epoch": 6.0,
109
+ "eval_accuracy": 0.8694029850746269,
110
+ "eval_f1": 0.7286821705426356,
111
+ "eval_loss": 0.7378106117248535,
112
+ "eval_precision": 0.7580645161290323,
113
+ "eval_recall": 0.7014925373134329,
114
+ "eval_runtime": 1.0705,
115
+ "eval_samples_per_second": 250.361,
116
+ "eval_steps_per_second": 4.671,
117
+ "step": 354
118
+ },
119
+ {
120
+ "epoch": 7.0,
121
+ "learning_rate": 9.880163895209396e-06,
122
+ "loss": 0.0159,
123
+ "step": 413
124
+ },
125
+ {
126
+ "epoch": 7.0,
127
+ "eval_accuracy": 0.8544776119402985,
128
+ "eval_f1": 0.6722689075630253,
129
+ "eval_loss": 0.8860062956809998,
130
+ "eval_precision": 0.7692307692307693,
131
+ "eval_recall": 0.5970149253731343,
132
+ "eval_runtime": 1.0639,
133
+ "eval_samples_per_second": 251.895,
134
+ "eval_steps_per_second": 4.7,
135
+ "step": 413
136
+ },
137
+ {
138
+ "epoch": 8.0,
139
+ "learning_rate": 9.843577222724305e-06,
140
+ "loss": 0.0165,
141
+ "step": 472
142
+ },
143
+ {
144
+ "epoch": 8.0,
145
+ "eval_accuracy": 0.8694029850746269,
146
+ "eval_f1": 0.7445255474452556,
147
+ "eval_loss": 0.8260650634765625,
148
+ "eval_precision": 0.7285714285714285,
149
+ "eval_recall": 0.7611940298507462,
150
+ "eval_runtime": 1.0763,
151
+ "eval_samples_per_second": 249.008,
152
+ "eval_steps_per_second": 4.646,
153
+ "step": 472
154
+ },
155
+ {
156
+ "epoch": 9.0,
157
+ "learning_rate": 9.80221052429255e-06,
158
+ "loss": 0.0175,
159
+ "step": 531
160
+ },
161
+ {
162
+ "epoch": 9.0,
163
+ "eval_accuracy": 0.8731343283582089,
164
+ "eval_f1": 0.7424242424242424,
165
+ "eval_loss": 0.8732029795646667,
166
+ "eval_precision": 0.7538461538461538,
167
+ "eval_recall": 0.7313432835820896,
168
+ "eval_runtime": 1.0644,
169
+ "eval_samples_per_second": 251.793,
170
+ "eval_steps_per_second": 4.698,
171
+ "step": 531
172
+ },
173
+ {
174
+ "epoch": 10.0,
175
+ "learning_rate": 9.756104623851222e-06,
176
+ "loss": 0.0062,
177
+ "step": 590
178
+ },
179
+ {
180
+ "epoch": 10.0,
181
+ "eval_accuracy": 0.8656716417910447,
182
+ "eval_f1": 0.7272727272727273,
183
+ "eval_loss": 0.9647775888442993,
184
+ "eval_precision": 0.7384615384615385,
185
+ "eval_recall": 0.7164179104477612,
186
+ "eval_runtime": 1.0637,
187
+ "eval_samples_per_second": 251.945,
188
+ "eval_steps_per_second": 4.7,
189
+ "step": 590
190
+ },
191
+ {
192
+ "epoch": 11.0,
193
+ "learning_rate": 9.705305022357621e-06,
194
+ "loss": 0.0003,
195
+ "step": 649
196
+ },
197
+ {
198
+ "epoch": 11.0,
199
+ "eval_accuracy": 0.8619402985074627,
200
+ "eval_f1": 0.7175572519083969,
201
+ "eval_loss": 1.0107759237289429,
202
+ "eval_precision": 0.734375,
203
+ "eval_recall": 0.7014925373134329,
204
+ "eval_runtime": 1.0637,
205
+ "eval_samples_per_second": 251.951,
206
+ "eval_steps_per_second": 4.701,
207
+ "step": 649
208
+ },
209
+ {
210
+ "epoch": 11.0,
211
+ "step": 649,
212
+ "total_flos": 1358837045406720.0,
213
+ "train_loss": 0.11986877576079226,
214
+ "train_runtime": 294.5192,
215
+ "train_samples_per_second": 637.649,
216
+ "train_steps_per_second": 20.033
217
+ }
218
+ ],
219
+ "logging_steps": 500,
220
+ "max_steps": 5900,
221
+ "num_input_tokens_seen": 0,
222
+ "num_train_epochs": 100,
223
+ "save_steps": 500,
224
+ "total_flos": 1358837045406720.0,
225
+ "train_batch_size": 32,
226
+ "trial_name": null,
227
+ "trial_params": null
228
+ }