nttx commited on
Commit
d635734
·
verified ·
1 Parent(s): 3087d10

Training in progress, step 1650, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc0563610e9795da6193abe9970ea5bdd798757bad57849692ef086caa273f7d
3
  size 1101095848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de088b2de1ee181bd71f905f8d5857efe95f9d48787209fce8bc6015c4c58137
3
  size 1101095848
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3284ff4276b76c44987eb43330358b6c7c1a8f3bf83a04fd1796b7d3c6aac59c
3
  size 559894868
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:647ed4da156e09a8fd511aca2fdf0669f08f7f5c5d40e062eafcfb42b63efa62
3
  size 559894868
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3681057530ddf641436a54f3c5d867353c0b63200499627cd2751802959141cc
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf99c4a322b12e03b16f1e7ad8423bdee73f9697406f3e5b3166adc0a2af306b
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:befb5de8641401f08357d93e09a6b263a413b6c62f0e4dc381df697735030b77
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:090e504a81765b504142cff403afcf92b817f0e48945f0e4e014e79a6565dad0
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 1.826446294784546,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-1350",
4
- "epoch": 1.0871534698314913,
5
  "eval_steps": 150,
6
- "global_step": 1500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -305,6 +305,35 @@
305
  "eval_samples_per_second": 13.401,
306
  "eval_steps_per_second": 3.356,
307
  "step": 1500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
308
  }
309
  ],
310
  "logging_steps": 50,
@@ -319,7 +348,7 @@
319
  "early_stopping_threshold": 0.0
320
  },
321
  "attributes": {
322
- "early_stopping_patience_counter": 1
323
  }
324
  },
325
  "TrainerControl": {
@@ -328,12 +357,12 @@
328
  "should_evaluate": false,
329
  "should_log": false,
330
  "should_save": true,
331
- "should_training_stop": false
332
  },
333
  "attributes": {}
334
  }
335
  },
336
- "total_flos": 1.0828424646607503e+18,
337
  "train_batch_size": 4,
338
  "trial_name": null,
339
  "trial_params": null
 
1
  {
2
  "best_metric": 1.826446294784546,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-1350",
4
+ "epoch": 1.1958688168146403,
5
  "eval_steps": 150,
6
+ "global_step": 1650,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
305
  "eval_samples_per_second": 13.401,
306
  "eval_steps_per_second": 3.356,
307
  "step": 1500
308
+ },
309
+ {
310
+ "epoch": 1.1233919188258743,
311
+ "grad_norm": 0.8776970505714417,
312
+ "learning_rate": 9.733794785622253e-05,
313
+ "loss": 1.3963,
314
+ "step": 1550
315
+ },
316
+ {
317
+ "epoch": 1.1596303678202573,
318
+ "grad_norm": 0.9756889939308167,
319
+ "learning_rate": 9.202138944469168e-05,
320
+ "loss": 1.393,
321
+ "step": 1600
322
+ },
323
+ {
324
+ "epoch": 1.1958688168146403,
325
+ "grad_norm": 0.8379771113395691,
326
+ "learning_rate": 8.672744727162781e-05,
327
+ "loss": 1.4002,
328
+ "step": 1650
329
+ },
330
+ {
331
+ "epoch": 1.1958688168146403,
332
+ "eval_loss": 1.837064504623413,
333
+ "eval_runtime": 86.6691,
334
+ "eval_samples_per_second": 13.407,
335
+ "eval_steps_per_second": 3.358,
336
+ "step": 1650
337
  }
338
  ],
339
  "logging_steps": 50,
 
348
  "early_stopping_threshold": 0.0
349
  },
350
  "attributes": {
351
+ "early_stopping_patience_counter": 2
352
  }
353
  },
354
  "TrainerControl": {
 
357
  "should_evaluate": false,
358
  "should_log": false,
359
  "should_save": true,
360
+ "should_training_stop": true
361
  },
362
  "attributes": {}
363
  }
364
  },
365
+ "total_flos": 1.1904824490055434e+18,
366
  "train_batch_size": 4,
367
  "trial_name": null,
368
  "trial_params": null