htlou's picture
Upload folder using huggingface_hub
5bec34e verified
raw
history blame
17.8 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.984771573604061,
"eval_steps": 50,
"global_step": 441,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0338409475465313,
"grad_norm": 18.879603600162138,
"learning_rate": 5e-07,
"loss": 1.7268,
"step": 5
},
{
"epoch": 0.0676818950930626,
"grad_norm": 12.562690441872693,
"learning_rate": 1e-06,
"loss": 1.5925,
"step": 10
},
{
"epoch": 0.10152284263959391,
"grad_norm": 8.05925907057448,
"learning_rate": 9.996679701338661e-07,
"loss": 1.269,
"step": 15
},
{
"epoch": 0.1353637901861252,
"grad_norm": 4.106846211376464,
"learning_rate": 9.986723215107924e-07,
"loss": 1.1436,
"step": 20
},
{
"epoch": 0.1692047377326565,
"grad_norm": 4.065415826389198,
"learning_rate": 9.97014376471095e-07,
"loss": 1.0624,
"step": 25
},
{
"epoch": 0.20304568527918782,
"grad_norm": 3.5507350425506274,
"learning_rate": 9.946963369638524e-07,
"loss": 1.0302,
"step": 30
},
{
"epoch": 0.23688663282571912,
"grad_norm": 3.620116416000427,
"learning_rate": 9.917212816224536e-07,
"loss": 1.0156,
"step": 35
},
{
"epoch": 0.2707275803722504,
"grad_norm": 3.392556379414998,
"learning_rate": 9.880931616758056e-07,
"loss": 0.9731,
"step": 40
},
{
"epoch": 0.30456852791878175,
"grad_norm": 3.6283098411860224,
"learning_rate": 9.838167957006293e-07,
"loss": 0.9453,
"step": 45
},
{
"epoch": 0.338409475465313,
"grad_norm": 3.643003094042682,
"learning_rate": 9.788978632218138e-07,
"loss": 0.9333,
"step": 50
},
{
"epoch": 0.338409475465313,
"eval_loss": 0.9387685656547546,
"eval_runtime": 73.7594,
"eval_samples_per_second": 56.942,
"eval_steps_per_second": 0.895,
"step": 50
},
{
"epoch": 0.37225042301184436,
"grad_norm": 3.5187728461211143,
"learning_rate": 9.73342897169329e-07,
"loss": 0.933,
"step": 55
},
{
"epoch": 0.40609137055837563,
"grad_norm": 3.221032807973684,
"learning_rate": 9.671592752017137e-07,
"loss": 0.9244,
"step": 60
},
{
"epoch": 0.43993231810490696,
"grad_norm": 3.6068076034822982,
"learning_rate": 9.603552099076648e-07,
"loss": 0.9217,
"step": 65
},
{
"epoch": 0.47377326565143824,
"grad_norm": 3.522728399930581,
"learning_rate": 9.52939737898737e-07,
"loss": 0.9251,
"step": 70
},
{
"epoch": 0.5076142131979695,
"grad_norm": 3.4156164712414805,
"learning_rate": 9.449227078076443e-07,
"loss": 0.8951,
"step": 75
},
{
"epoch": 0.5414551607445008,
"grad_norm": 3.5241377443178594,
"learning_rate": 9.363147672080985e-07,
"loss": 0.893,
"step": 80
},
{
"epoch": 0.5752961082910322,
"grad_norm": 3.480593037844269,
"learning_rate": 9.271273484735592e-07,
"loss": 0.8702,
"step": 85
},
{
"epoch": 0.6091370558375635,
"grad_norm": 3.3606090648530906,
"learning_rate": 9.173726535936766e-07,
"loss": 0.8931,
"step": 90
},
{
"epoch": 0.6429780033840947,
"grad_norm": 3.3802398568272545,
"learning_rate": 9.070636379685915e-07,
"loss": 0.8689,
"step": 95
},
{
"epoch": 0.676818950930626,
"grad_norm": 3.7340957238116013,
"learning_rate": 8.962139932026156e-07,
"loss": 0.8736,
"step": 100
},
{
"epoch": 0.676818950930626,
"eval_loss": 0.8820675015449524,
"eval_runtime": 73.2211,
"eval_samples_per_second": 57.361,
"eval_steps_per_second": 0.901,
"step": 100
},
{
"epoch": 0.7106598984771574,
"grad_norm": 3.491575137149294,
"learning_rate": 8.848381289201459e-07,
"loss": 0.8658,
"step": 105
},
{
"epoch": 0.7445008460236887,
"grad_norm": 3.3971585698046023,
"learning_rate": 8.72951153627962e-07,
"loss": 0.85,
"step": 110
},
{
"epoch": 0.7783417935702199,
"grad_norm": 3.4606145381962574,
"learning_rate": 8.605688546493238e-07,
"loss": 0.8831,
"step": 115
},
{
"epoch": 0.8121827411167513,
"grad_norm": 3.312370481580055,
"learning_rate": 8.477076771565202e-07,
"loss": 0.8542,
"step": 120
},
{
"epoch": 0.8460236886632826,
"grad_norm": 3.479568204762634,
"learning_rate": 8.343847023297169e-07,
"loss": 0.8538,
"step": 125
},
{
"epoch": 0.8798646362098139,
"grad_norm": 3.470397617651747,
"learning_rate": 8.206176246711065e-07,
"loss": 0.8568,
"step": 130
},
{
"epoch": 0.9137055837563451,
"grad_norm": 3.4164690427045974,
"learning_rate": 8.064247285044972e-07,
"loss": 0.8566,
"step": 135
},
{
"epoch": 0.9475465313028765,
"grad_norm": 3.5885722316358994,
"learning_rate": 7.918248636915459e-07,
"loss": 0.8671,
"step": 140
},
{
"epoch": 0.9813874788494078,
"grad_norm": 3.573742166264298,
"learning_rate": 7.768374205968906e-07,
"loss": 0.8607,
"step": 145
},
{
"epoch": 1.015228426395939,
"grad_norm": 3.7440804811259376,
"learning_rate": 7.614823043354285e-07,
"loss": 0.8459,
"step": 150
},
{
"epoch": 1.015228426395939,
"eval_loss": 0.8576342463493347,
"eval_runtime": 73.2015,
"eval_samples_per_second": 57.376,
"eval_steps_per_second": 0.902,
"step": 150
},
{
"epoch": 1.0490693739424704,
"grad_norm": 3.836769449038388,
"learning_rate": 7.457799083359471e-07,
"loss": 0.7928,
"step": 155
},
{
"epoch": 1.0829103214890017,
"grad_norm": 3.776692818561464,
"learning_rate": 7.297510872562131e-07,
"loss": 0.7801,
"step": 160
},
{
"epoch": 1.116751269035533,
"grad_norm": 3.5436254123817075,
"learning_rate": 7.134171292854955e-07,
"loss": 0.7915,
"step": 165
},
{
"epoch": 1.1505922165820643,
"grad_norm": 3.637774293664913,
"learning_rate": 6.967997278713089e-07,
"loss": 0.7831,
"step": 170
},
{
"epoch": 1.1844331641285957,
"grad_norm": 3.6828430568432817,
"learning_rate": 6.79920952907921e-07,
"loss": 0.7911,
"step": 175
},
{
"epoch": 1.218274111675127,
"grad_norm": 3.3501836574132673,
"learning_rate": 6.628032214248982e-07,
"loss": 0.7816,
"step": 180
},
{
"epoch": 1.252115059221658,
"grad_norm": 3.678603436999248,
"learning_rate": 6.454692678146119e-07,
"loss": 0.7742,
"step": 185
},
{
"epoch": 1.2859560067681894,
"grad_norm": 3.469023528162151,
"learning_rate": 6.279421136382494e-07,
"loss": 0.7932,
"step": 190
},
{
"epoch": 1.3197969543147208,
"grad_norm": 3.5311154474090007,
"learning_rate": 6.102450370504299e-07,
"loss": 0.7852,
"step": 195
},
{
"epoch": 1.353637901861252,
"grad_norm": 3.509395500338986,
"learning_rate": 5.924015418830354e-07,
"loss": 0.7682,
"step": 200
},
{
"epoch": 1.353637901861252,
"eval_loss": 0.8485246300697327,
"eval_runtime": 73.3335,
"eval_samples_per_second": 57.273,
"eval_steps_per_second": 0.9,
"step": 200
},
{
"epoch": 1.3874788494077834,
"grad_norm": 3.5828255600676226,
"learning_rate": 5.74435326429313e-07,
"loss": 0.7784,
"step": 205
},
{
"epoch": 1.4213197969543148,
"grad_norm": 3.8406423248810526,
"learning_rate": 5.563702519697108e-07,
"loss": 0.7789,
"step": 210
},
{
"epoch": 1.455160744500846,
"grad_norm": 3.5396413573689482,
"learning_rate": 5.382303110812466e-07,
"loss": 0.7843,
"step": 215
},
{
"epoch": 1.4890016920473772,
"grad_norm": 3.9366608589473184,
"learning_rate": 5.200395957725005e-07,
"loss": 0.7734,
"step": 220
},
{
"epoch": 1.5228426395939088,
"grad_norm": 3.6167320799092644,
"learning_rate": 5.018222654865471e-07,
"loss": 0.777,
"step": 225
},
{
"epoch": 1.5566835871404399,
"grad_norm": 3.508615499961612,
"learning_rate": 4.836025150143318e-07,
"loss": 0.7777,
"step": 230
},
{
"epoch": 1.5905245346869712,
"grad_norm": 3.8323856939840133,
"learning_rate": 4.654045423610952e-07,
"loss": 0.7744,
"step": 235
},
{
"epoch": 1.6243654822335025,
"grad_norm": 3.48551950556471,
"learning_rate": 4.4725251660853357e-07,
"loss": 0.7812,
"step": 240
},
{
"epoch": 1.6582064297800339,
"grad_norm": 3.2586294819951815,
"learning_rate": 4.2917054581536926e-07,
"loss": 0.768,
"step": 245
},
{
"epoch": 1.6920473773265652,
"grad_norm": 3.587138715732753,
"learning_rate": 4.1118264499897003e-07,
"loss": 0.7797,
"step": 250
},
{
"epoch": 1.6920473773265652,
"eval_loss": 0.8384872078895569,
"eval_runtime": 73.1511,
"eval_samples_per_second": 57.415,
"eval_steps_per_second": 0.902,
"step": 250
},
{
"epoch": 1.7258883248730963,
"grad_norm": 3.497575413714801,
"learning_rate": 3.9331270424053616e-07,
"loss": 0.7617,
"step": 255
},
{
"epoch": 1.7597292724196278,
"grad_norm": 3.410968439354172,
"learning_rate": 3.755844569562191e-07,
"loss": 0.7687,
"step": 260
},
{
"epoch": 1.793570219966159,
"grad_norm": 3.609149780281463,
"learning_rate": 3.580214483763093e-07,
"loss": 0.7602,
"step": 265
},
{
"epoch": 1.8274111675126905,
"grad_norm": 3.493512578978129,
"learning_rate": 3.406470042743574e-07,
"loss": 0.7717,
"step": 270
},
{
"epoch": 1.8612521150592216,
"grad_norm": 3.6002811799241714,
"learning_rate": 3.23484199987761e-07,
"loss": 0.7496,
"step": 275
},
{
"epoch": 1.895093062605753,
"grad_norm": 3.6855175892280303,
"learning_rate": 3.065558297709588e-07,
"loss": 0.7695,
"step": 280
},
{
"epoch": 1.9289340101522843,
"grad_norm": 3.6642552773973254,
"learning_rate": 2.898843765219388e-07,
"loss": 0.7628,
"step": 285
},
{
"epoch": 1.9627749576988156,
"grad_norm": 3.7846627756438913,
"learning_rate": 2.7349198192226295e-07,
"loss": 0.7732,
"step": 290
},
{
"epoch": 1.996615905245347,
"grad_norm": 3.4513200002016022,
"learning_rate": 2.574004170302696e-07,
"loss": 0.7684,
"step": 295
},
{
"epoch": 2.030456852791878,
"grad_norm": 3.8311631709862968,
"learning_rate": 2.4163105336650643e-07,
"loss": 0.7246,
"step": 300
},
{
"epoch": 2.030456852791878,
"eval_loss": 0.8336632251739502,
"eval_runtime": 73.1743,
"eval_samples_per_second": 57.397,
"eval_steps_per_second": 0.902,
"step": 300
},
{
"epoch": 2.0642978003384096,
"grad_norm": 3.9190275346300107,
"learning_rate": 2.2620483452979887e-07,
"loss": 0.7305,
"step": 305
},
{
"epoch": 2.0981387478849407,
"grad_norm": 3.6908653316747446,
"learning_rate": 2.1114224838164806e-07,
"loss": 0.7351,
"step": 310
},
{
"epoch": 2.1319796954314723,
"grad_norm": 3.802531438077045,
"learning_rate": 1.964632998359036e-07,
"loss": 0.7367,
"step": 315
},
{
"epoch": 2.1658206429780034,
"grad_norm": 3.775491751588902,
"learning_rate": 1.8218748428984782e-07,
"loss": 0.7233,
"step": 320
},
{
"epoch": 2.199661590524535,
"grad_norm": 3.5862593065525,
"learning_rate": 1.6833376173198005e-07,
"loss": 0.7233,
"step": 325
},
{
"epoch": 2.233502538071066,
"grad_norm": 3.5427938243892,
"learning_rate": 1.5492053156088498e-07,
"loss": 0.7098,
"step": 330
},
{
"epoch": 2.267343485617597,
"grad_norm": 3.61152784910624,
"learning_rate": 1.4196560814863567e-07,
"loss": 0.7129,
"step": 335
},
{
"epoch": 2.3011844331641287,
"grad_norm": 3.807265016749999,
"learning_rate": 1.294861971811773e-07,
"loss": 0.7074,
"step": 340
},
{
"epoch": 2.33502538071066,
"grad_norm": 3.6719299115770783,
"learning_rate": 1.1749887280712161e-07,
"loss": 0.7121,
"step": 345
},
{
"epoch": 2.3688663282571913,
"grad_norm": 3.5588978514727305,
"learning_rate": 1.0601955562529774e-07,
"loss": 0.7082,
"step": 350
},
{
"epoch": 2.3688663282571913,
"eval_loss": 0.8359954357147217,
"eval_runtime": 73.1022,
"eval_samples_per_second": 57.454,
"eval_steps_per_second": 0.903,
"step": 350
},
{
"epoch": 2.4027072758037225,
"grad_norm": 3.72273582563589,
"learning_rate": 9.506349154029425e-08,
"loss": 0.7117,
"step": 355
},
{
"epoch": 2.436548223350254,
"grad_norm": 3.777312137275298,
"learning_rate": 8.46452315140772e-08,
"loss": 0.7238,
"step": 360
},
{
"epoch": 2.470389170896785,
"grad_norm": 3.9355287879984404,
"learning_rate": 7.477861224057403e-08,
"loss": 0.7385,
"step": 365
},
{
"epoch": 2.504230118443316,
"grad_norm": 3.612978118989858,
"learning_rate": 6.547673776889095e-08,
"loss": 0.7239,
"step": 370
},
{
"epoch": 2.5380710659898478,
"grad_norm": 3.73334551587451,
"learning_rate": 5.6751962099570396e-08,
"loss": 0.6999,
"step": 375
},
{
"epoch": 2.571912013536379,
"grad_norm": 3.7980379085271974,
"learning_rate": 4.861587277700274e-08,
"loss": 0.7253,
"step": 380
},
{
"epoch": 2.6057529610829104,
"grad_norm": 3.8953973449204105,
"learning_rate": 4.107927549978235e-08,
"loss": 0.7267,
"step": 385
},
{
"epoch": 2.6395939086294415,
"grad_norm": 4.0344872537944525,
"learning_rate": 3.4152179769449396e-08,
"loss": 0.7179,
"step": 390
},
{
"epoch": 2.673434856175973,
"grad_norm": 3.8358198567516935,
"learning_rate": 2.784378559667622e-08,
"loss": 0.7043,
"step": 395
},
{
"epoch": 2.707275803722504,
"grad_norm": 3.5246037765259532,
"learning_rate": 2.2162471282553553e-08,
"loss": 0.7007,
"step": 400
},
{
"epoch": 2.707275803722504,
"eval_loss": 0.8346139788627625,
"eval_runtime": 73.1876,
"eval_samples_per_second": 57.387,
"eval_steps_per_second": 0.902,
"step": 400
},
{
"epoch": 2.7411167512690353,
"grad_norm": 3.808011429657169,
"learning_rate": 1.7115782291206082e-08,
"loss": 0.7232,
"step": 405
},
{
"epoch": 2.774957698815567,
"grad_norm": 3.8868492324851003,
"learning_rate": 1.2710421228514733e-08,
"loss": 0.7213,
"step": 410
},
{
"epoch": 2.808798646362098,
"grad_norm": 3.8016855221337953,
"learning_rate": 8.952238940255153e-09,
"loss": 0.7114,
"step": 415
},
{
"epoch": 2.8426395939086295,
"grad_norm": 3.6712328006089763,
"learning_rate": 5.846226741475557e-09,
"loss": 0.7148,
"step": 420
},
{
"epoch": 2.8764805414551606,
"grad_norm": 3.7375800014295417,
"learning_rate": 3.3965097874343872e-09,
"loss": 0.731,
"step": 425
},
{
"epoch": 2.910321489001692,
"grad_norm": 3.6510115083695562,
"learning_rate": 1.6063415949008618e-09,
"loss": 0.7146,
"step": 430
},
{
"epoch": 2.9441624365482233,
"grad_norm": 4.038300563245875,
"learning_rate": 4.780997210962478e-10,
"loss": 0.7113,
"step": 435
},
{
"epoch": 2.9780033840947544,
"grad_norm": 3.6454952565628376,
"learning_rate": 1.328260601385356e-11,
"loss": 0.701,
"step": 440
},
{
"epoch": 2.984771573604061,
"step": 441,
"total_flos": 2600101931384832.0,
"train_loss": 0.8247822573125497,
"train_runtime": 6397.793,
"train_samples_per_second": 17.725,
"train_steps_per_second": 0.069
}
],
"logging_steps": 5,
"max_steps": 441,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 10000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2600101931384832.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}