|
{ |
|
"best_metric": 0.0942394882440567, |
|
"best_model_checkpoint": "/content/train/Qwen2.5-VL-3B-Instruct-unsloth-r8-rslora/checkpoint-320", |
|
"epoch": 2.4218455743879472, |
|
"eval_steps": 10, |
|
"global_step": 320, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.007532956685499058, |
|
"grad_norm": 1.6337342262268066, |
|
"learning_rate": 1e-05, |
|
"loss": 2.8339, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.015065913370998116, |
|
"grad_norm": 1.7995260953903198, |
|
"learning_rate": 2e-05, |
|
"loss": 3.1612, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.022598870056497175, |
|
"grad_norm": 1.8331536054611206, |
|
"learning_rate": 3e-05, |
|
"loss": 2.9529, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.030131826741996232, |
|
"grad_norm": 1.8774369955062866, |
|
"learning_rate": 4e-05, |
|
"loss": 3.1371, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.03766478342749529, |
|
"grad_norm": 2.1215269565582275, |
|
"learning_rate": 5e-05, |
|
"loss": 3.4216, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.04519774011299435, |
|
"grad_norm": 1.7083489894866943, |
|
"learning_rate": 6e-05, |
|
"loss": 2.5018, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.05273069679849341, |
|
"grad_norm": 1.9553663730621338, |
|
"learning_rate": 7e-05, |
|
"loss": 2.8832, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.060263653483992465, |
|
"grad_norm": 2.085784673690796, |
|
"learning_rate": 8e-05, |
|
"loss": 2.9757, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.06779661016949153, |
|
"grad_norm": 1.949724793434143, |
|
"learning_rate": 9e-05, |
|
"loss": 2.7228, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.07532956685499058, |
|
"grad_norm": 2.1280415058135986, |
|
"learning_rate": 0.0001, |
|
"loss": 2.9347, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.07532956685499058, |
|
"eval_loss": 2.670034646987915, |
|
"eval_runtime": 54.8057, |
|
"eval_samples_per_second": 1.368, |
|
"eval_steps_per_second": 0.693, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.08286252354048965, |
|
"grad_norm": 2.418717861175537, |
|
"learning_rate": 9.999834399079165e-05, |
|
"loss": 3.0651, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0903954802259887, |
|
"grad_norm": 1.9917370080947876, |
|
"learning_rate": 9.99933760728612e-05, |
|
"loss": 2.4549, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.09792843691148775, |
|
"grad_norm": 1.8027087450027466, |
|
"learning_rate": 9.99850965752854e-05, |
|
"loss": 2.2586, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.10546139359698682, |
|
"grad_norm": 1.6262425184249878, |
|
"learning_rate": 9.997350604650123e-05, |
|
"loss": 2.0871, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.11299435028248588, |
|
"grad_norm": 1.5385862588882446, |
|
"learning_rate": 9.995860525426954e-05, |
|
"loss": 2.0329, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.12052730696798493, |
|
"grad_norm": 1.1733174324035645, |
|
"learning_rate": 9.994039518562432e-05, |
|
"loss": 1.6894, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.128060263653484, |
|
"grad_norm": 1.1716901063919067, |
|
"learning_rate": 9.991887704680724e-05, |
|
"loss": 1.8195, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.13559322033898305, |
|
"grad_norm": 1.3725332021713257, |
|
"learning_rate": 9.989405226318772e-05, |
|
"loss": 1.9845, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.1431261770244821, |
|
"grad_norm": 0.8536051511764526, |
|
"learning_rate": 9.986592247916858e-05, |
|
"loss": 1.5969, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.15065913370998116, |
|
"grad_norm": 0.6615162491798401, |
|
"learning_rate": 9.983448955807708e-05, |
|
"loss": 1.6763, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.15065913370998116, |
|
"eval_loss": 1.6168410778045654, |
|
"eval_runtime": 43.9067, |
|
"eval_samples_per_second": 1.708, |
|
"eval_steps_per_second": 0.865, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.15819209039548024, |
|
"grad_norm": 0.601297914981842, |
|
"learning_rate": 9.979975558204147e-05, |
|
"loss": 1.6742, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.1657250470809793, |
|
"grad_norm": 0.4877667725086212, |
|
"learning_rate": 9.976172285185314e-05, |
|
"loss": 1.6194, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.17325800376647835, |
|
"grad_norm": 0.4728071391582489, |
|
"learning_rate": 9.972039388681413e-05, |
|
"loss": 1.6243, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.1807909604519774, |
|
"grad_norm": 0.37902191281318665, |
|
"learning_rate": 9.967577142457032e-05, |
|
"loss": 1.4569, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.18832391713747645, |
|
"grad_norm": 0.45830923318862915, |
|
"learning_rate": 9.962785842093003e-05, |
|
"loss": 1.6214, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.1958568738229755, |
|
"grad_norm": 0.39010411500930786, |
|
"learning_rate": 9.957665804966829e-05, |
|
"loss": 1.345, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.2033898305084746, |
|
"grad_norm": 0.5287721753120422, |
|
"learning_rate": 9.952217370231653e-05, |
|
"loss": 1.6019, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.21092278719397364, |
|
"grad_norm": 0.4444209635257721, |
|
"learning_rate": 9.946440898793801e-05, |
|
"loss": 1.4566, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.2184557438794727, |
|
"grad_norm": 0.4856286346912384, |
|
"learning_rate": 9.940336773288865e-05, |
|
"loss": 1.5087, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.22598870056497175, |
|
"grad_norm": 0.4327958822250366, |
|
"learning_rate": 9.933905398056372e-05, |
|
"loss": 1.2924, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.22598870056497175, |
|
"eval_loss": 1.4212392568588257, |
|
"eval_runtime": 44.0456, |
|
"eval_samples_per_second": 1.703, |
|
"eval_steps_per_second": 0.863, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.2335216572504708, |
|
"grad_norm": 0.49079710245132446, |
|
"learning_rate": 9.92714719911298e-05, |
|
"loss": 1.3501, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.24105461393596986, |
|
"grad_norm": 0.8166887760162354, |
|
"learning_rate": 9.920062624124282e-05, |
|
"loss": 1.4564, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.24858757062146894, |
|
"grad_norm": 0.6213160157203674, |
|
"learning_rate": 9.912652142375132e-05, |
|
"loss": 1.5757, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.256120527306968, |
|
"grad_norm": 0.49864110350608826, |
|
"learning_rate": 9.904916244738571e-05, |
|
"loss": 1.3778, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.263653483992467, |
|
"grad_norm": 0.5232111215591431, |
|
"learning_rate": 9.896855443643308e-05, |
|
"loss": 1.3466, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.2711864406779661, |
|
"grad_norm": 0.4295203983783722, |
|
"learning_rate": 9.888470273039775e-05, |
|
"loss": 1.3928, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.2787193973634652, |
|
"grad_norm": 0.3830944895744324, |
|
"learning_rate": 9.879761288364766e-05, |
|
"loss": 1.4017, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.2862523540489642, |
|
"grad_norm": 0.3481033146381378, |
|
"learning_rate": 9.870729066504629e-05, |
|
"loss": 1.3344, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.2937853107344633, |
|
"grad_norm": 0.28261324763298035, |
|
"learning_rate": 9.861374205757068e-05, |
|
"loss": 1.2005, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.3013182674199623, |
|
"grad_norm": 0.31103572249412537, |
|
"learning_rate": 9.851697325791505e-05, |
|
"loss": 1.2026, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.3013182674199623, |
|
"eval_loss": 1.2615383863449097, |
|
"eval_runtime": 43.98, |
|
"eval_samples_per_second": 1.705, |
|
"eval_steps_per_second": 0.864, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.3088512241054614, |
|
"grad_norm": 0.41986966133117676, |
|
"learning_rate": 9.841699067608033e-05, |
|
"loss": 1.2891, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.3163841807909605, |
|
"grad_norm": 0.3304821848869324, |
|
"learning_rate": 9.831380093494957e-05, |
|
"loss": 1.1725, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.3239171374764595, |
|
"grad_norm": 0.38536179065704346, |
|
"learning_rate": 9.820741086984924e-05, |
|
"loss": 1.2084, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.3314500941619586, |
|
"grad_norm": 0.3935360908508301, |
|
"learning_rate": 9.809782752809644e-05, |
|
"loss": 1.2919, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.3389830508474576, |
|
"grad_norm": 0.3400743305683136, |
|
"learning_rate": 9.798505816853208e-05, |
|
"loss": 1.2405, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.3465160075329567, |
|
"grad_norm": 0.3844086825847626, |
|
"learning_rate": 9.786911026104007e-05, |
|
"loss": 1.2489, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.3540489642184557, |
|
"grad_norm": 0.3005635738372803, |
|
"learning_rate": 9.774999148605251e-05, |
|
"loss": 1.1564, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.3615819209039548, |
|
"grad_norm": 0.2923331558704376, |
|
"learning_rate": 9.762770973404094e-05, |
|
"loss": 1.1403, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.3691148775894539, |
|
"grad_norm": 0.30060315132141113, |
|
"learning_rate": 9.750227310499366e-05, |
|
"loss": 1.1172, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.3766478342749529, |
|
"grad_norm": 0.2783910036087036, |
|
"learning_rate": 9.737368990787916e-05, |
|
"loss": 1.0288, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.3766478342749529, |
|
"eval_loss": 1.122483491897583, |
|
"eval_runtime": 43.9499, |
|
"eval_samples_per_second": 1.706, |
|
"eval_steps_per_second": 0.865, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.384180790960452, |
|
"grad_norm": 0.2709607779979706, |
|
"learning_rate": 9.72419686600958e-05, |
|
"loss": 1.0827, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.391713747645951, |
|
"grad_norm": 0.27541661262512207, |
|
"learning_rate": 9.710711808690754e-05, |
|
"loss": 1.172, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.3992467043314501, |
|
"grad_norm": 0.25708749890327454, |
|
"learning_rate": 9.696914712086603e-05, |
|
"loss": 1.058, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.4067796610169492, |
|
"grad_norm": 0.29954883456230164, |
|
"learning_rate": 9.682806490121885e-05, |
|
"loss": 1.0781, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.4143126177024482, |
|
"grad_norm": 0.28139016032218933, |
|
"learning_rate": 9.668388077330421e-05, |
|
"loss": 1.0746, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.4218455743879473, |
|
"grad_norm": 0.2823486328125, |
|
"learning_rate": 9.653660428793188e-05, |
|
"loss": 1.052, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.4293785310734463, |
|
"grad_norm": 0.28060609102249146, |
|
"learning_rate": 9.638624520075046e-05, |
|
"loss": 1.0088, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.4369114877589454, |
|
"grad_norm": 0.26586028933525085, |
|
"learning_rate": 9.623281347160127e-05, |
|
"loss": 0.9935, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.4444444444444444, |
|
"grad_norm": 0.2685753405094147, |
|
"learning_rate": 9.607631926385859e-05, |
|
"loss": 1.002, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.4519774011299435, |
|
"grad_norm": 0.24896016716957092, |
|
"learning_rate": 9.591677294375636e-05, |
|
"loss": 0.9303, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.4519774011299435, |
|
"eval_loss": 0.9904586672782898, |
|
"eval_runtime": 43.9768, |
|
"eval_samples_per_second": 1.705, |
|
"eval_steps_per_second": 0.864, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.4595103578154426, |
|
"grad_norm": 0.2723137140274048, |
|
"learning_rate": 9.575418507970161e-05, |
|
"loss": 1.0173, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.4670433145009416, |
|
"grad_norm": 0.28442397713661194, |
|
"learning_rate": 9.558856644157432e-05, |
|
"loss": 0.9589, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.4745762711864407, |
|
"grad_norm": 0.25184300541877747, |
|
"learning_rate": 9.541992800001409e-05, |
|
"loss": 0.9013, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.4821092278719397, |
|
"grad_norm": 0.2699225842952728, |
|
"learning_rate": 9.52482809256934e-05, |
|
"loss": 0.9436, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.4896421845574388, |
|
"grad_norm": 0.24603509902954102, |
|
"learning_rate": 9.507363658857768e-05, |
|
"loss": 0.907, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.4971751412429379, |
|
"grad_norm": 0.24069125950336456, |
|
"learning_rate": 9.489600655717217e-05, |
|
"loss": 0.9147, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.504708097928437, |
|
"grad_norm": 0.2531817555427551, |
|
"learning_rate": 9.471540259775554e-05, |
|
"loss": 0.8429, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.512241054613936, |
|
"grad_norm": 0.29864469170570374, |
|
"learning_rate": 9.453183667360062e-05, |
|
"loss": 0.9324, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.519774011299435, |
|
"grad_norm": 0.28371143341064453, |
|
"learning_rate": 9.43453209441818e-05, |
|
"loss": 0.8185, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.527306967984934, |
|
"grad_norm": 0.3073464035987854, |
|
"learning_rate": 9.415586776436973e-05, |
|
"loss": 0.9043, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.527306967984934, |
|
"eval_loss": 0.8400442600250244, |
|
"eval_runtime": 43.8471, |
|
"eval_samples_per_second": 1.71, |
|
"eval_steps_per_second": 0.867, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.5348399246704332, |
|
"grad_norm": 0.31865832209587097, |
|
"learning_rate": 9.396348968361281e-05, |
|
"loss": 0.8984, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.5423728813559322, |
|
"grad_norm": 0.30962345004081726, |
|
"learning_rate": 9.376819944510598e-05, |
|
"loss": 0.8379, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.5499058380414312, |
|
"grad_norm": 0.40249139070510864, |
|
"learning_rate": 9.357000998494656e-05, |
|
"loss": 0.8553, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.5574387947269304, |
|
"grad_norm": 0.33023273944854736, |
|
"learning_rate": 9.336893443127738e-05, |
|
"loss": 0.762, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.5649717514124294, |
|
"grad_norm": 0.2891211211681366, |
|
"learning_rate": 9.31649861034172e-05, |
|
"loss": 0.7443, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.5725047080979284, |
|
"grad_norm": 0.33249810338020325, |
|
"learning_rate": 9.295817851097837e-05, |
|
"loss": 0.789, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.5800376647834274, |
|
"grad_norm": 0.32798218727111816, |
|
"learning_rate": 9.274852535297198e-05, |
|
"loss": 0.7544, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.5875706214689266, |
|
"grad_norm": 0.38781678676605225, |
|
"learning_rate": 9.253604051690046e-05, |
|
"loss": 0.691, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.5951035781544256, |
|
"grad_norm": 0.3911118805408478, |
|
"learning_rate": 9.232073807783759e-05, |
|
"loss": 0.6637, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.6026365348399246, |
|
"grad_norm": 0.31271758675575256, |
|
"learning_rate": 9.210263229749626e-05, |
|
"loss": 0.6649, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.6026365348399246, |
|
"eval_loss": 0.657122015953064, |
|
"eval_runtime": 43.9006, |
|
"eval_samples_per_second": 1.708, |
|
"eval_steps_per_second": 0.866, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.6101694915254238, |
|
"grad_norm": 0.3603556454181671, |
|
"learning_rate": 9.188173762328367e-05, |
|
"loss": 0.6501, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.6177024482109228, |
|
"grad_norm": 0.39841267466545105, |
|
"learning_rate": 9.165806868734444e-05, |
|
"loss": 0.6572, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.6252354048964218, |
|
"grad_norm": 0.34062060713768005, |
|
"learning_rate": 9.143164030559122e-05, |
|
"loss": 0.6706, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.632768361581921, |
|
"grad_norm": 0.31477442383766174, |
|
"learning_rate": 9.120246747672347e-05, |
|
"loss": 0.629, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.64030131826742, |
|
"grad_norm": 0.28587597608566284, |
|
"learning_rate": 9.097056538123376e-05, |
|
"loss": 0.5698, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.647834274952919, |
|
"grad_norm": 0.2795222997665405, |
|
"learning_rate": 9.073594938040231e-05, |
|
"loss": 0.573, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.655367231638418, |
|
"grad_norm": 0.23265816271305084, |
|
"learning_rate": 9.049863501527947e-05, |
|
"loss": 0.626, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.6629001883239172, |
|
"grad_norm": 0.26728200912475586, |
|
"learning_rate": 9.025863800565613e-05, |
|
"loss": 0.6103, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.6704331450094162, |
|
"grad_norm": 0.24294917285442352, |
|
"learning_rate": 9.001597424902267e-05, |
|
"loss": 0.5109, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.6779661016949152, |
|
"grad_norm": 0.27384141087532043, |
|
"learning_rate": 8.977065981951566e-05, |
|
"loss": 0.5266, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.6779661016949152, |
|
"eval_loss": 0.5219850540161133, |
|
"eval_runtime": 43.852, |
|
"eval_samples_per_second": 1.71, |
|
"eval_steps_per_second": 0.867, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.6854990583804144, |
|
"grad_norm": 0.2532135248184204, |
|
"learning_rate": 8.952271096685332e-05, |
|
"loss": 0.5027, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.6930320150659134, |
|
"grad_norm": 0.25869038701057434, |
|
"learning_rate": 8.927214411525895e-05, |
|
"loss": 0.5846, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.7005649717514124, |
|
"grad_norm": 0.25413447618484497, |
|
"learning_rate": 8.90189758623731e-05, |
|
"loss": 0.5171, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.7080979284369114, |
|
"grad_norm": 0.2756996154785156, |
|
"learning_rate": 8.876322297815405e-05, |
|
"loss": 0.5104, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.7156308851224106, |
|
"grad_norm": 0.25177237391471863, |
|
"learning_rate": 8.850490240376711e-05, |
|
"loss": 0.4921, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.7231638418079096, |
|
"grad_norm": 0.22342444956302643, |
|
"learning_rate": 8.824403125046225e-05, |
|
"loss": 0.45, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.7306967984934086, |
|
"grad_norm": 0.2055438607931137, |
|
"learning_rate": 8.798062679844077e-05, |
|
"loss": 0.4527, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.7382297551789078, |
|
"grad_norm": 0.21503853797912598, |
|
"learning_rate": 8.771470649571056e-05, |
|
"loss": 0.5282, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.7457627118644068, |
|
"grad_norm": 0.1799071580171585, |
|
"learning_rate": 8.744628795693047e-05, |
|
"loss": 0.5086, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.7532956685499058, |
|
"grad_norm": 0.20049640536308289, |
|
"learning_rate": 8.717538896224332e-05, |
|
"loss": 0.5187, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.7532956685499058, |
|
"eval_loss": 0.457345575094223, |
|
"eval_runtime": 43.8906, |
|
"eval_samples_per_second": 1.709, |
|
"eval_steps_per_second": 0.866, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.7608286252354048, |
|
"grad_norm": 0.19504646956920624, |
|
"learning_rate": 8.690202745609835e-05, |
|
"loss": 0.3967, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.768361581920904, |
|
"grad_norm": 0.18335066735744476, |
|
"learning_rate": 8.662622154606237e-05, |
|
"loss": 0.4931, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.775894538606403, |
|
"grad_norm": 0.2755972743034363, |
|
"learning_rate": 8.634798950162048e-05, |
|
"loss": 0.5392, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.783427495291902, |
|
"grad_norm": 0.18094103038311005, |
|
"learning_rate": 8.606734975296578e-05, |
|
"loss": 0.4579, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.7909604519774012, |
|
"grad_norm": 0.20168878138065338, |
|
"learning_rate": 8.578432088977859e-05, |
|
"loss": 0.3906, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.7984934086629002, |
|
"grad_norm": 0.20075590908527374, |
|
"learning_rate": 8.549892165999505e-05, |
|
"loss": 0.4453, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.8060263653483992, |
|
"grad_norm": 0.19419971108436584, |
|
"learning_rate": 8.521117096856528e-05, |
|
"loss": 0.4881, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.8135593220338984, |
|
"grad_norm": 0.25679534673690796, |
|
"learning_rate": 8.492108787620105e-05, |
|
"loss": 0.4528, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.8210922787193974, |
|
"grad_norm": 0.17266082763671875, |
|
"learning_rate": 8.462869159811327e-05, |
|
"loss": 0.4072, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.8286252354048964, |
|
"grad_norm": 0.20196932554244995, |
|
"learning_rate": 8.433400150273906e-05, |
|
"loss": 0.4506, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.8286252354048964, |
|
"eval_loss": 0.4128032326698303, |
|
"eval_runtime": 43.8756, |
|
"eval_samples_per_second": 1.709, |
|
"eval_steps_per_second": 0.866, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.8361581920903954, |
|
"grad_norm": 0.18674571812152863, |
|
"learning_rate": 8.403703711045892e-05, |
|
"loss": 0.4425, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.8436911487758946, |
|
"grad_norm": 0.19765950739383698, |
|
"learning_rate": 8.373781809230355e-05, |
|
"loss": 0.4913, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.8512241054613936, |
|
"grad_norm": 0.157321497797966, |
|
"learning_rate": 8.343636426865096e-05, |
|
"loss": 0.4163, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.8587570621468926, |
|
"grad_norm": 0.20251426100730896, |
|
"learning_rate": 8.313269560791342e-05, |
|
"loss": 0.3659, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.8662900188323918, |
|
"grad_norm": 0.16248001158237457, |
|
"learning_rate": 8.28268322252149e-05, |
|
"loss": 0.4309, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.8738229755178908, |
|
"grad_norm": 0.16724838316440582, |
|
"learning_rate": 8.251879438105854e-05, |
|
"loss": 0.4167, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.8813559322033898, |
|
"grad_norm": 0.20615121722221375, |
|
"learning_rate": 8.220860247998456e-05, |
|
"loss": 0.3559, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"grad_norm": 0.2065141499042511, |
|
"learning_rate": 8.189627706921877e-05, |
|
"loss": 0.3626, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.896421845574388, |
|
"grad_norm": 0.17293284833431244, |
|
"learning_rate": 8.15818388373114e-05, |
|
"loss": 0.4325, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.903954802259887, |
|
"grad_norm": 0.19183604419231415, |
|
"learning_rate": 8.126530861276677e-05, |
|
"loss": 0.3387, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.903954802259887, |
|
"eval_loss": 0.38071778416633606, |
|
"eval_runtime": 43.8754, |
|
"eval_samples_per_second": 1.709, |
|
"eval_steps_per_second": 0.866, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.911487758945386, |
|
"grad_norm": 0.1738622784614563, |
|
"learning_rate": 8.094670736266353e-05, |
|
"loss": 0.4149, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.9190207156308852, |
|
"grad_norm": 0.23153111338615417, |
|
"learning_rate": 8.062605619126584e-05, |
|
"loss": 0.4434, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.9265536723163842, |
|
"grad_norm": 0.26979267597198486, |
|
"learning_rate": 8.030337633862542e-05, |
|
"loss": 0.3639, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.9340866290018832, |
|
"grad_norm": 0.19837673008441925, |
|
"learning_rate": 7.997868917917453e-05, |
|
"loss": 0.3336, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.9416195856873822, |
|
"grad_norm": 0.18979325890541077, |
|
"learning_rate": 7.965201622031021e-05, |
|
"loss": 0.3715, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.9491525423728814, |
|
"grad_norm": 0.20212630927562714, |
|
"learning_rate": 7.932337910096961e-05, |
|
"loss": 0.333, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.9566854990583804, |
|
"grad_norm": 0.22705507278442383, |
|
"learning_rate": 7.899279959019654e-05, |
|
"loss": 0.3456, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.9642184557438794, |
|
"grad_norm": 0.24515411257743835, |
|
"learning_rate": 7.866029958569956e-05, |
|
"loss": 0.3336, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.9717514124293786, |
|
"grad_norm": 0.2214023470878601, |
|
"learning_rate": 7.832590111240145e-05, |
|
"loss": 0.3951, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.9792843691148776, |
|
"grad_norm": 0.20626534521579742, |
|
"learning_rate": 7.798962632098024e-05, |
|
"loss": 0.3636, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.9792843691148776, |
|
"eval_loss": 0.3497096598148346, |
|
"eval_runtime": 43.8538, |
|
"eval_samples_per_second": 1.71, |
|
"eval_steps_per_second": 0.867, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.9868173258003766, |
|
"grad_norm": 0.21280796825885773, |
|
"learning_rate": 7.765149748640197e-05, |
|
"loss": 0.2931, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.9943502824858758, |
|
"grad_norm": 0.18561464548110962, |
|
"learning_rate": 7.73115370064452e-05, |
|
"loss": 0.3363, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 1.0075329566854991, |
|
"grad_norm": 0.3493395447731018, |
|
"learning_rate": 7.696976740021733e-05, |
|
"loss": 0.7106, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 1.015065913370998, |
|
"grad_norm": 0.24458995461463928, |
|
"learning_rate": 7.6626211306663e-05, |
|
"loss": 0.2806, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 1.0225988700564972, |
|
"grad_norm": 0.2562405467033386, |
|
"learning_rate": 7.628089148306434e-05, |
|
"loss": 0.3048, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.0301318267419963, |
|
"grad_norm": 0.22578994929790497, |
|
"learning_rate": 7.59338308035337e-05, |
|
"loss": 0.3142, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 1.0376647834274952, |
|
"grad_norm": 0.20230619609355927, |
|
"learning_rate": 7.558505225749827e-05, |
|
"loss": 0.3104, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 1.0451977401129944, |
|
"grad_norm": 0.22089356184005737, |
|
"learning_rate": 7.523457894817745e-05, |
|
"loss": 0.2626, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 1.0527306967984935, |
|
"grad_norm": 0.23944571614265442, |
|
"learning_rate": 7.488243409105233e-05, |
|
"loss": 0.2984, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 1.0602636534839924, |
|
"grad_norm": 0.2213144749403, |
|
"learning_rate": 7.452864101232798e-05, |
|
"loss": 0.3714, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.0602636534839924, |
|
"eval_loss": 0.31596410274505615, |
|
"eval_runtime": 43.9549, |
|
"eval_samples_per_second": 1.706, |
|
"eval_steps_per_second": 0.865, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.0677966101694916, |
|
"grad_norm": 0.2673434019088745, |
|
"learning_rate": 7.417322314738822e-05, |
|
"loss": 0.3455, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 1.0753295668549905, |
|
"grad_norm": 0.25528621673583984, |
|
"learning_rate": 7.381620403924333e-05, |
|
"loss": 0.311, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 1.0828625235404896, |
|
"grad_norm": 0.3240588903427124, |
|
"learning_rate": 7.345760733697055e-05, |
|
"loss": 0.3363, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 1.0903954802259888, |
|
"grad_norm": 0.31120482087135315, |
|
"learning_rate": 7.30974567941475e-05, |
|
"loss": 0.3529, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 1.0979284369114877, |
|
"grad_norm": 0.2683728039264679, |
|
"learning_rate": 7.273577626727884e-05, |
|
"loss": 0.2926, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.1054613935969868, |
|
"grad_norm": 0.2490655928850174, |
|
"learning_rate": 7.237258971421587e-05, |
|
"loss": 0.3029, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 1.112994350282486, |
|
"grad_norm": 0.2618696093559265, |
|
"learning_rate": 7.20079211925696e-05, |
|
"loss": 0.278, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 1.1205273069679849, |
|
"grad_norm": 0.25481289625167847, |
|
"learning_rate": 7.164179485811727e-05, |
|
"loss": 0.2936, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 1.128060263653484, |
|
"grad_norm": 0.36896899342536926, |
|
"learning_rate": 7.127423496320212e-05, |
|
"loss": 0.2876, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 1.1355932203389831, |
|
"grad_norm": 0.3275793194770813, |
|
"learning_rate": 7.090526585512696e-05, |
|
"loss": 0.3279, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.1355932203389831, |
|
"eval_loss": 0.2823197543621063, |
|
"eval_runtime": 43.9363, |
|
"eval_samples_per_second": 1.707, |
|
"eval_steps_per_second": 0.865, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.143126177024482, |
|
"grad_norm": 0.263179749250412, |
|
"learning_rate": 7.053491197454142e-05, |
|
"loss": 0.2342, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 1.1506591337099812, |
|
"grad_norm": 0.2754746377468109, |
|
"learning_rate": 7.016319785382296e-05, |
|
"loss": 0.3234, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 1.1581920903954803, |
|
"grad_norm": 0.29197824001312256, |
|
"learning_rate": 6.979014811545189e-05, |
|
"loss": 0.2458, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 1.1657250470809792, |
|
"grad_norm": 0.45912623405456543, |
|
"learning_rate": 6.941578747038023e-05, |
|
"loss": 0.2357, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 1.1732580037664784, |
|
"grad_norm": 0.36572179198265076, |
|
"learning_rate": 6.904014071639503e-05, |
|
"loss": 0.283, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 1.1807909604519775, |
|
"grad_norm": 0.29460588097572327, |
|
"learning_rate": 6.866323273647563e-05, |
|
"loss": 0.2653, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 1.1883239171374764, |
|
"grad_norm": 0.2520604729652405, |
|
"learning_rate": 6.828508849714546e-05, |
|
"loss": 0.3155, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 1.1958568738229756, |
|
"grad_norm": 0.2615698575973511, |
|
"learning_rate": 6.79057330468182e-05, |
|
"loss": 0.1852, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 1.2033898305084745, |
|
"grad_norm": 0.30008143186569214, |
|
"learning_rate": 6.752519151413861e-05, |
|
"loss": 0.2543, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 1.2109227871939736, |
|
"grad_norm": 0.3347761929035187, |
|
"learning_rate": 6.7143489106318e-05, |
|
"loss": 0.3602, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.2109227871939736, |
|
"eval_loss": 0.25164613127708435, |
|
"eval_runtime": 44.1497, |
|
"eval_samples_per_second": 1.699, |
|
"eval_steps_per_second": 0.861, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.2184557438794728, |
|
"grad_norm": 0.3353792726993561, |
|
"learning_rate": 6.676065110746444e-05, |
|
"loss": 0.2489, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 1.2259887005649717, |
|
"grad_norm": 0.32569241523742676, |
|
"learning_rate": 6.637670287690799e-05, |
|
"loss": 0.194, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 1.2335216572504708, |
|
"grad_norm": 0.336117684841156, |
|
"learning_rate": 6.599166984752087e-05, |
|
"loss": 0.1816, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 1.24105461393597, |
|
"grad_norm": 0.37371379137039185, |
|
"learning_rate": 6.560557752403277e-05, |
|
"loss": 0.2807, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 1.2485875706214689, |
|
"grad_norm": 0.2796350419521332, |
|
"learning_rate": 6.52184514813414e-05, |
|
"loss": 0.2149, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 1.256120527306968, |
|
"grad_norm": 0.4764145314693451, |
|
"learning_rate": 6.483031736281843e-05, |
|
"loss": 0.1903, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 1.2636534839924671, |
|
"grad_norm": 0.36564409732818604, |
|
"learning_rate": 6.444120087861081e-05, |
|
"loss": 0.2638, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 1.271186440677966, |
|
"grad_norm": 0.34570133686065674, |
|
"learning_rate": 6.40511278039378e-05, |
|
"loss": 0.2568, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 1.2787193973634652, |
|
"grad_norm": 0.3726440966129303, |
|
"learning_rate": 6.366012397738355e-05, |
|
"loss": 0.2396, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 1.286252354048964, |
|
"grad_norm": 0.22987572848796844, |
|
"learning_rate": 6.326821529918553e-05, |
|
"loss": 0.2168, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.286252354048964, |
|
"eval_loss": 0.22047367691993713, |
|
"eval_runtime": 43.9518, |
|
"eval_samples_per_second": 1.706, |
|
"eval_steps_per_second": 0.865, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.2937853107344632, |
|
"grad_norm": 0.38589489459991455, |
|
"learning_rate": 6.287542772951897e-05, |
|
"loss": 0.2671, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 1.3013182674199624, |
|
"grad_norm": 0.26084208488464355, |
|
"learning_rate": 6.248178728677711e-05, |
|
"loss": 0.2628, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 1.3088512241054615, |
|
"grad_norm": 0.3167526423931122, |
|
"learning_rate": 6.208732004584791e-05, |
|
"loss": 0.2128, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 1.3163841807909604, |
|
"grad_norm": 0.3979116976261139, |
|
"learning_rate": 6.16920521363867e-05, |
|
"loss": 0.217, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 1.3239171374764596, |
|
"grad_norm": 0.3168756663799286, |
|
"learning_rate": 6.129600974108538e-05, |
|
"loss": 0.2284, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 1.3314500941619585, |
|
"grad_norm": 0.30765026807785034, |
|
"learning_rate": 6.089921909393812e-05, |
|
"loss": 0.2698, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 1.3389830508474576, |
|
"grad_norm": 0.3305312991142273, |
|
"learning_rate": 6.050170647850351e-05, |
|
"loss": 0.257, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 1.3465160075329567, |
|
"grad_norm": 0.6542951464653015, |
|
"learning_rate": 6.0103498226163603e-05, |
|
"loss": 0.2666, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 1.3540489642184557, |
|
"grad_norm": 0.3485718071460724, |
|
"learning_rate": 5.970462071437973e-05, |
|
"loss": 0.1579, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 1.3615819209039548, |
|
"grad_norm": 0.4073808789253235, |
|
"learning_rate": 5.93051003649452e-05, |
|
"loss": 0.1643, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.3615819209039548, |
|
"eval_loss": 0.19270579516887665, |
|
"eval_runtime": 44.023, |
|
"eval_samples_per_second": 1.704, |
|
"eval_steps_per_second": 0.863, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.369114877589454, |
|
"grad_norm": 0.37884190678596497, |
|
"learning_rate": 5.890496364223509e-05, |
|
"loss": 0.2015, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 1.3766478342749529, |
|
"grad_norm": 0.2823837101459503, |
|
"learning_rate": 5.850423705145334e-05, |
|
"loss": 0.1774, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 1.384180790960452, |
|
"grad_norm": 0.33302009105682373, |
|
"learning_rate": 5.8102947136876876e-05, |
|
"loss": 0.2202, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 1.3917137476459511, |
|
"grad_norm": 0.30890604853630066, |
|
"learning_rate": 5.770112048009747e-05, |
|
"loss": 0.1372, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 1.39924670433145, |
|
"grad_norm": 0.3033400774002075, |
|
"learning_rate": 5.7298783698260874e-05, |
|
"loss": 0.1857, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 1.4067796610169492, |
|
"grad_norm": 0.2779131829738617, |
|
"learning_rate": 5.68959634423037e-05, |
|
"loss": 0.187, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 1.414312617702448, |
|
"grad_norm": 0.485245943069458, |
|
"learning_rate": 5.64926863951881e-05, |
|
"loss": 0.2437, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 1.4218455743879472, |
|
"grad_norm": 0.37200334668159485, |
|
"learning_rate": 5.60889792701342e-05, |
|
"loss": 0.1923, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 1.4293785310734464, |
|
"grad_norm": 0.32306909561157227, |
|
"learning_rate": 5.568486880885068e-05, |
|
"loss": 0.1713, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 1.4369114877589455, |
|
"grad_norm": 0.3483230471611023, |
|
"learning_rate": 5.52803817797633e-05, |
|
"loss": 0.1336, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.4369114877589455, |
|
"eval_loss": 0.17033681273460388, |
|
"eval_runtime": 43.9231, |
|
"eval_samples_per_second": 1.708, |
|
"eval_steps_per_second": 0.865, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.4444444444444444, |
|
"grad_norm": 0.3841436505317688, |
|
"learning_rate": 5.487554497624189e-05, |
|
"loss": 0.1797, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 1.4519774011299436, |
|
"grad_norm": 0.3441992998123169, |
|
"learning_rate": 5.4470385214825416e-05, |
|
"loss": 0.2176, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 1.4595103578154425, |
|
"grad_norm": 0.5087612867355347, |
|
"learning_rate": 5.406492933344571e-05, |
|
"loss": 0.2663, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 1.4670433145009416, |
|
"grad_norm": 0.3248097598552704, |
|
"learning_rate": 5.365920418964973e-05, |
|
"loss": 0.1739, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 1.4745762711864407, |
|
"grad_norm": 0.4247044622898102, |
|
"learning_rate": 5.3253236658820396e-05, |
|
"loss": 0.1454, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 1.4821092278719397, |
|
"grad_norm": 0.2898694574832916, |
|
"learning_rate": 5.28470536323965e-05, |
|
"loss": 0.1609, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 1.4896421845574388, |
|
"grad_norm": 0.41251513361930847, |
|
"learning_rate": 5.244068201609133e-05, |
|
"loss": 0.143, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 1.497175141242938, |
|
"grad_norm": 0.525326132774353, |
|
"learning_rate": 5.2034148728110424e-05, |
|
"loss": 0.1938, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 1.5047080979284368, |
|
"grad_norm": 0.36119720339775085, |
|
"learning_rate": 5.162748069736851e-05, |
|
"loss": 0.1518, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 1.512241054613936, |
|
"grad_norm": 0.38599205017089844, |
|
"learning_rate": 5.1220704861705774e-05, |
|
"loss": 0.1198, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.512241054613936, |
|
"eval_loss": 0.15222826600074768, |
|
"eval_runtime": 44.0589, |
|
"eval_samples_per_second": 1.702, |
|
"eval_steps_per_second": 0.862, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.5197740112994351, |
|
"grad_norm": 0.36092710494995117, |
|
"learning_rate": 5.081384816610336e-05, |
|
"loss": 0.1485, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 1.527306967984934, |
|
"grad_norm": 0.3418121039867401, |
|
"learning_rate": 5.0406937560898646e-05, |
|
"loss": 0.1368, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 1.5348399246704332, |
|
"grad_norm": 0.39115968346595764, |
|
"learning_rate": 5e-05, |
|
"loss": 0.1843, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 1.542372881355932, |
|
"grad_norm": 0.4022577106952667, |
|
"learning_rate": 4.9593062439101365e-05, |
|
"loss": 0.128, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 1.5499058380414312, |
|
"grad_norm": 0.3265657126903534, |
|
"learning_rate": 4.918615183389665e-05, |
|
"loss": 0.1365, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 1.5574387947269304, |
|
"grad_norm": 0.32065853476524353, |
|
"learning_rate": 4.877929513829424e-05, |
|
"loss": 0.1495, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 1.5649717514124295, |
|
"grad_norm": 0.2755489647388458, |
|
"learning_rate": 4.8372519302631486e-05, |
|
"loss": 0.1105, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 1.5725047080979284, |
|
"grad_norm": 0.5003211498260498, |
|
"learning_rate": 4.796585127188958e-05, |
|
"loss": 0.1284, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 1.5800376647834273, |
|
"grad_norm": 0.43929675221443176, |
|
"learning_rate": 4.755931798390867e-05, |
|
"loss": 0.1889, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 1.5875706214689265, |
|
"grad_norm": 0.2825930118560791, |
|
"learning_rate": 4.715294636760352e-05, |
|
"loss": 0.1364, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.5875706214689265, |
|
"eval_loss": 0.13791552186012268, |
|
"eval_runtime": 43.9041, |
|
"eval_samples_per_second": 1.708, |
|
"eval_steps_per_second": 0.866, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.5951035781544256, |
|
"grad_norm": 0.3584269881248474, |
|
"learning_rate": 4.674676334117962e-05, |
|
"loss": 0.1261, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 1.6026365348399247, |
|
"grad_norm": 0.3463616967201233, |
|
"learning_rate": 4.634079581035029e-05, |
|
"loss": 0.185, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 1.6101694915254239, |
|
"grad_norm": 0.3442526161670685, |
|
"learning_rate": 4.59350706665543e-05, |
|
"loss": 0.104, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 1.6177024482109228, |
|
"grad_norm": 0.3831660747528076, |
|
"learning_rate": 4.55296147851746e-05, |
|
"loss": 0.1294, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 1.6252354048964217, |
|
"grad_norm": 0.3194776475429535, |
|
"learning_rate": 4.512445502375813e-05, |
|
"loss": 0.1068, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 1.6327683615819208, |
|
"grad_norm": 0.3435104787349701, |
|
"learning_rate": 4.471961822023671e-05, |
|
"loss": 0.0864, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 1.64030131826742, |
|
"grad_norm": 0.2775423526763916, |
|
"learning_rate": 4.431513119114934e-05, |
|
"loss": 0.1858, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 1.6478342749529191, |
|
"grad_norm": 0.3337323069572449, |
|
"learning_rate": 4.391102072986581e-05, |
|
"loss": 0.1946, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 1.655367231638418, |
|
"grad_norm": 0.32516106963157654, |
|
"learning_rate": 4.350731360481191e-05, |
|
"loss": 0.1014, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 1.6629001883239172, |
|
"grad_norm": 0.2920524775981903, |
|
"learning_rate": 4.3104036557696295e-05, |
|
"loss": 0.2689, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.6629001883239172, |
|
"eval_loss": 0.12488219887018204, |
|
"eval_runtime": 43.9386, |
|
"eval_samples_per_second": 1.707, |
|
"eval_steps_per_second": 0.865, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.670433145009416, |
|
"grad_norm": 0.4062795042991638, |
|
"learning_rate": 4.270121630173913e-05, |
|
"loss": 0.1495, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 1.6779661016949152, |
|
"grad_norm": 0.30497950315475464, |
|
"learning_rate": 4.229887951990255e-05, |
|
"loss": 0.1369, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 1.6854990583804144, |
|
"grad_norm": 0.26471954584121704, |
|
"learning_rate": 4.189705286312314e-05, |
|
"loss": 0.0776, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 1.6930320150659135, |
|
"grad_norm": 0.22838452458381653, |
|
"learning_rate": 4.149576294854668e-05, |
|
"loss": 0.0942, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.7005649717514124, |
|
"grad_norm": 0.2828744649887085, |
|
"learning_rate": 4.1095036357764915e-05, |
|
"loss": 0.0787, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.7080979284369113, |
|
"grad_norm": 0.3124982714653015, |
|
"learning_rate": 4.069489963505482e-05, |
|
"loss": 0.13, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 1.7156308851224105, |
|
"grad_norm": 0.3018610179424286, |
|
"learning_rate": 4.029537928562028e-05, |
|
"loss": 0.1263, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 1.7231638418079096, |
|
"grad_norm": 0.28386422991752625, |
|
"learning_rate": 3.98965017738364e-05, |
|
"loss": 0.089, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.7306967984934087, |
|
"grad_norm": 0.24454158544540405, |
|
"learning_rate": 3.9498293521496503e-05, |
|
"loss": 0.1525, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 1.7382297551789079, |
|
"grad_norm": 0.29519349336624146, |
|
"learning_rate": 3.9100780906061896e-05, |
|
"loss": 0.1383, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.7382297551789079, |
|
"eval_loss": 0.11655016243457794, |
|
"eval_runtime": 43.924, |
|
"eval_samples_per_second": 1.707, |
|
"eval_steps_per_second": 0.865, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.7457627118644068, |
|
"grad_norm": 0.2894473075866699, |
|
"learning_rate": 3.8703990258914614e-05, |
|
"loss": 0.1709, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 1.7532956685499057, |
|
"grad_norm": 0.31395775079727173, |
|
"learning_rate": 3.83079478636133e-05, |
|
"loss": 0.0944, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.7608286252354048, |
|
"grad_norm": 0.2754004895687103, |
|
"learning_rate": 3.791267995415208e-05, |
|
"loss": 0.1075, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 1.768361581920904, |
|
"grad_norm": 0.27213406562805176, |
|
"learning_rate": 3.7518212713222906e-05, |
|
"loss": 0.1264, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 1.7758945386064031, |
|
"grad_norm": 0.44370901584625244, |
|
"learning_rate": 3.7124572270481056e-05, |
|
"loss": 0.1066, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.783427495291902, |
|
"grad_norm": 0.27574431896209717, |
|
"learning_rate": 3.673178470081448e-05, |
|
"loss": 0.1138, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 1.7909604519774012, |
|
"grad_norm": 0.25503936409950256, |
|
"learning_rate": 3.633987602261647e-05, |
|
"loss": 0.0913, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 1.7984934086629, |
|
"grad_norm": 0.33877211809158325, |
|
"learning_rate": 3.594887219606221e-05, |
|
"loss": 0.1084, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 1.8060263653483992, |
|
"grad_norm": 0.3899424374103546, |
|
"learning_rate": 3.55587991213892e-05, |
|
"loss": 0.203, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 1.8135593220338984, |
|
"grad_norm": 0.3390370011329651, |
|
"learning_rate": 3.516968263718159e-05, |
|
"loss": 0.1145, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.8135593220338984, |
|
"eval_loss": 0.11021808534860611, |
|
"eval_runtime": 43.9451, |
|
"eval_samples_per_second": 1.707, |
|
"eval_steps_per_second": 0.865, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.8210922787193975, |
|
"grad_norm": 0.33010998368263245, |
|
"learning_rate": 3.47815485186586e-05, |
|
"loss": 0.0726, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 1.8286252354048964, |
|
"grad_norm": 0.280018150806427, |
|
"learning_rate": 3.439442247596724e-05, |
|
"loss": 0.1107, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 1.8361581920903953, |
|
"grad_norm": 0.2712264657020569, |
|
"learning_rate": 3.400833015247913e-05, |
|
"loss": 0.0737, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 1.8436911487758945, |
|
"grad_norm": 0.2368086278438568, |
|
"learning_rate": 3.3623297123092006e-05, |
|
"loss": 0.1474, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.8512241054613936, |
|
"grad_norm": 0.29197973012924194, |
|
"learning_rate": 3.323934889253556e-05, |
|
"loss": 0.078, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 1.8587570621468927, |
|
"grad_norm": 0.2464105784893036, |
|
"learning_rate": 3.285651089368202e-05, |
|
"loss": 0.1029, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 1.8662900188323919, |
|
"grad_norm": 0.24727390706539154, |
|
"learning_rate": 3.2474808485861397e-05, |
|
"loss": 0.0552, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 1.8738229755178908, |
|
"grad_norm": 0.23096492886543274, |
|
"learning_rate": 3.209426695318182e-05, |
|
"loss": 0.0818, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.8813559322033897, |
|
"grad_norm": 0.23165109753608704, |
|
"learning_rate": 3.171491150285456e-05, |
|
"loss": 0.1493, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 1.8888888888888888, |
|
"grad_norm": 0.3371273875236511, |
|
"learning_rate": 3.133676726352438e-05, |
|
"loss": 0.1037, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.8888888888888888, |
|
"eval_loss": 0.10490331053733826, |
|
"eval_runtime": 43.9307, |
|
"eval_samples_per_second": 1.707, |
|
"eval_steps_per_second": 0.865, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.896421845574388, |
|
"grad_norm": 0.2399485558271408, |
|
"learning_rate": 3.0959859283604984e-05, |
|
"loss": 0.0929, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 1.9039548022598871, |
|
"grad_norm": 0.26773276925086975, |
|
"learning_rate": 3.0584212529619775e-05, |
|
"loss": 0.0637, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 1.911487758945386, |
|
"grad_norm": 0.41889140009880066, |
|
"learning_rate": 3.0209851884548117e-05, |
|
"loss": 0.1263, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 1.9190207156308852, |
|
"grad_norm": 0.2959991991519928, |
|
"learning_rate": 2.9836802146177034e-05, |
|
"loss": 0.1339, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 1.926553672316384, |
|
"grad_norm": 0.24563738703727722, |
|
"learning_rate": 2.9465088025458586e-05, |
|
"loss": 0.1496, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 1.9340866290018832, |
|
"grad_norm": 0.24098043143749237, |
|
"learning_rate": 2.9094734144873036e-05, |
|
"loss": 0.0539, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 1.9416195856873824, |
|
"grad_norm": 0.3055378198623657, |
|
"learning_rate": 2.8725765036797892e-05, |
|
"loss": 0.1087, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 1.9491525423728815, |
|
"grad_norm": 0.27689605951309204, |
|
"learning_rate": 2.835820514188273e-05, |
|
"loss": 0.1004, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 1.9566854990583804, |
|
"grad_norm": 0.30273178219795227, |
|
"learning_rate": 2.7992078807430422e-05, |
|
"loss": 0.1261, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 1.9642184557438793, |
|
"grad_norm": 0.3399432897567749, |
|
"learning_rate": 2.7627410285784163e-05, |
|
"loss": 0.1318, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.9642184557438793, |
|
"eval_loss": 0.10169798880815506, |
|
"eval_runtime": 43.9845, |
|
"eval_samples_per_second": 1.705, |
|
"eval_steps_per_second": 0.864, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.9717514124293785, |
|
"grad_norm": 0.2374039739370346, |
|
"learning_rate": 2.7264223732721167e-05, |
|
"loss": 0.0795, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 1.9792843691148776, |
|
"grad_norm": 0.2347286343574524, |
|
"learning_rate": 2.6902543205852492e-05, |
|
"loss": 0.1129, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 1.9868173258003767, |
|
"grad_norm": 0.2325945943593979, |
|
"learning_rate": 2.6542392663029463e-05, |
|
"loss": 0.117, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 1.9943502824858759, |
|
"grad_norm": 0.2451542764902115, |
|
"learning_rate": 2.618379596075668e-05, |
|
"loss": 0.0742, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 2.007532956685499, |
|
"grad_norm": 0.4711216390132904, |
|
"learning_rate": 2.582677685261179e-05, |
|
"loss": 0.2505, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 2.0150659133709983, |
|
"grad_norm": 0.25324326753616333, |
|
"learning_rate": 2.5471358987672017e-05, |
|
"loss": 0.0925, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 2.022598870056497, |
|
"grad_norm": 0.24556219577789307, |
|
"learning_rate": 2.511756590894765e-05, |
|
"loss": 0.069, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 2.030131826741996, |
|
"grad_norm": 0.20775027573108673, |
|
"learning_rate": 2.476542105182254e-05, |
|
"loss": 0.0687, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 2.0376647834274952, |
|
"grad_norm": 0.2542414367198944, |
|
"learning_rate": 2.4414947742501744e-05, |
|
"loss": 0.0769, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 2.0451977401129944, |
|
"grad_norm": 0.26574474573135376, |
|
"learning_rate": 2.4066169196466326e-05, |
|
"loss": 0.0834, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.0451977401129944, |
|
"eval_loss": 0.0993381068110466, |
|
"eval_runtime": 43.9044, |
|
"eval_samples_per_second": 1.708, |
|
"eval_steps_per_second": 0.866, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.0527306967984935, |
|
"grad_norm": 0.3231644034385681, |
|
"learning_rate": 2.3719108516935683e-05, |
|
"loss": 0.0774, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 2.0602636534839927, |
|
"grad_norm": 0.22078053653240204, |
|
"learning_rate": 2.3373788693337024e-05, |
|
"loss": 0.0884, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 2.0677966101694913, |
|
"grad_norm": 0.22590897977352142, |
|
"learning_rate": 2.303023259978267e-05, |
|
"loss": 0.0781, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 2.0753295668549905, |
|
"grad_norm": 0.23885948956012726, |
|
"learning_rate": 2.268846299355481e-05, |
|
"loss": 0.0543, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 2.0828625235404896, |
|
"grad_norm": 0.2120169848203659, |
|
"learning_rate": 2.2348502513598035e-05, |
|
"loss": 0.0925, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 2.0903954802259888, |
|
"grad_norm": 0.2760743498802185, |
|
"learning_rate": 2.2010373679019776e-05, |
|
"loss": 0.1427, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 2.097928436911488, |
|
"grad_norm": 0.3039568066596985, |
|
"learning_rate": 2.167409888759856e-05, |
|
"loss": 0.1766, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 2.105461393596987, |
|
"grad_norm": 0.2834893763065338, |
|
"learning_rate": 2.133970041430044e-05, |
|
"loss": 0.0816, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 2.1129943502824857, |
|
"grad_norm": 0.2928575277328491, |
|
"learning_rate": 2.1007200409803462e-05, |
|
"loss": 0.1076, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 2.120527306967985, |
|
"grad_norm": 0.2669938802719116, |
|
"learning_rate": 2.067662089903039e-05, |
|
"loss": 0.0889, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.120527306967985, |
|
"eval_loss": 0.09731351584196091, |
|
"eval_runtime": 44.078, |
|
"eval_samples_per_second": 1.702, |
|
"eval_steps_per_second": 0.862, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.128060263653484, |
|
"grad_norm": 0.28592410683631897, |
|
"learning_rate": 2.03479837796898e-05, |
|
"loss": 0.0685, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 2.135593220338983, |
|
"grad_norm": 0.32021915912628174, |
|
"learning_rate": 2.002131082082549e-05, |
|
"loss": 0.0927, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 2.1431261770244823, |
|
"grad_norm": 0.2118028998374939, |
|
"learning_rate": 1.9696623661374618e-05, |
|
"loss": 0.0634, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 2.150659133709981, |
|
"grad_norm": 0.22075195610523224, |
|
"learning_rate": 1.937394380873418e-05, |
|
"loss": 0.1464, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 2.15819209039548, |
|
"grad_norm": 0.24222980439662933, |
|
"learning_rate": 1.905329263733649e-05, |
|
"loss": 0.1251, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 2.1657250470809792, |
|
"grad_norm": 0.28131377696990967, |
|
"learning_rate": 1.873469138723325e-05, |
|
"loss": 0.1184, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 2.1732580037664784, |
|
"grad_norm": 0.22016428411006927, |
|
"learning_rate": 1.8418161162688615e-05, |
|
"loss": 0.0399, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 2.1807909604519775, |
|
"grad_norm": 0.19806838035583496, |
|
"learning_rate": 1.8103722930781247e-05, |
|
"loss": 0.0756, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 2.1883239171374766, |
|
"grad_norm": 0.332441121339798, |
|
"learning_rate": 1.779139752001545e-05, |
|
"loss": 0.0938, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 2.1958568738229753, |
|
"grad_norm": 0.19058367609977722, |
|
"learning_rate": 1.748120561894147e-05, |
|
"loss": 0.0643, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.1958568738229753, |
|
"eval_loss": 0.09646200388669968, |
|
"eval_runtime": 43.99, |
|
"eval_samples_per_second": 1.705, |
|
"eval_steps_per_second": 0.864, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.2033898305084745, |
|
"grad_norm": 0.168031245470047, |
|
"learning_rate": 1.7173167774785092e-05, |
|
"loss": 0.0524, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 2.2109227871939736, |
|
"grad_norm": 0.23299571871757507, |
|
"learning_rate": 1.6867304392086575e-05, |
|
"loss": 0.0775, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 2.2184557438794728, |
|
"grad_norm": 0.258078008890152, |
|
"learning_rate": 1.6563635731349057e-05, |
|
"loss": 0.0662, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 2.225988700564972, |
|
"grad_norm": 0.21041317284107208, |
|
"learning_rate": 1.6262181907696454e-05, |
|
"loss": 0.1229, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 2.2335216572504706, |
|
"grad_norm": 0.1758544147014618, |
|
"learning_rate": 1.5962962889541105e-05, |
|
"loss": 0.0487, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 2.2410546139359697, |
|
"grad_norm": 0.24289867281913757, |
|
"learning_rate": 1.5665998497260958e-05, |
|
"loss": 0.103, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 2.248587570621469, |
|
"grad_norm": 0.41178449988365173, |
|
"learning_rate": 1.5371308401886757e-05, |
|
"loss": 0.1671, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 2.256120527306968, |
|
"grad_norm": 0.22324170172214508, |
|
"learning_rate": 1.5078912123798961e-05, |
|
"loss": 0.089, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 2.263653483992467, |
|
"grad_norm": 0.2477371245622635, |
|
"learning_rate": 1.4788829031434732e-05, |
|
"loss": 0.0397, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 2.2711864406779663, |
|
"grad_norm": 0.21119807660579681, |
|
"learning_rate": 1.4501078340004953e-05, |
|
"loss": 0.0677, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.2711864406779663, |
|
"eval_loss": 0.09540043771266937, |
|
"eval_runtime": 43.9302, |
|
"eval_samples_per_second": 1.707, |
|
"eval_steps_per_second": 0.865, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.2787193973634654, |
|
"grad_norm": 0.24356690049171448, |
|
"learning_rate": 1.4215679110221413e-05, |
|
"loss": 0.1493, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 2.286252354048964, |
|
"grad_norm": 0.180653914809227, |
|
"learning_rate": 1.3932650247034218e-05, |
|
"loss": 0.0703, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 2.2937853107344632, |
|
"grad_norm": 0.3206869959831238, |
|
"learning_rate": 1.3652010498379519e-05, |
|
"loss": 0.1167, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 2.3013182674199624, |
|
"grad_norm": 0.21196141839027405, |
|
"learning_rate": 1.337377845393763e-05, |
|
"loss": 0.0721, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 2.3088512241054615, |
|
"grad_norm": 0.34639954566955566, |
|
"learning_rate": 1.309797254390167e-05, |
|
"loss": 0.1114, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 2.3163841807909606, |
|
"grad_norm": 0.28393760323524475, |
|
"learning_rate": 1.2824611037756684e-05, |
|
"loss": 0.0527, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 2.3239171374764593, |
|
"grad_norm": 0.2436118721961975, |
|
"learning_rate": 1.255371204306956e-05, |
|
"loss": 0.0825, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 2.3314500941619585, |
|
"grad_norm": 0.2877277135848999, |
|
"learning_rate": 1.2285293504289447e-05, |
|
"loss": 0.0721, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 2.3389830508474576, |
|
"grad_norm": 0.18935468792915344, |
|
"learning_rate": 1.2019373201559247e-05, |
|
"loss": 0.0826, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 2.3465160075329567, |
|
"grad_norm": 0.3491658568382263, |
|
"learning_rate": 1.1755968749537754e-05, |
|
"loss": 0.1784, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.3465160075329567, |
|
"eval_loss": 0.09475459903478622, |
|
"eval_runtime": 43.9621, |
|
"eval_samples_per_second": 1.706, |
|
"eval_steps_per_second": 0.864, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.354048964218456, |
|
"grad_norm": 0.31381261348724365, |
|
"learning_rate": 1.1495097596232901e-05, |
|
"loss": 0.1266, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 2.361581920903955, |
|
"grad_norm": 0.23978573083877563, |
|
"learning_rate": 1.1236777021845956e-05, |
|
"loss": 0.1047, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 2.3691148775894537, |
|
"grad_norm": 0.24144862592220306, |
|
"learning_rate": 1.0981024137626922e-05, |
|
"loss": 0.0572, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 2.376647834274953, |
|
"grad_norm": 0.21653713285923004, |
|
"learning_rate": 1.0727855884741056e-05, |
|
"loss": 0.0672, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 2.384180790960452, |
|
"grad_norm": 0.22251500189304352, |
|
"learning_rate": 1.0477289033146675e-05, |
|
"loss": 0.0598, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 2.391713747645951, |
|
"grad_norm": 0.297451913356781, |
|
"learning_rate": 1.022934018048432e-05, |
|
"loss": 0.0518, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 2.3992467043314503, |
|
"grad_norm": 0.29483604431152344, |
|
"learning_rate": 9.984025750977339e-06, |
|
"loss": 0.0875, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 2.406779661016949, |
|
"grad_norm": 0.3049258887767792, |
|
"learning_rate": 9.741361994343867e-06, |
|
"loss": 0.0843, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 2.414312617702448, |
|
"grad_norm": 0.21589063107967377, |
|
"learning_rate": 9.501364984720557e-06, |
|
"loss": 0.0588, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 2.4218455743879472, |
|
"grad_norm": 0.2357209175825119, |
|
"learning_rate": 9.264050619597697e-06, |
|
"loss": 0.105, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.4218455743879472, |
|
"eval_loss": 0.0942394882440567, |
|
"eval_runtime": 43.9875, |
|
"eval_samples_per_second": 1.705, |
|
"eval_steps_per_second": 0.864, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.4218455743879472, |
|
"step": 320, |
|
"total_flos": 4.419854408240333e+16, |
|
"train_loss": 0.537436925212387, |
|
"train_runtime": 3691.1633, |
|
"train_samples_per_second": 0.863, |
|
"train_steps_per_second": 0.107 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 396, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 4, |
|
"early_stopping_threshold": 0.0015 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 4 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 4.419854408240333e+16, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|