{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9991031390134529, "eval_steps": 500, "global_step": 557, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002810468997013877, "grad_norm": 0.08777584135532379, "learning_rate": 1.111111111111111e-06, "loss": 1.6284, "step": 1 }, { "epoch": 0.005620937994027754, "grad_norm": 0.08739772439002991, "learning_rate": 2.222222222222222e-06, "loss": 1.6386, "step": 2 }, { "epoch": 0.00843140699104163, "grad_norm": 0.07890280336141586, "learning_rate": 3.3333333333333333e-06, "loss": 1.6366, "step": 3 }, { "epoch": 0.011241875988055507, "grad_norm": 0.050849832594394684, "learning_rate": 4.444444444444444e-06, "loss": 1.6428, "step": 4 }, { "epoch": 0.014052344985069383, "grad_norm": 0.026262454688549042, "learning_rate": 5.555555555555557e-06, "loss": 1.6335, "step": 5 }, { "epoch": 0.01686281398208326, "grad_norm": 0.017007650807499886, "learning_rate": 6.666666666666667e-06, "loss": 1.6298, "step": 6 }, { "epoch": 0.019673282979097137, "grad_norm": 0.021000387147068977, "learning_rate": 7.77777777777778e-06, "loss": 1.6349, "step": 7 }, { "epoch": 0.022483751976111015, "grad_norm": 0.039422884583473206, "learning_rate": 8.888888888888888e-06, "loss": 1.6161, "step": 8 }, { "epoch": 0.02529422097312489, "grad_norm": 0.03416357561945915, "learning_rate": 1e-05, "loss": 1.6133, "step": 9 }, { "epoch": 0.028104689970138767, "grad_norm": 0.023096244782209396, "learning_rate": 1.1111111111111113e-05, "loss": 1.6017, "step": 10 }, { "epoch": 0.030915158967152644, "grad_norm": 0.028580009937286377, "learning_rate": 1.2222222222222224e-05, "loss": 1.6117, "step": 11 }, { "epoch": 0.03372562796416652, "grad_norm": 0.02169407717883587, "learning_rate": 1.3333333333333333e-05, "loss": 1.605, "step": 12 }, { "epoch": 0.036536096961180396, "grad_norm": 0.02079104818403721, "learning_rate": 1.4444444444444446e-05, "loss": 1.6006, "step": 13 }, { "epoch": 0.039346565958194274, "grad_norm": 0.02266255021095276, "learning_rate": 1.555555555555556e-05, "loss": 1.6076, "step": 14 }, { "epoch": 0.04215703495520815, "grad_norm": 0.023310977965593338, "learning_rate": 1.6666666666666667e-05, "loss": 1.6051, "step": 15 }, { "epoch": 0.04496750395222203, "grad_norm": 0.02227545529603958, "learning_rate": 1.7777777777777777e-05, "loss": 1.5976, "step": 16 }, { "epoch": 0.04777797294923591, "grad_norm": 0.02008313126862049, "learning_rate": 1.888888888888889e-05, "loss": 1.6048, "step": 17 }, { "epoch": 0.05058844194624978, "grad_norm": 0.020431598648428917, "learning_rate": 2e-05, "loss": 1.6033, "step": 18 }, { "epoch": 0.053398910943263656, "grad_norm": 0.019984332844614983, "learning_rate": 1.999956548296958e-05, "loss": 1.5915, "step": 19 }, { "epoch": 0.05620937994027753, "grad_norm": 0.020108483731746674, "learning_rate": 1.9998261969639324e-05, "loss": 1.5938, "step": 20 }, { "epoch": 0.05901984893729141, "grad_norm": 0.016152869910001755, "learning_rate": 1.9996089573288985e-05, "loss": 1.5931, "step": 21 }, { "epoch": 0.06183031793430529, "grad_norm": 0.016522705554962158, "learning_rate": 1.99930484827072e-05, "loss": 1.5953, "step": 22 }, { "epoch": 0.06464078693131917, "grad_norm": 0.017014645040035248, "learning_rate": 1.9989138962175105e-05, "loss": 1.5898, "step": 23 }, { "epoch": 0.06745125592833304, "grad_norm": 0.016037074849009514, "learning_rate": 1.9984361351443343e-05, "loss": 1.5904, "step": 24 }, { "epoch": 0.07026172492534692, "grad_norm": 0.01631341129541397, "learning_rate": 1.9978716065702566e-05, "loss": 1.5913, "step": 25 }, { "epoch": 0.07307219392236079, "grad_norm": 0.014392613433301449, "learning_rate": 1.9972203595547334e-05, "loss": 1.5893, "step": 26 }, { "epoch": 0.07588266291937468, "grad_norm": 0.015151307918131351, "learning_rate": 1.996482450693348e-05, "loss": 1.5912, "step": 27 }, { "epoch": 0.07869313191638855, "grad_norm": 0.015482740476727486, "learning_rate": 1.9956579441128942e-05, "loss": 1.5847, "step": 28 }, { "epoch": 0.08150360091340242, "grad_norm": 0.01364535465836525, "learning_rate": 1.994746911465802e-05, "loss": 1.585, "step": 29 }, { "epoch": 0.0843140699104163, "grad_norm": 0.014720232225954533, "learning_rate": 1.9937494319239112e-05, "loss": 1.5773, "step": 30 }, { "epoch": 0.08712453890743017, "grad_norm": 0.012587211094796658, "learning_rate": 1.9926655921715924e-05, "loss": 1.5766, "step": 31 }, { "epoch": 0.08993500790444406, "grad_norm": 0.013409728184342384, "learning_rate": 1.9914954863982106e-05, "loss": 1.5764, "step": 32 }, { "epoch": 0.09274547690145793, "grad_norm": 0.013652725145220757, "learning_rate": 1.990239216289944e-05, "loss": 1.5818, "step": 33 }, { "epoch": 0.09555594589847181, "grad_norm": 0.015371325425803661, "learning_rate": 1.9888968910209433e-05, "loss": 1.5834, "step": 34 }, { "epoch": 0.09836641489548568, "grad_norm": 0.013734452426433563, "learning_rate": 1.9874686272438467e-05, "loss": 1.5719, "step": 35 }, { "epoch": 0.10117688389249956, "grad_norm": 0.012110204435884953, "learning_rate": 1.9859545490796414e-05, "loss": 1.5797, "step": 36 }, { "epoch": 0.10398735288951344, "grad_norm": 0.013209005817770958, "learning_rate": 1.9843547881068763e-05, "loss": 1.5789, "step": 37 }, { "epoch": 0.10679782188652731, "grad_norm": 0.01261916570365429, "learning_rate": 1.9826694833502295e-05, "loss": 1.5716, "step": 38 }, { "epoch": 0.1096082908835412, "grad_norm": 0.012995808385312557, "learning_rate": 1.9808987812684247e-05, "loss": 1.5682, "step": 39 }, { "epoch": 0.11241875988055507, "grad_norm": 0.013399253599345684, "learning_rate": 1.979042835741503e-05, "loss": 1.5711, "step": 40 }, { "epoch": 0.11522922887756895, "grad_norm": 0.012696914374828339, "learning_rate": 1.9771018080574534e-05, "loss": 1.5742, "step": 41 }, { "epoch": 0.11803969787458282, "grad_norm": 0.011595464311540127, "learning_rate": 1.9750758668981925e-05, "loss": 1.581, "step": 42 }, { "epoch": 0.12085016687159669, "grad_norm": 0.01301959715783596, "learning_rate": 1.9729651883249075e-05, "loss": 1.5821, "step": 43 }, { "epoch": 0.12366063586861058, "grad_norm": 0.012804425321519375, "learning_rate": 1.9707699557627554e-05, "loss": 1.5921, "step": 44 }, { "epoch": 0.12647110486562446, "grad_norm": 0.012021846137940884, "learning_rate": 1.968490359984923e-05, "loss": 1.5795, "step": 45 }, { "epoch": 0.12928157386263833, "grad_norm": 0.013476583175361156, "learning_rate": 1.9661265990960486e-05, "loss": 1.5755, "step": 46 }, { "epoch": 0.1320920428596522, "grad_norm": 0.011448990553617477, "learning_rate": 1.9636788785150037e-05, "loss": 1.5688, "step": 47 }, { "epoch": 0.13490251185666607, "grad_norm": 0.014088047668337822, "learning_rate": 1.9611474109570446e-05, "loss": 1.5785, "step": 48 }, { "epoch": 0.13771298085367994, "grad_norm": 0.013553822413086891, "learning_rate": 1.9585324164153236e-05, "loss": 1.5812, "step": 49 }, { "epoch": 0.14052344985069384, "grad_norm": 0.011343201622366905, "learning_rate": 1.9558341221417744e-05, "loss": 1.5778, "step": 50 }, { "epoch": 0.14333391884770771, "grad_norm": 0.01123133860528469, "learning_rate": 1.9530527626273592e-05, "loss": 1.5758, "step": 51 }, { "epoch": 0.14614438784472158, "grad_norm": 0.01259944774210453, "learning_rate": 1.9501885795816937e-05, "loss": 1.5731, "step": 52 }, { "epoch": 0.14895485684173546, "grad_norm": 0.011086889542639256, "learning_rate": 1.9472418219120403e-05, "loss": 1.576, "step": 53 }, { "epoch": 0.15176532583874935, "grad_norm": 0.011971558444201946, "learning_rate": 1.9442127457016768e-05, "loss": 1.571, "step": 54 }, { "epoch": 0.15457579483576322, "grad_norm": 0.012076422572135925, "learning_rate": 1.9411016141876438e-05, "loss": 1.5743, "step": 55 }, { "epoch": 0.1573862638327771, "grad_norm": 0.01147051528096199, "learning_rate": 1.9379086977378664e-05, "loss": 1.5701, "step": 56 }, { "epoch": 0.16019673282979097, "grad_norm": 0.013713942840695381, "learning_rate": 1.9346342738276593e-05, "loss": 1.5632, "step": 57 }, { "epoch": 0.16300720182680484, "grad_norm": 0.012207642197608948, "learning_rate": 1.9312786270156135e-05, "loss": 1.5684, "step": 58 }, { "epoch": 0.16581767082381874, "grad_norm": 0.01164440531283617, "learning_rate": 1.927842048918867e-05, "loss": 1.572, "step": 59 }, { "epoch": 0.1686281398208326, "grad_norm": 0.014730863273143768, "learning_rate": 1.9243248381877605e-05, "loss": 1.5737, "step": 60 }, { "epoch": 0.17143860881784648, "grad_norm": 0.01201706100255251, "learning_rate": 1.9207273004798873e-05, "loss": 1.5708, "step": 61 }, { "epoch": 0.17424907781486035, "grad_norm": 0.012422306463122368, "learning_rate": 1.9170497484335276e-05, "loss": 1.5666, "step": 62 }, { "epoch": 0.17705954681187422, "grad_norm": 0.014150385744869709, "learning_rate": 1.9132925016404805e-05, "loss": 1.5652, "step": 63 }, { "epoch": 0.17987001580888812, "grad_norm": 0.01331349741667509, "learning_rate": 1.9094558866182892e-05, "loss": 1.5613, "step": 64 }, { "epoch": 0.182680484805902, "grad_norm": 0.014323413372039795, "learning_rate": 1.9055402367818673e-05, "loss": 1.5643, "step": 65 }, { "epoch": 0.18549095380291586, "grad_norm": 0.010913478210568428, "learning_rate": 1.901545892414523e-05, "loss": 1.5768, "step": 66 }, { "epoch": 0.18830142279992973, "grad_norm": 0.013672353699803352, "learning_rate": 1.897473200638386e-05, "loss": 1.5707, "step": 67 }, { "epoch": 0.19111189179694363, "grad_norm": 0.011649075895547867, "learning_rate": 1.8933225153842446e-05, "loss": 1.5678, "step": 68 }, { "epoch": 0.1939223607939575, "grad_norm": 0.012527153827250004, "learning_rate": 1.8890941973607843e-05, "loss": 1.5637, "step": 69 }, { "epoch": 0.19673282979097137, "grad_norm": 0.013441166840493679, "learning_rate": 1.8847886140232438e-05, "loss": 1.5543, "step": 70 }, { "epoch": 0.19954329878798524, "grad_norm": 0.014678705483675003, "learning_rate": 1.8804061395414795e-05, "loss": 1.5594, "step": 71 }, { "epoch": 0.2023537677849991, "grad_norm": 0.015578769147396088, "learning_rate": 1.875947154767452e-05, "loss": 1.5766, "step": 72 }, { "epoch": 0.205164236782013, "grad_norm": 0.013635441660881042, "learning_rate": 1.8714120472021252e-05, "loss": 1.5622, "step": 73 }, { "epoch": 0.20797470577902688, "grad_norm": 0.011884918436408043, "learning_rate": 1.8668012109617933e-05, "loss": 1.5539, "step": 74 }, { "epoch": 0.21078517477604075, "grad_norm": 0.013244202360510826, "learning_rate": 1.862115046743831e-05, "loss": 1.5641, "step": 75 }, { "epoch": 0.21359564377305462, "grad_norm": 0.011851202696561813, "learning_rate": 1.85735396179187e-05, "loss": 1.554, "step": 76 }, { "epoch": 0.2164061127700685, "grad_norm": 0.012839299626648426, "learning_rate": 1.8525183698604098e-05, "loss": 1.5663, "step": 77 }, { "epoch": 0.2192165817670824, "grad_norm": 0.013278639875352383, "learning_rate": 1.8476086911788588e-05, "loss": 1.5674, "step": 78 }, { "epoch": 0.22202705076409626, "grad_norm": 0.011714858934283257, "learning_rate": 1.8426253524150176e-05, "loss": 1.5628, "step": 79 }, { "epoch": 0.22483751976111013, "grad_norm": 0.01387378666549921, "learning_rate": 1.8375687866379988e-05, "loss": 1.5746, "step": 80 }, { "epoch": 0.227647988758124, "grad_norm": 0.014316901564598083, "learning_rate": 1.8324394332805913e-05, "loss": 1.56, "step": 81 }, { "epoch": 0.2304584577551379, "grad_norm": 0.013246218673884869, "learning_rate": 1.8272377381010726e-05, "loss": 1.5604, "step": 82 }, { "epoch": 0.23326892675215177, "grad_norm": 0.014117361046373844, "learning_rate": 1.8219641531444713e-05, "loss": 1.5535, "step": 83 }, { "epoch": 0.23607939574916564, "grad_norm": 0.013312868773937225, "learning_rate": 1.8166191367032828e-05, "loss": 1.5571, "step": 84 }, { "epoch": 0.23888986474617951, "grad_norm": 0.012200911529362202, "learning_rate": 1.811203153277641e-05, "loss": 1.5636, "step": 85 }, { "epoch": 0.24170033374319339, "grad_norm": 0.012060770764946938, "learning_rate": 1.8057166735349533e-05, "loss": 1.5548, "step": 86 }, { "epoch": 0.24451080274020728, "grad_norm": 0.01396957691758871, "learning_rate": 1.800160174268996e-05, "loss": 1.566, "step": 87 }, { "epoch": 0.24732127173722115, "grad_norm": 0.017935393378138542, "learning_rate": 1.7945341383584818e-05, "loss": 1.5705, "step": 88 }, { "epoch": 0.25013174073423505, "grad_norm": 0.01408262737095356, "learning_rate": 1.7888390547250944e-05, "loss": 1.5701, "step": 89 }, { "epoch": 0.2529422097312489, "grad_norm": 0.014612286351621151, "learning_rate": 1.7830754182909985e-05, "loss": 1.56, "step": 90 }, { "epoch": 0.2557526787282628, "grad_norm": 0.016053492203354836, "learning_rate": 1.7772437299358324e-05, "loss": 1.5655, "step": 91 }, { "epoch": 0.25856314772527667, "grad_norm": 0.014997400343418121, "learning_rate": 1.771344496453177e-05, "loss": 1.5692, "step": 92 }, { "epoch": 0.26137361672229054, "grad_norm": 0.013961107470095158, "learning_rate": 1.7653782305065158e-05, "loss": 1.56, "step": 93 }, { "epoch": 0.2641840857193044, "grad_norm": 0.015686677768826485, "learning_rate": 1.7593454505846807e-05, "loss": 1.5561, "step": 94 }, { "epoch": 0.2669945547163183, "grad_norm": 0.013397088274359703, "learning_rate": 1.753246680956795e-05, "loss": 1.5651, "step": 95 }, { "epoch": 0.26980502371333215, "grad_norm": 0.014148705638945103, "learning_rate": 1.7470824516267125e-05, "loss": 1.5538, "step": 96 }, { "epoch": 0.272615492710346, "grad_norm": 0.012879609130322933, "learning_rate": 1.7408532982869573e-05, "loss": 1.5592, "step": 97 }, { "epoch": 0.2754259617073599, "grad_norm": 0.012351011857390404, "learning_rate": 1.7345597622721727e-05, "loss": 1.5633, "step": 98 }, { "epoch": 0.2782364307043738, "grad_norm": 0.010927798226475716, "learning_rate": 1.7282023905120743e-05, "loss": 1.5623, "step": 99 }, { "epoch": 0.2810468997013877, "grad_norm": 0.015191658399999142, "learning_rate": 1.721781735483921e-05, "loss": 1.5585, "step": 100 }, { "epoch": 0.28385736869840156, "grad_norm": 0.01361748855561018, "learning_rate": 1.7152983551645054e-05, "loss": 1.5553, "step": 101 }, { "epoch": 0.28666783769541543, "grad_norm": 0.01628255844116211, "learning_rate": 1.708752812981659e-05, "loss": 1.5681, "step": 102 }, { "epoch": 0.2894783066924293, "grad_norm": 0.011448734439909458, "learning_rate": 1.702145677765293e-05, "loss": 1.5619, "step": 103 }, { "epoch": 0.29228877568944317, "grad_norm": 0.0159872155636549, "learning_rate": 1.6954775236979616e-05, "loss": 1.5528, "step": 104 }, { "epoch": 0.29509924468645704, "grad_norm": 0.014284605160355568, "learning_rate": 1.6887489302649657e-05, "loss": 1.547, "step": 105 }, { "epoch": 0.2979097136834709, "grad_norm": 0.014883718453347683, "learning_rate": 1.6819604822039924e-05, "loss": 1.5608, "step": 106 }, { "epoch": 0.3007201826804848, "grad_norm": 0.01631166599690914, "learning_rate": 1.6751127694543012e-05, "loss": 1.561, "step": 107 }, { "epoch": 0.3035306516774987, "grad_norm": 0.014855324290692806, "learning_rate": 1.6682063871054534e-05, "loss": 1.557, "step": 108 }, { "epoch": 0.3063411206745126, "grad_norm": 0.014950740151107311, "learning_rate": 1.661241935345599e-05, "loss": 1.5686, "step": 109 }, { "epoch": 0.30915158967152645, "grad_norm": 0.015889016911387444, "learning_rate": 1.654220019409317e-05, "loss": 1.5525, "step": 110 }, { "epoch": 0.3119620586685403, "grad_norm": 0.012233362533152103, "learning_rate": 1.6471412495250195e-05, "loss": 1.5586, "step": 111 }, { "epoch": 0.3147725276655542, "grad_norm": 0.013401877135038376, "learning_rate": 1.640006240861921e-05, "loss": 1.5645, "step": 112 }, { "epoch": 0.31758299666256806, "grad_norm": 0.012674611061811447, "learning_rate": 1.632815613476576e-05, "loss": 1.5564, "step": 113 }, { "epoch": 0.32039346565958193, "grad_norm": 0.013244451023638248, "learning_rate": 1.6255699922589968e-05, "loss": 1.5554, "step": 114 }, { "epoch": 0.3232039346565958, "grad_norm": 0.013073718175292015, "learning_rate": 1.6182700068783463e-05, "loss": 1.5561, "step": 115 }, { "epoch": 0.3260144036536097, "grad_norm": 0.012644422240555286, "learning_rate": 1.610916291728218e-05, "loss": 1.5538, "step": 116 }, { "epoch": 0.32882487265062355, "grad_norm": 0.013174659572541714, "learning_rate": 1.6035094858715065e-05, "loss": 1.5509, "step": 117 }, { "epoch": 0.33163534164763747, "grad_norm": 0.015327691100537777, "learning_rate": 1.5960502329848683e-05, "loss": 1.558, "step": 118 }, { "epoch": 0.33444581064465134, "grad_norm": 0.011017784476280212, "learning_rate": 1.588539181302786e-05, "loss": 1.5614, "step": 119 }, { "epoch": 0.3372562796416652, "grad_norm": 0.015872148796916008, "learning_rate": 1.580976983561235e-05, "loss": 1.5557, "step": 120 }, { "epoch": 0.3400667486386791, "grad_norm": 0.01222875714302063, "learning_rate": 1.5733642969409553e-05, "loss": 1.5603, "step": 121 }, { "epoch": 0.34287721763569295, "grad_norm": 0.015869170427322388, "learning_rate": 1.5657017830103448e-05, "loss": 1.5581, "step": 122 }, { "epoch": 0.3456876866327068, "grad_norm": 0.01336450781673193, "learning_rate": 1.5579901076679625e-05, "loss": 1.5663, "step": 123 }, { "epoch": 0.3484981556297207, "grad_norm": 0.015786062926054, "learning_rate": 1.5502299410846626e-05, "loss": 1.5518, "step": 124 }, { "epoch": 0.35130862462673457, "grad_norm": 0.01679217629134655, "learning_rate": 1.5424219576453526e-05, "loss": 1.5535, "step": 125 }, { "epoch": 0.35411909362374844, "grad_norm": 0.011585685424506664, "learning_rate": 1.5345668358903886e-05, "loss": 1.5715, "step": 126 }, { "epoch": 0.35692956262076236, "grad_norm": 0.015714086592197418, "learning_rate": 1.5266652584566056e-05, "loss": 1.5489, "step": 127 }, { "epoch": 0.35974003161777623, "grad_norm": 0.013761353679001331, "learning_rate": 1.5187179120179969e-05, "loss": 1.5515, "step": 128 }, { "epoch": 0.3625505006147901, "grad_norm": 0.013305050320923328, "learning_rate": 1.5107254872260366e-05, "loss": 1.5596, "step": 129 }, { "epoch": 0.365360969611804, "grad_norm": 0.013329329900443554, "learning_rate": 1.5026886786496624e-05, "loss": 1.5488, "step": 130 }, { "epoch": 0.36817143860881785, "grad_norm": 0.014221809804439545, "learning_rate": 1.4946081847149134e-05, "loss": 1.5536, "step": 131 }, { "epoch": 0.3709819076058317, "grad_norm": 0.011903174221515656, "learning_rate": 1.4864847076442358e-05, "loss": 1.5612, "step": 132 }, { "epoch": 0.3737923766028456, "grad_norm": 0.012618706561625004, "learning_rate": 1.4783189533954555e-05, "loss": 1.5578, "step": 133 }, { "epoch": 0.37660284559985946, "grad_norm": 0.012923737987875938, "learning_rate": 1.4701116316004307e-05, "loss": 1.5619, "step": 134 }, { "epoch": 0.37941331459687333, "grad_norm": 0.012171992100775242, "learning_rate": 1.46186345550338e-05, "loss": 1.5596, "step": 135 }, { "epoch": 0.38222378359388726, "grad_norm": 0.014030322432518005, "learning_rate": 1.4535751418989e-05, "loss": 1.5541, "step": 136 }, { "epoch": 0.3850342525909011, "grad_norm": 0.012220495380461216, "learning_rate": 1.4452474110696738e-05, "loss": 1.55, "step": 137 }, { "epoch": 0.387844721587915, "grad_norm": 0.013919586315751076, "learning_rate": 1.4368809867238754e-05, "loss": 1.5607, "step": 138 }, { "epoch": 0.39065519058492887, "grad_norm": 0.011982360854744911, "learning_rate": 1.4284765959322772e-05, "loss": 1.5464, "step": 139 }, { "epoch": 0.39346565958194274, "grad_norm": 0.012515905313193798, "learning_rate": 1.4200349690650654e-05, "loss": 1.5645, "step": 140 }, { "epoch": 0.3962761285789566, "grad_norm": 0.011968614533543587, "learning_rate": 1.411556839728367e-05, "loss": 1.5559, "step": 141 }, { "epoch": 0.3990865975759705, "grad_norm": 0.010865562595427036, "learning_rate": 1.4030429447004992e-05, "loss": 1.5617, "step": 142 }, { "epoch": 0.40189706657298435, "grad_norm": 0.013639074750244617, "learning_rate": 1.3944940238679384e-05, "loss": 1.5469, "step": 143 }, { "epoch": 0.4047075355699982, "grad_norm": 0.0128669748082757, "learning_rate": 1.3859108201610236e-05, "loss": 1.5466, "step": 144 }, { "epoch": 0.4075180045670121, "grad_norm": 0.011727554723620415, "learning_rate": 1.3772940794893916e-05, "loss": 1.5523, "step": 145 }, { "epoch": 0.410328473564026, "grad_norm": 0.015109344385564327, "learning_rate": 1.368644550677157e-05, "loss": 1.5554, "step": 146 }, { "epoch": 0.4131389425610399, "grad_norm": 0.013233068399131298, "learning_rate": 1.3599629853978342e-05, "loss": 1.5594, "step": 147 }, { "epoch": 0.41594941155805376, "grad_norm": 0.014392644166946411, "learning_rate": 1.3512501381090158e-05, "loss": 1.5603, "step": 148 }, { "epoch": 0.41875988055506763, "grad_norm": 0.011739206500351429, "learning_rate": 1.3425067659868084e-05, "loss": 1.5508, "step": 149 }, { "epoch": 0.4215703495520815, "grad_norm": 0.016567695885896683, "learning_rate": 1.3337336288600297e-05, "loss": 1.5529, "step": 150 }, { "epoch": 0.4243808185490954, "grad_norm": 0.01213662326335907, "learning_rate": 1.324931489144178e-05, "loss": 1.5468, "step": 151 }, { "epoch": 0.42719128754610924, "grad_norm": 0.01404926273971796, "learning_rate": 1.3161011117751756e-05, "loss": 1.5564, "step": 152 }, { "epoch": 0.4300017565431231, "grad_norm": 0.013470892794430256, "learning_rate": 1.3072432641428931e-05, "loss": 1.5541, "step": 153 }, { "epoch": 0.432812225540137, "grad_norm": 0.01260072086006403, "learning_rate": 1.2983587160244602e-05, "loss": 1.5453, "step": 154 }, { "epoch": 0.4356226945371509, "grad_norm": 0.012238179333508015, "learning_rate": 1.2894482395173695e-05, "loss": 1.5541, "step": 155 }, { "epoch": 0.4384331635341648, "grad_norm": 0.014110546559095383, "learning_rate": 1.2805126089723798e-05, "loss": 1.5606, "step": 156 }, { "epoch": 0.44124363253117865, "grad_norm": 0.011262672953307629, "learning_rate": 1.2715526009262209e-05, "loss": 1.5603, "step": 157 }, { "epoch": 0.4440541015281925, "grad_norm": 0.014026992954313755, "learning_rate": 1.2625689940341102e-05, "loss": 1.5462, "step": 158 }, { "epoch": 0.4468645705252064, "grad_norm": 0.011127503588795662, "learning_rate": 1.2535625690020861e-05, "loss": 1.5497, "step": 159 }, { "epoch": 0.44967503952222027, "grad_norm": 0.014035338535904884, "learning_rate": 1.24453410851916e-05, "loss": 1.5551, "step": 160 }, { "epoch": 0.45248550851923414, "grad_norm": 0.01174125261604786, "learning_rate": 1.2354843971892998e-05, "loss": 1.5494, "step": 161 }, { "epoch": 0.455295977516248, "grad_norm": 0.013476992025971413, "learning_rate": 1.2264142214632441e-05, "loss": 1.5622, "step": 162 }, { "epoch": 0.4581064465132619, "grad_norm": 0.012524358928203583, "learning_rate": 1.2173243695701575e-05, "loss": 1.5512, "step": 163 }, { "epoch": 0.4609169155102758, "grad_norm": 0.011893150396645069, "learning_rate": 1.2082156314491298e-05, "loss": 1.5601, "step": 164 }, { "epoch": 0.4637273845072897, "grad_norm": 0.011717031709849834, "learning_rate": 1.1990887986805295e-05, "loss": 1.5401, "step": 165 }, { "epoch": 0.46653785350430355, "grad_norm": 0.01049530878663063, "learning_rate": 1.1899446644172106e-05, "loss": 1.5434, "step": 166 }, { "epoch": 0.4693483225013174, "grad_norm": 0.013684769161045551, "learning_rate": 1.1807840233155863e-05, "loss": 1.5473, "step": 167 }, { "epoch": 0.4721587914983313, "grad_norm": 0.010303209535777569, "learning_rate": 1.1716076714665701e-05, "loss": 1.5534, "step": 168 }, { "epoch": 0.47496926049534516, "grad_norm": 0.013134041801095009, "learning_rate": 1.1624164063263931e-05, "loss": 1.54, "step": 169 }, { "epoch": 0.47777972949235903, "grad_norm": 0.011170011013746262, "learning_rate": 1.1532110266473026e-05, "loss": 1.5519, "step": 170 }, { "epoch": 0.4805901984893729, "grad_norm": 0.01223431620746851, "learning_rate": 1.1439923324081465e-05, "loss": 1.5461, "step": 171 }, { "epoch": 0.48340066748638677, "grad_norm": 0.01103509496897459, "learning_rate": 1.1347611247448544e-05, "loss": 1.5469, "step": 172 }, { "epoch": 0.48621113648340064, "grad_norm": 0.012365833856165409, "learning_rate": 1.1255182058808143e-05, "loss": 1.5546, "step": 173 }, { "epoch": 0.48902160548041457, "grad_norm": 0.011856972239911556, "learning_rate": 1.1162643790571574e-05, "loss": 1.5493, "step": 174 }, { "epoch": 0.49183207447742844, "grad_norm": 0.011247235350310802, "learning_rate": 1.1070004484629543e-05, "loss": 1.5485, "step": 175 }, { "epoch": 0.4946425434744423, "grad_norm": 0.010246271267533302, "learning_rate": 1.0977272191653272e-05, "loss": 1.5385, "step": 176 }, { "epoch": 0.4974530124714562, "grad_norm": 0.011663992889225483, "learning_rate": 1.0884454970394871e-05, "loss": 1.5543, "step": 177 }, { "epoch": 0.5002634814684701, "grad_norm": 0.011617216281592846, "learning_rate": 1.0791560886987016e-05, "loss": 1.5621, "step": 178 }, { "epoch": 0.503073950465484, "grad_norm": 0.010127879679203033, "learning_rate": 1.069859801424196e-05, "loss": 1.547, "step": 179 }, { "epoch": 0.5058844194624978, "grad_norm": 0.01151992380619049, "learning_rate": 1.0605574430949983e-05, "loss": 1.5496, "step": 180 }, { "epoch": 0.5086948884595117, "grad_norm": 0.012575927190482616, "learning_rate": 1.0512498221177319e-05, "loss": 1.5523, "step": 181 }, { "epoch": 0.5115053574565256, "grad_norm": 0.011552478186786175, "learning_rate": 1.0419377473563621e-05, "loss": 1.5537, "step": 182 }, { "epoch": 0.5143158264535395, "grad_norm": 0.010599706321954727, "learning_rate": 1.0326220280619036e-05, "loss": 1.5448, "step": 183 }, { "epoch": 0.5171262954505533, "grad_norm": 0.010817637667059898, "learning_rate": 1.0233034738020933e-05, "loss": 1.5503, "step": 184 }, { "epoch": 0.5199367644475672, "grad_norm": 0.012649599462747574, "learning_rate": 1.0139828943910358e-05, "loss": 1.5486, "step": 185 }, { "epoch": 0.5227472334445811, "grad_norm": 0.012253242544829845, "learning_rate": 1.004661099818829e-05, "loss": 1.5514, "step": 186 }, { "epoch": 0.5255577024415949, "grad_norm": 0.011790635995566845, "learning_rate": 9.953389001811716e-06, "loss": 1.5518, "step": 187 }, { "epoch": 0.5283681714386088, "grad_norm": 0.013327023014426231, "learning_rate": 9.860171056089646e-06, "loss": 1.5424, "step": 188 }, { "epoch": 0.5311786404356227, "grad_norm": 0.0112162996083498, "learning_rate": 9.766965261979072e-06, "loss": 1.5488, "step": 189 }, { "epoch": 0.5339891094326366, "grad_norm": 0.012566052377223969, "learning_rate": 9.673779719380967e-06, "loss": 1.5452, "step": 190 }, { "epoch": 0.5367995784296504, "grad_norm": 0.012440058402717113, "learning_rate": 9.580622526436382e-06, "loss": 1.5457, "step": 191 }, { "epoch": 0.5396100474266643, "grad_norm": 0.011921185068786144, "learning_rate": 9.487501778822685e-06, "loss": 1.5413, "step": 192 }, { "epoch": 0.5424205164236782, "grad_norm": 0.011896064504981041, "learning_rate": 9.394425569050018e-06, "loss": 1.5423, "step": 193 }, { "epoch": 0.545230985420692, "grad_norm": 0.01354946568608284, "learning_rate": 9.30140198575804e-06, "loss": 1.5523, "step": 194 }, { "epoch": 0.5480414544177059, "grad_norm": 0.011999576352536678, "learning_rate": 9.208439113012984e-06, "loss": 1.5367, "step": 195 }, { "epoch": 0.5508519234147198, "grad_norm": 0.012224970385432243, "learning_rate": 9.115545029605129e-06, "loss": 1.5564, "step": 196 }, { "epoch": 0.5536623924117338, "grad_norm": 0.012914243154227734, "learning_rate": 9.022727808346731e-06, "loss": 1.5426, "step": 197 }, { "epoch": 0.5564728614087476, "grad_norm": 0.012274730019271374, "learning_rate": 8.92999551537046e-06, "loss": 1.5454, "step": 198 }, { "epoch": 0.5592833304057615, "grad_norm": 0.010419441387057304, "learning_rate": 8.837356209428428e-06, "loss": 1.5388, "step": 199 }, { "epoch": 0.5620937994027754, "grad_norm": 0.012639000080525875, "learning_rate": 8.744817941191862e-06, "loss": 1.5451, "step": 200 }, { "epoch": 0.5649042683997892, "grad_norm": 0.01181803084909916, "learning_rate": 8.652388752551458e-06, "loss": 1.548, "step": 201 }, { "epoch": 0.5677147373968031, "grad_norm": 0.01122727058827877, "learning_rate": 8.560076675918537e-06, "loss": 1.5531, "step": 202 }, { "epoch": 0.570525206393817, "grad_norm": 0.011986492201685905, "learning_rate": 8.467889733526977e-06, "loss": 1.5334, "step": 203 }, { "epoch": 0.5733356753908309, "grad_norm": 0.01114217285066843, "learning_rate": 8.375835936736072e-06, "loss": 1.539, "step": 204 }, { "epoch": 0.5761461443878447, "grad_norm": 0.011127611622214317, "learning_rate": 8.283923285334304e-06, "loss": 1.546, "step": 205 }, { "epoch": 0.5789566133848586, "grad_norm": 0.011980608105659485, "learning_rate": 8.19215976684414e-06, "loss": 1.5556, "step": 206 }, { "epoch": 0.5817670823818725, "grad_norm": 0.011237296275794506, "learning_rate": 8.100553355827897e-06, "loss": 1.5368, "step": 207 }, { "epoch": 0.5845775513788863, "grad_norm": 0.01083777192980051, "learning_rate": 8.009112013194707e-06, "loss": 1.5404, "step": 208 }, { "epoch": 0.5873880203759002, "grad_norm": 0.01118643768131733, "learning_rate": 7.917843685508702e-06, "loss": 1.5405, "step": 209 }, { "epoch": 0.5901984893729141, "grad_norm": 0.012426355853676796, "learning_rate": 7.826756304298428e-06, "loss": 1.5497, "step": 210 }, { "epoch": 0.593008958369928, "grad_norm": 0.011632665991783142, "learning_rate": 7.73585778536756e-06, "loss": 1.5395, "step": 211 }, { "epoch": 0.5958194273669418, "grad_norm": 0.010038619861006737, "learning_rate": 7.645156028107005e-06, "loss": 1.5388, "step": 212 }, { "epoch": 0.5986298963639557, "grad_norm": 0.01042498741298914, "learning_rate": 7.554658914808404e-06, "loss": 1.5446, "step": 213 }, { "epoch": 0.6014403653609696, "grad_norm": 0.009714511223137379, "learning_rate": 7.464374309979143e-06, "loss": 1.545, "step": 214 }, { "epoch": 0.6042508343579834, "grad_norm": 0.011047974228858948, "learning_rate": 7.3743100596589e-06, "loss": 1.5496, "step": 215 }, { "epoch": 0.6070613033549974, "grad_norm": 0.009808655828237534, "learning_rate": 7.284473990737795e-06, "loss": 1.542, "step": 216 }, { "epoch": 0.6098717723520113, "grad_norm": 0.010092736221849918, "learning_rate": 7.194873910276205e-06, "loss": 1.5517, "step": 217 }, { "epoch": 0.6126822413490252, "grad_norm": 0.01010197214782238, "learning_rate": 7.1055176048263085e-06, "loss": 1.541, "step": 218 }, { "epoch": 0.615492710346039, "grad_norm": 0.011126919649541378, "learning_rate": 7.0164128397554e-06, "loss": 1.5582, "step": 219 }, { "epoch": 0.6183031793430529, "grad_norm": 0.01078337524086237, "learning_rate": 6.92756735857107e-06, "loss": 1.5438, "step": 220 }, { "epoch": 0.6211136483400668, "grad_norm": 0.010446320287883282, "learning_rate": 6.838988882248243e-06, "loss": 1.5448, "step": 221 }, { "epoch": 0.6239241173370806, "grad_norm": 0.011228025890886784, "learning_rate": 6.750685108558221e-06, "loss": 1.5347, "step": 222 }, { "epoch": 0.6267345863340945, "grad_norm": 0.012325327843427658, "learning_rate": 6.662663711399705e-06, "loss": 1.5537, "step": 223 }, { "epoch": 0.6295450553311084, "grad_norm": 0.01042882353067398, "learning_rate": 6.574932340131917e-06, "loss": 1.55, "step": 224 }, { "epoch": 0.6323555243281223, "grad_norm": 0.012044312432408333, "learning_rate": 6.487498618909845e-06, "loss": 1.5485, "step": 225 }, { "epoch": 0.6351659933251361, "grad_norm": 0.01117675006389618, "learning_rate": 6.400370146021662e-06, "loss": 1.5488, "step": 226 }, { "epoch": 0.63797646232215, "grad_norm": 0.010941598564386368, "learning_rate": 6.313554493228431e-06, "loss": 1.5456, "step": 227 }, { "epoch": 0.6407869313191639, "grad_norm": 0.011893962509930134, "learning_rate": 6.227059205106085e-06, "loss": 1.543, "step": 228 }, { "epoch": 0.6435974003161777, "grad_norm": 0.010805683210492134, "learning_rate": 6.14089179838977e-06, "loss": 1.5479, "step": 229 }, { "epoch": 0.6464078693131916, "grad_norm": 0.011175237596035004, "learning_rate": 6.0550597613206205e-06, "loss": 1.5399, "step": 230 }, { "epoch": 0.6492183383102055, "grad_norm": 0.011044224724173546, "learning_rate": 5.969570552995014e-06, "loss": 1.547, "step": 231 }, { "epoch": 0.6520288073072193, "grad_norm": 0.010453018359839916, "learning_rate": 5.8844316027163315e-06, "loss": 1.5468, "step": 232 }, { "epoch": 0.6548392763042332, "grad_norm": 0.010159061290323734, "learning_rate": 5.799650309349348e-06, "loss": 1.536, "step": 233 }, { "epoch": 0.6576497453012471, "grad_norm": 0.010464577935636044, "learning_rate": 5.715234040677229e-06, "loss": 1.5529, "step": 234 }, { "epoch": 0.6604602142982611, "grad_norm": 0.01091894879937172, "learning_rate": 5.631190132761247e-06, "loss": 1.5445, "step": 235 }, { "epoch": 0.6632706832952749, "grad_norm": 0.011955167166888714, "learning_rate": 5.547525889303265e-06, "loss": 1.54, "step": 236 }, { "epoch": 0.6660811522922888, "grad_norm": 0.010030088014900684, "learning_rate": 5.464248581011002e-06, "loss": 1.5453, "step": 237 }, { "epoch": 0.6688916212893027, "grad_norm": 0.009817596524953842, "learning_rate": 5.381365444966205e-06, "loss": 1.5367, "step": 238 }, { "epoch": 0.6717020902863166, "grad_norm": 0.010095087811350822, "learning_rate": 5.298883683995697e-06, "loss": 1.5559, "step": 239 }, { "epoch": 0.6745125592833304, "grad_norm": 0.010223811492323875, "learning_rate": 5.216810466045448e-06, "loss": 1.5406, "step": 240 }, { "epoch": 0.6773230282803443, "grad_norm": 0.009899747557938099, "learning_rate": 5.135152923557647e-06, "loss": 1.5395, "step": 241 }, { "epoch": 0.6801334972773582, "grad_norm": 0.009352842345833778, "learning_rate": 5.053918152850868e-06, "loss": 1.5423, "step": 242 }, { "epoch": 0.682943966274372, "grad_norm": 0.0106782466173172, "learning_rate": 4.973113213503379e-06, "loss": 1.5397, "step": 243 }, { "epoch": 0.6857544352713859, "grad_norm": 0.010458718985319138, "learning_rate": 4.8927451277396365e-06, "loss": 1.5264, "step": 244 }, { "epoch": 0.6885649042683998, "grad_norm": 0.01009244006127119, "learning_rate": 4.812820879820034e-06, "loss": 1.5512, "step": 245 }, { "epoch": 0.6913753732654137, "grad_norm": 0.009827625937759876, "learning_rate": 4.733347415433946e-06, "loss": 1.5421, "step": 246 }, { "epoch": 0.6941858422624275, "grad_norm": 0.010725892148911953, "learning_rate": 4.654331641096118e-06, "loss": 1.5447, "step": 247 }, { "epoch": 0.6969963112594414, "grad_norm": 0.009530246257781982, "learning_rate": 4.575780423546476e-06, "loss": 1.5429, "step": 248 }, { "epoch": 0.6998067802564553, "grad_norm": 0.009370286948978901, "learning_rate": 4.497700589153379e-06, "loss": 1.54, "step": 249 }, { "epoch": 0.7026172492534691, "grad_norm": 0.010170893743634224, "learning_rate": 4.420098923320378e-06, "loss": 1.5448, "step": 250 }, { "epoch": 0.705427718250483, "grad_norm": 0.01000099815428257, "learning_rate": 4.342982169896555e-06, "loss": 1.5377, "step": 251 }, { "epoch": 0.7082381872474969, "grad_norm": 0.010095086880028248, "learning_rate": 4.266357030590449e-06, "loss": 1.5321, "step": 252 }, { "epoch": 0.7110486562445109, "grad_norm": 0.009943116456270218, "learning_rate": 4.1902301643876555e-06, "loss": 1.543, "step": 253 }, { "epoch": 0.7138591252415247, "grad_norm": 0.009373177774250507, "learning_rate": 4.114608186972143e-06, "loss": 1.5432, "step": 254 }, { "epoch": 0.7166695942385386, "grad_norm": 0.009264045394957066, "learning_rate": 4.0394976701513235e-06, "loss": 1.5501, "step": 255 }, { "epoch": 0.7194800632355525, "grad_norm": 0.00987312849611044, "learning_rate": 3.96490514128494e-06, "loss": 1.5376, "step": 256 }, { "epoch": 0.7222905322325663, "grad_norm": 0.010003172792494297, "learning_rate": 3.890837082717822e-06, "loss": 1.5424, "step": 257 }, { "epoch": 0.7251010012295802, "grad_norm": 0.009270971640944481, "learning_rate": 3.817299931216537e-06, "loss": 1.5453, "step": 258 }, { "epoch": 0.7279114702265941, "grad_norm": 0.009572354145348072, "learning_rate": 3.74430007741003e-06, "loss": 1.5325, "step": 259 }, { "epoch": 0.730721939223608, "grad_norm": 0.009282928891479969, "learning_rate": 3.671843865234238e-06, "loss": 1.5447, "step": 260 }, { "epoch": 0.7335324082206218, "grad_norm": 0.0098671093583107, "learning_rate": 3.599937591380791e-06, "loss": 1.5466, "step": 261 }, { "epoch": 0.7363428772176357, "grad_norm": 0.010173565708100796, "learning_rate": 3.5285875047498075e-06, "loss": 1.5347, "step": 262 }, { "epoch": 0.7391533462146496, "grad_norm": 0.0094713494181633, "learning_rate": 3.4577998059068354e-06, "loss": 1.542, "step": 263 }, { "epoch": 0.7419638152116634, "grad_norm": 0.009647979401051998, "learning_rate": 3.3875806465440152e-06, "loss": 1.5437, "step": 264 }, { "epoch": 0.7447742842086773, "grad_norm": 0.009664146229624748, "learning_rate": 3.3179361289454694e-06, "loss": 1.5406, "step": 265 }, { "epoch": 0.7475847532056912, "grad_norm": 0.010614863596856594, "learning_rate": 3.2488723054569905e-06, "loss": 1.5512, "step": 266 }, { "epoch": 0.750395222202705, "grad_norm": 0.009987459518015385, "learning_rate": 3.1803951779600774e-06, "loss": 1.5341, "step": 267 }, { "epoch": 0.7532056911997189, "grad_norm": 0.009676804766058922, "learning_rate": 3.112510697350348e-06, "loss": 1.5457, "step": 268 }, { "epoch": 0.7560161601967328, "grad_norm": 0.010411749593913555, "learning_rate": 3.04522476302039e-06, "loss": 1.5456, "step": 269 }, { "epoch": 0.7588266291937467, "grad_norm": 0.010124873369932175, "learning_rate": 2.978543222347076e-06, "loss": 1.5452, "step": 270 }, { "epoch": 0.7616370981907605, "grad_norm": 0.009868764318525791, "learning_rate": 2.912471870183411e-06, "loss": 1.555, "step": 271 }, { "epoch": 0.7644475671877745, "grad_norm": 0.009340250864624977, "learning_rate": 2.847016448354948e-06, "loss": 1.5419, "step": 272 }, { "epoch": 0.7672580361847884, "grad_norm": 0.00967357400804758, "learning_rate": 2.782182645160789e-06, "loss": 1.5433, "step": 273 }, { "epoch": 0.7700685051818023, "grad_norm": 0.009733307175338268, "learning_rate": 2.71797609487926e-06, "loss": 1.5378, "step": 274 }, { "epoch": 0.7728789741788161, "grad_norm": 0.009725394658744335, "learning_rate": 2.6544023772782736e-06, "loss": 1.5373, "step": 275 }, { "epoch": 0.77568944317583, "grad_norm": 0.009589461609721184, "learning_rate": 2.591467017130426e-06, "loss": 1.5331, "step": 276 }, { "epoch": 0.7784999121728439, "grad_norm": 0.009730999357998371, "learning_rate": 2.5291754837328787e-06, "loss": 1.5374, "step": 277 }, { "epoch": 0.7813103811698577, "grad_norm": 0.009977961890399456, "learning_rate": 2.4675331904320533e-06, "loss": 1.5386, "step": 278 }, { "epoch": 0.7841208501668716, "grad_norm": 0.009760740213096142, "learning_rate": 2.4065454941531963e-06, "loss": 1.5521, "step": 279 }, { "epoch": 0.7869313191638855, "grad_norm": 0.009903214871883392, "learning_rate": 2.346217694934847e-06, "loss": 1.5408, "step": 280 }, { "epoch": 0.7897417881608993, "grad_norm": 0.010044459253549576, "learning_rate": 2.286555035468233e-06, "loss": 1.54, "step": 281 }, { "epoch": 0.7925522571579132, "grad_norm": 0.009294719435274601, "learning_rate": 2.22756270064168e-06, "loss": 1.5417, "step": 282 }, { "epoch": 0.7953627261549271, "grad_norm": 0.009228096343576908, "learning_rate": 2.16924581709002e-06, "loss": 1.5285, "step": 283 }, { "epoch": 0.798173195151941, "grad_norm": 0.009288915432989597, "learning_rate": 2.1116094527490594e-06, "loss": 1.5431, "step": 284 }, { "epoch": 0.8009836641489548, "grad_norm": 0.010039190761744976, "learning_rate": 2.0546586164151827e-06, "loss": 1.5324, "step": 285 }, { "epoch": 0.8037941331459687, "grad_norm": 0.009471042081713676, "learning_rate": 1.9983982573100413e-06, "loss": 1.5375, "step": 286 }, { "epoch": 0.8066046021429826, "grad_norm": 0.009732691571116447, "learning_rate": 1.94283326465047e-06, "loss": 1.5425, "step": 287 }, { "epoch": 0.8094150711399964, "grad_norm": 0.00961157213896513, "learning_rate": 1.887968467223591e-06, "loss": 1.5378, "step": 288 }, { "epoch": 0.8122255401370103, "grad_norm": 0.009570587426424026, "learning_rate": 1.8338086329671734e-06, "loss": 1.546, "step": 289 }, { "epoch": 0.8150360091340242, "grad_norm": 0.009907620027661324, "learning_rate": 1.7803584685552877e-06, "loss": 1.545, "step": 290 }, { "epoch": 0.8178464781310382, "grad_norm": 0.009468463249504566, "learning_rate": 1.7276226189892763e-06, "loss": 1.5387, "step": 291 }, { "epoch": 0.820656947128052, "grad_norm": 0.008901305496692657, "learning_rate": 1.6756056671940902e-06, "loss": 1.541, "step": 292 }, { "epoch": 0.8234674161250659, "grad_norm": 0.009180068038403988, "learning_rate": 1.624312133620013e-06, "loss": 1.543, "step": 293 }, { "epoch": 0.8262778851220798, "grad_norm": 0.0097258435562253, "learning_rate": 1.5737464758498243e-06, "loss": 1.5323, "step": 294 }, { "epoch": 0.8290883541190937, "grad_norm": 0.00945055577903986, "learning_rate": 1.523913088211415e-06, "loss": 1.5366, "step": 295 }, { "epoch": 0.8318988231161075, "grad_norm": 0.009719719178974628, "learning_rate": 1.474816301395906e-06, "loss": 1.5489, "step": 296 }, { "epoch": 0.8347092921131214, "grad_norm": 0.00904281809926033, "learning_rate": 1.4264603820813006e-06, "loss": 1.5413, "step": 297 }, { "epoch": 0.8375197611101353, "grad_norm": 0.009636901319026947, "learning_rate": 1.3788495325616912e-06, "loss": 1.5493, "step": 298 }, { "epoch": 0.8403302301071491, "grad_norm": 0.00919989962130785, "learning_rate": 1.3319878903820682e-06, "loss": 1.5387, "step": 299 }, { "epoch": 0.843140699104163, "grad_norm": 0.009369692765176296, "learning_rate": 1.2858795279787517e-06, "loss": 1.54, "step": 300 }, { "epoch": 0.8459511681011769, "grad_norm": 0.009285050444304943, "learning_rate": 1.2405284523254823e-06, "loss": 1.5496, "step": 301 }, { "epoch": 0.8487616370981907, "grad_norm": 0.009485352784395218, "learning_rate": 1.195938604585205e-06, "loss": 1.5274, "step": 302 }, { "epoch": 0.8515721060952046, "grad_norm": 0.00928215030580759, "learning_rate": 1.152113859767565e-06, "loss": 1.5437, "step": 303 }, { "epoch": 0.8543825750922185, "grad_norm": 0.009152165614068508, "learning_rate": 1.109058026392158e-06, "loss": 1.5395, "step": 304 }, { "epoch": 0.8571930440892324, "grad_norm": 0.009171172976493835, "learning_rate": 1.0667748461575544e-06, "loss": 1.5487, "step": 305 }, { "epoch": 0.8600035130862462, "grad_norm": 0.00959504209458828, "learning_rate": 1.0252679936161392e-06, "loss": 1.5374, "step": 306 }, { "epoch": 0.8628139820832601, "grad_norm": 0.00906702782958746, "learning_rate": 9.845410758547724e-07, "loss": 1.5412, "step": 307 }, { "epoch": 0.865624451080274, "grad_norm": 0.008921015076339245, "learning_rate": 9.445976321813277e-07, "loss": 1.5295, "step": 308 }, { "epoch": 0.868434920077288, "grad_norm": 0.009434668347239494, "learning_rate": 9.054411338171099e-07, "loss": 1.5474, "step": 309 }, { "epoch": 0.8712453890743018, "grad_norm": 0.009549295529723167, "learning_rate": 8.670749835951964e-07, "loss": 1.5425, "step": 310 }, { "epoch": 0.8740558580713157, "grad_norm": 0.009337035939097404, "learning_rate": 8.29502515664723e-07, "loss": 1.5412, "step": 311 }, { "epoch": 0.8768663270683296, "grad_norm": 0.008889071643352509, "learning_rate": 7.927269952011285e-07, "loss": 1.5589, "step": 312 }, { "epoch": 0.8796767960653434, "grad_norm": 0.008982825092971325, "learning_rate": 7.567516181223966e-07, "loss": 1.5504, "step": 313 }, { "epoch": 0.8824872650623573, "grad_norm": 0.009038092568516731, "learning_rate": 7.215795108113343e-07, "loss": 1.5391, "step": 314 }, { "epoch": 0.8852977340593712, "grad_norm": 0.0092889703810215, "learning_rate": 6.872137298438653e-07, "loss": 1.5354, "step": 315 }, { "epoch": 0.888108203056385, "grad_norm": 0.009272675029933453, "learning_rate": 6.536572617234082e-07, "loss": 1.5449, "step": 316 }, { "epoch": 0.8909186720533989, "grad_norm": 0.00896420981734991, "learning_rate": 6.209130226213378e-07, "loss": 1.5544, "step": 317 }, { "epoch": 0.8937291410504128, "grad_norm": 0.00960630364716053, "learning_rate": 5.889838581235641e-07, "loss": 1.5425, "step": 318 }, { "epoch": 0.8965396100474267, "grad_norm": 0.009224542416632175, "learning_rate": 5.578725429832344e-07, "loss": 1.5287, "step": 319 }, { "epoch": 0.8993500790444405, "grad_norm": 0.009561055339872837, "learning_rate": 5.275817808796013e-07, "loss": 1.5399, "step": 320 }, { "epoch": 0.9021605480414544, "grad_norm": 0.00883413664996624, "learning_rate": 4.981142041830645e-07, "loss": 1.5373, "step": 321 }, { "epoch": 0.9049710170384683, "grad_norm": 0.0092536062002182, "learning_rate": 4.6947237372640954e-07, "loss": 1.5372, "step": 322 }, { "epoch": 0.9077814860354821, "grad_norm": 0.008912206627428532, "learning_rate": 4.416587785822568e-07, "loss": 1.5412, "step": 323 }, { "epoch": 0.910591955032496, "grad_norm": 0.009592842310667038, "learning_rate": 4.1467583584676395e-07, "loss": 1.541, "step": 324 }, { "epoch": 0.9134024240295099, "grad_norm": 0.00898828823119402, "learning_rate": 3.885258904295575e-07, "loss": 1.539, "step": 325 }, { "epoch": 0.9162128930265238, "grad_norm": 0.009104624390602112, "learning_rate": 3.6321121484996447e-07, "loss": 1.5313, "step": 326 }, { "epoch": 0.9190233620235376, "grad_norm": 0.00905083492398262, "learning_rate": 3.3873400903951636e-07, "loss": 1.5396, "step": 327 }, { "epoch": 0.9218338310205516, "grad_norm": 0.009055250324308872, "learning_rate": 3.1509640015076946e-07, "loss": 1.5471, "step": 328 }, { "epoch": 0.9246443000175655, "grad_norm": 0.00856536254286766, "learning_rate": 2.923004423724474e-07, "loss": 1.5302, "step": 329 }, { "epoch": 0.9274547690145793, "grad_norm": 0.0095223942771554, "learning_rate": 2.703481167509281e-07, "loss": 1.5387, "step": 330 }, { "epoch": 0.9302652380115932, "grad_norm": 0.009137586690485477, "learning_rate": 2.4924133101807636e-07, "loss": 1.5379, "step": 331 }, { "epoch": 0.9330757070086071, "grad_norm": 0.009256749413907528, "learning_rate": 2.289819194254661e-07, "loss": 1.54, "step": 332 }, { "epoch": 0.935886176005621, "grad_norm": 0.00952975358814001, "learning_rate": 2.0957164258497031e-07, "loss": 1.5413, "step": 333 }, { "epoch": 0.9386966450026348, "grad_norm": 0.009139897301793098, "learning_rate": 1.9101218731575777e-07, "loss": 1.5547, "step": 334 }, { "epoch": 0.9415071139996487, "grad_norm": 0.009397204965353012, "learning_rate": 1.73305166497707e-07, "loss": 1.5374, "step": 335 }, { "epoch": 0.9443175829966626, "grad_norm": 0.009158180095255375, "learning_rate": 1.5645211893123846e-07, "loss": 1.5449, "step": 336 }, { "epoch": 0.9471280519936764, "grad_norm": 0.008892880752682686, "learning_rate": 1.4045450920358917e-07, "loss": 1.5418, "step": 337 }, { "epoch": 0.9499385209906903, "grad_norm": 0.008902773261070251, "learning_rate": 1.2531372756153458e-07, "loss": 1.5451, "step": 338 }, { "epoch": 0.9527489899877042, "grad_norm": 0.00923394039273262, "learning_rate": 1.1103108979056865e-07, "loss": 1.5329, "step": 339 }, { "epoch": 0.9555594589847181, "grad_norm": 0.009119280613958836, "learning_rate": 9.760783710056176e-08, "loss": 1.5399, "step": 340 }, { "epoch": 0.9583699279817319, "grad_norm": 0.009382793679833412, "learning_rate": 8.504513601789388e-08, "loss": 1.5372, "step": 341 }, { "epoch": 0.9611803969787458, "grad_norm": 0.0087265744805336, "learning_rate": 7.334407828407885e-08, "loss": 1.5415, "step": 342 }, { "epoch": 0.9639908659757597, "grad_norm": 0.008734731003642082, "learning_rate": 6.250568076088814e-08, "loss": 1.5388, "step": 343 }, { "epoch": 0.9668013349727735, "grad_norm": 0.00877598486840725, "learning_rate": 5.2530885341982586e-08, "loss": 1.5385, "step": 344 }, { "epoch": 0.9696118039697874, "grad_norm": 0.008809634484350681, "learning_rate": 4.3420558871060116e-08, "loss": 1.5447, "step": 345 }, { "epoch": 0.9724222729668013, "grad_norm": 0.00890128593891859, "learning_rate": 3.517549306652157e-08, "loss": 1.538, "step": 346 }, { "epoch": 0.9752327419638153, "grad_norm": 0.008788186125457287, "learning_rate": 2.7796404452666847e-08, "loss": 1.5506, "step": 347 }, { "epoch": 0.9780432109608291, "grad_norm": 0.009096509777009487, "learning_rate": 2.1283934297432472e-08, "loss": 1.5365, "step": 348 }, { "epoch": 0.980853679957843, "grad_norm": 0.00890890508890152, "learning_rate": 1.5638648556656198e-08, "loss": 1.5406, "step": 349 }, { "epoch": 0.9836641489548569, "grad_norm": 0.008733571507036686, "learning_rate": 1.0861037824896337e-08, "loss": 1.5474, "step": 350 }, { "epoch": 0.9864746179518707, "grad_norm": 0.00961330346763134, "learning_rate": 6.951517292800303e-09, "loss": 1.5417, "step": 351 }, { "epoch": 0.9892850869488846, "grad_norm": 0.009068558923900127, "learning_rate": 3.9104267110168235e-09, "loss": 1.5369, "step": 352 }, { "epoch": 0.9920955559458985, "grad_norm": 0.008888053707778454, "learning_rate": 1.738030360677323e-09, "loss": 1.5488, "step": 353 }, { "epoch": 0.9949060249429124, "grad_norm": 0.009277201257646084, "learning_rate": 4.3451703042207694e-10, "loss": 1.5375, "step": 354 }, { "epoch": 0.9977164939399262, "grad_norm": 0.009334351867437363, "learning_rate": 0.0, "loss": 1.5352, "step": 355 }, { "epoch": 0.6385650224215247, "grad_norm": 0.010235507041215897, "learning_rate": 2e-05, "loss": 1.5054, "step": 356 }, { "epoch": 0.6403587443946188, "grad_norm": 0.010560178197920322, "learning_rate": 2e-05, "loss": 1.5118, "step": 357 }, { "epoch": 0.6421524663677131, "grad_norm": 0.010353959165513515, "learning_rate": 2e-05, "loss": 1.5226, "step": 358 }, { "epoch": 0.6439461883408072, "grad_norm": 0.010382940992712975, "learning_rate": 2e-05, "loss": 1.5078, "step": 359 }, { "epoch": 0.6457399103139013, "grad_norm": 0.009856803342700005, "learning_rate": 2e-05, "loss": 1.5167, "step": 360 }, { "epoch": 0.6475336322869956, "grad_norm": 0.010195410810410976, "learning_rate": 2e-05, "loss": 1.5142, "step": 361 }, { "epoch": 0.6493273542600897, "grad_norm": 0.010302864946424961, "learning_rate": 2e-05, "loss": 1.5136, "step": 362 }, { "epoch": 0.6511210762331838, "grad_norm": 0.010046405717730522, "learning_rate": 2e-05, "loss": 1.5112, "step": 363 }, { "epoch": 0.6529147982062781, "grad_norm": 0.010849208571016788, "learning_rate": 2e-05, "loss": 1.5114, "step": 364 }, { "epoch": 0.6547085201793722, "grad_norm": 0.010421674698591232, "learning_rate": 2e-05, "loss": 1.5173, "step": 365 }, { "epoch": 0.6565022421524663, "grad_norm": 0.00989589188247919, "learning_rate": 2e-05, "loss": 1.5063, "step": 366 }, { "epoch": 0.6582959641255606, "grad_norm": 0.010465629398822784, "learning_rate": 2e-05, "loss": 1.5031, "step": 367 }, { "epoch": 0.6600896860986547, "grad_norm": 0.009964341297745705, "learning_rate": 2e-05, "loss": 1.5207, "step": 368 }, { "epoch": 0.6618834080717488, "grad_norm": 0.01189314667135477, "learning_rate": 2e-05, "loss": 1.5361, "step": 369 }, { "epoch": 0.6636771300448431, "grad_norm": 0.01012677513062954, "learning_rate": 2e-05, "loss": 1.5215, "step": 370 }, { "epoch": 0.6654708520179372, "grad_norm": 0.009877102449536324, "learning_rate": 2e-05, "loss": 1.5262, "step": 371 }, { "epoch": 0.6672645739910313, "grad_norm": 0.01000463031232357, "learning_rate": 2e-05, "loss": 1.5183, "step": 372 }, { "epoch": 0.6690582959641256, "grad_norm": 0.010188892483711243, "learning_rate": 2e-05, "loss": 1.5183, "step": 373 }, { "epoch": 0.6708520179372197, "grad_norm": 0.010129815898835659, "learning_rate": 2e-05, "loss": 1.5245, "step": 374 }, { "epoch": 0.672645739910314, "grad_norm": 0.010608335956931114, "learning_rate": 2e-05, "loss": 1.5169, "step": 375 }, { "epoch": 0.6744394618834081, "grad_norm": 0.010223207995295525, "learning_rate": 2e-05, "loss": 1.5185, "step": 376 }, { "epoch": 0.6762331838565022, "grad_norm": 0.010141369886696339, "learning_rate": 2e-05, "loss": 1.5161, "step": 377 }, { "epoch": 0.6780269058295965, "grad_norm": 0.01027351152151823, "learning_rate": 2e-05, "loss": 1.5119, "step": 378 }, { "epoch": 0.6798206278026906, "grad_norm": 0.010362266562879086, "learning_rate": 2e-05, "loss": 1.5126, "step": 379 }, { "epoch": 0.6816143497757847, "grad_norm": 0.010336722247302532, "learning_rate": 2e-05, "loss": 1.5173, "step": 380 }, { "epoch": 0.683408071748879, "grad_norm": 0.01007298193871975, "learning_rate": 2e-05, "loss": 1.5075, "step": 381 }, { "epoch": 0.6852017937219731, "grad_norm": 0.010275410488247871, "learning_rate": 2e-05, "loss": 1.5123, "step": 382 }, { "epoch": 0.6869955156950672, "grad_norm": 0.010203160345554352, "learning_rate": 2e-05, "loss": 1.5151, "step": 383 }, { "epoch": 0.6887892376681615, "grad_norm": 0.010127630084753036, "learning_rate": 2e-05, "loss": 1.5211, "step": 384 }, { "epoch": 0.6905829596412556, "grad_norm": 0.009799284860491753, "learning_rate": 2e-05, "loss": 1.5191, "step": 385 }, { "epoch": 0.6923766816143497, "grad_norm": 0.01014394499361515, "learning_rate": 2e-05, "loss": 1.5261, "step": 386 }, { "epoch": 0.694170403587444, "grad_norm": 0.010567774064838886, "learning_rate": 2e-05, "loss": 1.5232, "step": 387 }, { "epoch": 0.6959641255605381, "grad_norm": 0.010051852092146873, "learning_rate": 2e-05, "loss": 1.5212, "step": 388 }, { "epoch": 0.6977578475336322, "grad_norm": 0.010241293348371983, "learning_rate": 2e-05, "loss": 1.5094, "step": 389 }, { "epoch": 0.6995515695067265, "grad_norm": 0.0095717404037714, "learning_rate": 2e-05, "loss": 1.5115, "step": 390 }, { "epoch": 0.7013452914798206, "grad_norm": 0.00974031537771225, "learning_rate": 2e-05, "loss": 1.5195, "step": 391 }, { "epoch": 0.7031390134529149, "grad_norm": 0.010140657424926758, "learning_rate": 2e-05, "loss": 1.5048, "step": 392 }, { "epoch": 0.704932735426009, "grad_norm": 0.010055477730929852, "learning_rate": 2e-05, "loss": 1.5162, "step": 393 }, { "epoch": 0.7067264573991031, "grad_norm": 0.01005468424409628, "learning_rate": 2e-05, "loss": 1.5258, "step": 394 }, { "epoch": 0.7085201793721974, "grad_norm": 0.010284669697284698, "learning_rate": 2e-05, "loss": 1.5094, "step": 395 }, { "epoch": 0.7103139013452915, "grad_norm": 0.010200968012213707, "learning_rate": 2e-05, "loss": 1.5172, "step": 396 }, { "epoch": 0.7121076233183856, "grad_norm": 0.01015354972332716, "learning_rate": 2e-05, "loss": 1.5117, "step": 397 }, { "epoch": 0.7139013452914799, "grad_norm": 0.009913373738527298, "learning_rate": 2e-05, "loss": 1.5268, "step": 398 }, { "epoch": 0.715695067264574, "grad_norm": 0.010287330485880375, "learning_rate": 2e-05, "loss": 1.5211, "step": 399 }, { "epoch": 0.7174887892376681, "grad_norm": 0.01057345885783434, "learning_rate": 2e-05, "loss": 1.5199, "step": 400 }, { "epoch": 0.7192825112107624, "grad_norm": 0.010113878175616264, "learning_rate": 2e-05, "loss": 1.5168, "step": 401 }, { "epoch": 0.7210762331838565, "grad_norm": 0.009940318763256073, "learning_rate": 2e-05, "loss": 1.5175, "step": 402 }, { "epoch": 0.7228699551569506, "grad_norm": 0.010180394165217876, "learning_rate": 2e-05, "loss": 1.5211, "step": 403 }, { "epoch": 0.7246636771300449, "grad_norm": 0.00961736124008894, "learning_rate": 2e-05, "loss": 1.5228, "step": 404 }, { "epoch": 0.726457399103139, "grad_norm": 0.010378845036029816, "learning_rate": 2e-05, "loss": 1.522, "step": 405 }, { "epoch": 0.7282511210762331, "grad_norm": 0.010189516469836235, "learning_rate": 2e-05, "loss": 1.525, "step": 406 }, { "epoch": 0.7300448430493274, "grad_norm": 0.010004358366131783, "learning_rate": 2e-05, "loss": 1.5172, "step": 407 }, { "epoch": 0.7318385650224215, "grad_norm": 0.010387993417680264, "learning_rate": 2e-05, "loss": 1.5246, "step": 408 }, { "epoch": 0.7336322869955157, "grad_norm": 0.010004810988903046, "learning_rate": 2e-05, "loss": 1.5132, "step": 409 }, { "epoch": 0.7354260089686099, "grad_norm": 0.009845850057899952, "learning_rate": 2e-05, "loss": 1.5248, "step": 410 }, { "epoch": 0.737219730941704, "grad_norm": 0.010015097446739674, "learning_rate": 2e-05, "loss": 1.5196, "step": 411 }, { "epoch": 0.7390134529147983, "grad_norm": 0.009975203312933445, "learning_rate": 2e-05, "loss": 1.5096, "step": 412 }, { "epoch": 0.7408071748878924, "grad_norm": 0.010078891180455685, "learning_rate": 2e-05, "loss": 1.5162, "step": 413 }, { "epoch": 0.7426008968609865, "grad_norm": 0.011885426007211208, "learning_rate": 2e-05, "loss": 1.5189, "step": 414 }, { "epoch": 0.7443946188340808, "grad_norm": 0.009693853557109833, "learning_rate": 2e-05, "loss": 1.5194, "step": 415 }, { "epoch": 0.7461883408071749, "grad_norm": 0.010337116196751595, "learning_rate": 2e-05, "loss": 1.5191, "step": 416 }, { "epoch": 0.747982062780269, "grad_norm": 0.00993486400693655, "learning_rate": 2e-05, "loss": 1.5177, "step": 417 }, { "epoch": 0.7497757847533633, "grad_norm": 0.010143253020942211, "learning_rate": 2e-05, "loss": 1.514, "step": 418 }, { "epoch": 0.7515695067264574, "grad_norm": 0.010233073495328426, "learning_rate": 2e-05, "loss": 1.5154, "step": 419 }, { "epoch": 0.7533632286995515, "grad_norm": 0.009982983581721783, "learning_rate": 2e-05, "loss": 1.5223, "step": 420 }, { "epoch": 0.7551569506726458, "grad_norm": 0.010409766808152199, "learning_rate": 2e-05, "loss": 1.5152, "step": 421 }, { "epoch": 0.7569506726457399, "grad_norm": 0.0099264495074749, "learning_rate": 2e-05, "loss": 1.5185, "step": 422 }, { "epoch": 0.758744394618834, "grad_norm": 0.009928545914590359, "learning_rate": 2e-05, "loss": 1.4986, "step": 423 }, { "epoch": 0.7605381165919283, "grad_norm": 0.009940563701093197, "learning_rate": 2e-05, "loss": 1.5071, "step": 424 }, { "epoch": 0.7623318385650224, "grad_norm": 0.010767797008156776, "learning_rate": 2e-05, "loss": 1.5006, "step": 425 }, { "epoch": 0.7641255605381166, "grad_norm": 0.010551121085882187, "learning_rate": 2e-05, "loss": 1.5201, "step": 426 }, { "epoch": 0.7659192825112108, "grad_norm": 0.010118665173649788, "learning_rate": 2e-05, "loss": 1.5213, "step": 427 }, { "epoch": 0.7677130044843049, "grad_norm": 0.010247626341879368, "learning_rate": 2e-05, "loss": 1.5178, "step": 428 }, { "epoch": 0.7695067264573991, "grad_norm": 0.010188435204327106, "learning_rate": 2e-05, "loss": 1.5085, "step": 429 }, { "epoch": 0.7713004484304933, "grad_norm": 0.010428003035485744, "learning_rate": 2e-05, "loss": 1.5124, "step": 430 }, { "epoch": 0.7730941704035874, "grad_norm": 0.01012035645544529, "learning_rate": 2e-05, "loss": 1.5299, "step": 431 }, { "epoch": 0.7748878923766817, "grad_norm": 0.010584665462374687, "learning_rate": 2e-05, "loss": 1.5095, "step": 432 }, { "epoch": 0.7766816143497758, "grad_norm": 0.009979243390262127, "learning_rate": 2e-05, "loss": 1.5193, "step": 433 }, { "epoch": 0.7784753363228699, "grad_norm": 0.00958004966378212, "learning_rate": 2e-05, "loss": 1.5214, "step": 434 }, { "epoch": 0.7802690582959642, "grad_norm": 0.00973733700811863, "learning_rate": 2e-05, "loss": 1.5208, "step": 435 }, { "epoch": 0.7820627802690583, "grad_norm": 0.010465665720403194, "learning_rate": 2e-05, "loss": 1.5227, "step": 436 }, { "epoch": 0.7838565022421524, "grad_norm": 0.010098133236169815, "learning_rate": 2e-05, "loss": 1.5248, "step": 437 }, { "epoch": 0.7856502242152467, "grad_norm": 0.10259313136339188, "learning_rate": 2e-05, "loss": 1.5222, "step": 438 }, { "epoch": 0.7874439461883408, "grad_norm": 0.01040815282613039, "learning_rate": 2e-05, "loss": 1.5205, "step": 439 }, { "epoch": 0.7892376681614349, "grad_norm": 0.010325520299375057, "learning_rate": 2e-05, "loss": 1.5189, "step": 440 }, { "epoch": 0.7910313901345292, "grad_norm": 0.010079775005578995, "learning_rate": 2e-05, "loss": 1.5156, "step": 441 }, { "epoch": 0.7928251121076233, "grad_norm": 0.010167201980948448, "learning_rate": 2e-05, "loss": 1.5116, "step": 442 }, { "epoch": 0.7946188340807175, "grad_norm": 0.010806124657392502, "learning_rate": 2e-05, "loss": 1.5153, "step": 443 }, { "epoch": 0.7964125560538117, "grad_norm": 0.010324080474674702, "learning_rate": 2e-05, "loss": 1.5246, "step": 444 }, { "epoch": 0.7982062780269058, "grad_norm": 0.010092305950820446, "learning_rate": 2e-05, "loss": 1.5282, "step": 445 }, { "epoch": 0.8, "grad_norm": 0.01007048413157463, "learning_rate": 2e-05, "loss": 1.5108, "step": 446 }, { "epoch": 0.8017937219730942, "grad_norm": 0.010184276849031448, "learning_rate": 2e-05, "loss": 1.51, "step": 447 }, { "epoch": 0.8035874439461883, "grad_norm": 0.010521662421524525, "learning_rate": 2e-05, "loss": 1.5139, "step": 448 }, { "epoch": 0.8053811659192825, "grad_norm": 0.010600044392049313, "learning_rate": 2e-05, "loss": 1.5091, "step": 449 }, { "epoch": 0.8071748878923767, "grad_norm": 0.009714100509881973, "learning_rate": 2e-05, "loss": 1.5122, "step": 450 }, { "epoch": 0.8089686098654708, "grad_norm": 0.010295005515217781, "learning_rate": 2e-05, "loss": 1.52, "step": 451 }, { "epoch": 0.810762331838565, "grad_norm": 0.010034569539129734, "learning_rate": 2e-05, "loss": 1.5197, "step": 452 }, { "epoch": 0.8125560538116592, "grad_norm": 0.010086962021887302, "learning_rate": 2e-05, "loss": 1.5117, "step": 453 }, { "epoch": 0.8143497757847533, "grad_norm": 0.010277335532009602, "learning_rate": 2e-05, "loss": 1.5033, "step": 454 }, { "epoch": 0.8161434977578476, "grad_norm": 0.010540721006691456, "learning_rate": 2e-05, "loss": 1.5166, "step": 455 }, { "epoch": 0.8179372197309417, "grad_norm": 0.009755424223840237, "learning_rate": 2e-05, "loss": 1.5149, "step": 456 }, { "epoch": 0.8197309417040358, "grad_norm": 0.00984253827482462, "learning_rate": 2e-05, "loss": 1.5093, "step": 457 }, { "epoch": 0.8215246636771301, "grad_norm": 0.009836334735155106, "learning_rate": 2e-05, "loss": 1.5141, "step": 458 }, { "epoch": 0.8233183856502242, "grad_norm": 0.01032332144677639, "learning_rate": 2e-05, "loss": 1.5241, "step": 459 }, { "epoch": 0.8251121076233184, "grad_norm": 0.010635129176080227, "learning_rate": 2e-05, "loss": 1.5068, "step": 460 }, { "epoch": 0.8269058295964126, "grad_norm": 0.009664127603173256, "learning_rate": 2e-05, "loss": 1.5052, "step": 461 }, { "epoch": 0.8286995515695067, "grad_norm": 0.010554889217019081, "learning_rate": 2e-05, "loss": 1.5071, "step": 462 }, { "epoch": 0.8304932735426009, "grad_norm": 0.009871057234704494, "learning_rate": 2e-05, "loss": 1.5189, "step": 463 }, { "epoch": 0.8322869955156951, "grad_norm": 0.010431516915559769, "learning_rate": 2e-05, "loss": 1.5183, "step": 464 }, { "epoch": 0.8340807174887892, "grad_norm": 0.009860005229711533, "learning_rate": 2e-05, "loss": 1.5213, "step": 465 }, { "epoch": 0.8358744394618834, "grad_norm": 0.010233579203486443, "learning_rate": 2e-05, "loss": 1.5182, "step": 466 }, { "epoch": 0.8376681614349776, "grad_norm": 0.010311591438949108, "learning_rate": 2e-05, "loss": 1.5092, "step": 467 }, { "epoch": 0.8394618834080717, "grad_norm": 0.010733729228377342, "learning_rate": 2e-05, "loss": 1.5186, "step": 468 }, { "epoch": 0.841255605381166, "grad_norm": 0.009951340965926647, "learning_rate": 2e-05, "loss": 1.5097, "step": 469 }, { "epoch": 0.8430493273542601, "grad_norm": 0.01003777701407671, "learning_rate": 2e-05, "loss": 1.5173, "step": 470 }, { "epoch": 0.8448430493273542, "grad_norm": 0.009939250536262989, "learning_rate": 2e-05, "loss": 1.5108, "step": 471 }, { "epoch": 0.8466367713004485, "grad_norm": 0.009835812263190746, "learning_rate": 2e-05, "loss": 1.5272, "step": 472 }, { "epoch": 0.8484304932735426, "grad_norm": 0.010321546345949173, "learning_rate": 2e-05, "loss": 1.5193, "step": 473 }, { "epoch": 0.8502242152466367, "grad_norm": 0.01006554439663887, "learning_rate": 2e-05, "loss": 1.5165, "step": 474 }, { "epoch": 0.852017937219731, "grad_norm": 0.009972809813916683, "learning_rate": 2e-05, "loss": 1.5228, "step": 475 }, { "epoch": 0.8538116591928251, "grad_norm": 0.010388972237706184, "learning_rate": 2e-05, "loss": 1.5188, "step": 476 }, { "epoch": 0.8556053811659193, "grad_norm": 0.010111154057085514, "learning_rate": 2e-05, "loss": 1.5199, "step": 477 }, { "epoch": 0.8573991031390135, "grad_norm": 0.01029327604919672, "learning_rate": 2e-05, "loss": 1.516, "step": 478 }, { "epoch": 0.8591928251121076, "grad_norm": 0.010400544852018356, "learning_rate": 2e-05, "loss": 1.5218, "step": 479 }, { "epoch": 0.8609865470852018, "grad_norm": 0.0099885743111372, "learning_rate": 2e-05, "loss": 1.5155, "step": 480 }, { "epoch": 0.862780269058296, "grad_norm": 0.010007279925048351, "learning_rate": 2e-05, "loss": 1.5205, "step": 481 }, { "epoch": 0.8645739910313901, "grad_norm": 0.01053563691675663, "learning_rate": 2e-05, "loss": 1.5019, "step": 482 }, { "epoch": 0.8663677130044843, "grad_norm": 0.01031608134508133, "learning_rate": 2e-05, "loss": 1.5217, "step": 483 }, { "epoch": 0.8681614349775785, "grad_norm": 0.010082092136144638, "learning_rate": 2e-05, "loss": 1.5073, "step": 484 }, { "epoch": 0.8699551569506726, "grad_norm": 0.01012254785746336, "learning_rate": 2e-05, "loss": 1.5101, "step": 485 }, { "epoch": 0.8717488789237668, "grad_norm": 0.010539901442825794, "learning_rate": 2e-05, "loss": 1.5209, "step": 486 }, { "epoch": 0.873542600896861, "grad_norm": 0.009883386082947254, "learning_rate": 2e-05, "loss": 1.5275, "step": 487 }, { "epoch": 0.8753363228699551, "grad_norm": 0.010055874474346638, "learning_rate": 2e-05, "loss": 1.521, "step": 488 }, { "epoch": 0.8771300448430494, "grad_norm": 0.010441599413752556, "learning_rate": 2e-05, "loss": 1.5253, "step": 489 }, { "epoch": 0.8789237668161435, "grad_norm": 0.010321282781660557, "learning_rate": 2e-05, "loss": 1.5128, "step": 490 }, { "epoch": 0.8807174887892377, "grad_norm": 0.010404079221189022, "learning_rate": 2e-05, "loss": 1.5216, "step": 491 }, { "epoch": 0.8825112107623319, "grad_norm": 0.010680857114493847, "learning_rate": 2e-05, "loss": 1.5102, "step": 492 }, { "epoch": 0.884304932735426, "grad_norm": 0.009785238653421402, "learning_rate": 2e-05, "loss": 1.5152, "step": 493 }, { "epoch": 0.8860986547085202, "grad_norm": 0.010622934438288212, "learning_rate": 2e-05, "loss": 1.5134, "step": 494 }, { "epoch": 0.8878923766816144, "grad_norm": 0.009563595987856388, "learning_rate": 2e-05, "loss": 1.5213, "step": 495 }, { "epoch": 0.8896860986547085, "grad_norm": 0.009900403209030628, "learning_rate": 2e-05, "loss": 1.5254, "step": 496 }, { "epoch": 0.8914798206278027, "grad_norm": 0.010441206395626068, "learning_rate": 2e-05, "loss": 1.5042, "step": 497 }, { "epoch": 0.8932735426008969, "grad_norm": 0.010110273025929928, "learning_rate": 2e-05, "loss": 1.5141, "step": 498 }, { "epoch": 0.895067264573991, "grad_norm": 0.00976527575403452, "learning_rate": 2e-05, "loss": 1.5189, "step": 499 }, { "epoch": 0.8968609865470852, "grad_norm": 0.010270185768604279, "learning_rate": 2e-05, "loss": 1.5128, "step": 500 }, { "epoch": 0.8986547085201794, "grad_norm": 0.010477078147232533, "learning_rate": 2e-05, "loss": 1.5331, "step": 501 }, { "epoch": 0.9004484304932735, "grad_norm": 0.009786723181605339, "learning_rate": 2e-05, "loss": 1.5143, "step": 502 }, { "epoch": 0.9022421524663677, "grad_norm": 0.009838691912591457, "learning_rate": 2e-05, "loss": 1.5237, "step": 503 }, { "epoch": 0.9040358744394619, "grad_norm": 0.010305250994861126, "learning_rate": 2e-05, "loss": 1.5236, "step": 504 }, { "epoch": 0.905829596412556, "grad_norm": 0.010098317638039589, "learning_rate": 2e-05, "loss": 1.5189, "step": 505 }, { "epoch": 0.9076233183856502, "grad_norm": 0.010335841216146946, "learning_rate": 2e-05, "loss": 1.519, "step": 506 }, { "epoch": 0.9094170403587444, "grad_norm": 0.009809168055653572, "learning_rate": 2e-05, "loss": 1.5176, "step": 507 }, { "epoch": 0.9112107623318386, "grad_norm": 0.01069081760942936, "learning_rate": 2e-05, "loss": 1.5055, "step": 508 }, { "epoch": 0.9130044843049328, "grad_norm": 0.009927291423082352, "learning_rate": 2e-05, "loss": 1.5224, "step": 509 }, { "epoch": 0.9147982062780269, "grad_norm": 0.010560589842498302, "learning_rate": 2e-05, "loss": 1.5129, "step": 510 }, { "epoch": 0.9165919282511211, "grad_norm": 0.010154438205063343, "learning_rate": 2e-05, "loss": 1.52, "step": 511 }, { "epoch": 0.9183856502242153, "grad_norm": 0.010346156544983387, "learning_rate": 2e-05, "loss": 1.5194, "step": 512 }, { "epoch": 0.9201793721973094, "grad_norm": 0.010523281060159206, "learning_rate": 2e-05, "loss": 1.5187, "step": 513 }, { "epoch": 0.9219730941704036, "grad_norm": 0.010443002916872501, "learning_rate": 2e-05, "loss": 1.5059, "step": 514 }, { "epoch": 0.9237668161434978, "grad_norm": 0.010005362331867218, "learning_rate": 2e-05, "loss": 1.5102, "step": 515 }, { "epoch": 0.9255605381165919, "grad_norm": 0.010285025462508202, "learning_rate": 2e-05, "loss": 1.5217, "step": 516 }, { "epoch": 0.9273542600896861, "grad_norm": 0.010401098988950253, "learning_rate": 2e-05, "loss": 1.5243, "step": 517 }, { "epoch": 0.9291479820627803, "grad_norm": 0.010455128736793995, "learning_rate": 2e-05, "loss": 1.5054, "step": 518 }, { "epoch": 0.9309417040358744, "grad_norm": 0.00987928081303835, "learning_rate": 2e-05, "loss": 1.5053, "step": 519 }, { "epoch": 0.9327354260089686, "grad_norm": 0.010212692432105541, "learning_rate": 2e-05, "loss": 1.5151, "step": 520 }, { "epoch": 0.9345291479820628, "grad_norm": 0.010937588289380074, "learning_rate": 2e-05, "loss": 1.5099, "step": 521 }, { "epoch": 0.9363228699551569, "grad_norm": 0.010248001664876938, "learning_rate": 2e-05, "loss": 1.5256, "step": 522 }, { "epoch": 0.9381165919282511, "grad_norm": 0.010430903173983097, "learning_rate": 2e-05, "loss": 1.5056, "step": 523 }, { "epoch": 0.9399103139013453, "grad_norm": 0.0102499695494771, "learning_rate": 2e-05, "loss": 1.5285, "step": 524 }, { "epoch": 0.9417040358744395, "grad_norm": 0.010674213990569115, "learning_rate": 2e-05, "loss": 1.5136, "step": 525 }, { "epoch": 0.9434977578475336, "grad_norm": 0.010732615366578102, "learning_rate": 2e-05, "loss": 1.5119, "step": 526 }, { "epoch": 0.9452914798206278, "grad_norm": 0.009994648396968842, "learning_rate": 2e-05, "loss": 1.5228, "step": 527 }, { "epoch": 0.947085201793722, "grad_norm": 0.010234368033707142, "learning_rate": 2e-05, "loss": 1.5258, "step": 528 }, { "epoch": 0.9488789237668162, "grad_norm": 0.010327205993235111, "learning_rate": 2e-05, "loss": 1.5156, "step": 529 }, { "epoch": 0.9506726457399103, "grad_norm": 0.009836922399699688, "learning_rate": 2e-05, "loss": 1.5171, "step": 530 }, { "epoch": 0.9524663677130045, "grad_norm": 0.009962068870663643, "learning_rate": 2e-05, "loss": 1.5125, "step": 531 }, { "epoch": 0.9542600896860987, "grad_norm": 0.010127882473170757, "learning_rate": 2e-05, "loss": 1.5182, "step": 532 }, { "epoch": 0.9560538116591928, "grad_norm": 0.010251611471176147, "learning_rate": 2e-05, "loss": 1.5139, "step": 533 }, { "epoch": 0.957847533632287, "grad_norm": 0.010081682354211807, "learning_rate": 2e-05, "loss": 1.5239, "step": 534 }, { "epoch": 0.9596412556053812, "grad_norm": 0.010235367342829704, "learning_rate": 2e-05, "loss": 1.5159, "step": 535 }, { "epoch": 0.9614349775784753, "grad_norm": 0.009694702923297882, "learning_rate": 2e-05, "loss": 1.5174, "step": 536 }, { "epoch": 0.9632286995515695, "grad_norm": 0.010224996134638786, "learning_rate": 2e-05, "loss": 1.5171, "step": 537 }, { "epoch": 0.9650224215246637, "grad_norm": 0.010206632316112518, "learning_rate": 2e-05, "loss": 1.5223, "step": 538 }, { "epoch": 0.9668161434977578, "grad_norm": 0.010011864826083183, "learning_rate": 2e-05, "loss": 1.5282, "step": 539 }, { "epoch": 0.968609865470852, "grad_norm": 0.010364921763539314, "learning_rate": 2e-05, "loss": 1.5092, "step": 540 }, { "epoch": 0.9704035874439462, "grad_norm": 0.010109508410096169, "learning_rate": 2e-05, "loss": 1.5068, "step": 541 }, { "epoch": 0.9721973094170404, "grad_norm": 0.00964987464249134, "learning_rate": 2e-05, "loss": 1.5089, "step": 542 }, { "epoch": 0.9739910313901345, "grad_norm": 0.010244207456707954, "learning_rate": 2e-05, "loss": 1.5217, "step": 543 }, { "epoch": 0.9757847533632287, "grad_norm": 0.009797874838113785, "learning_rate": 2e-05, "loss": 1.5143, "step": 544 }, { "epoch": 0.9775784753363229, "grad_norm": 0.010056640952825546, "learning_rate": 2e-05, "loss": 1.5276, "step": 545 }, { "epoch": 0.979372197309417, "grad_norm": 0.009898710064589977, "learning_rate": 2e-05, "loss": 1.5222, "step": 546 }, { "epoch": 0.9811659192825112, "grad_norm": 0.0099082225933671, "learning_rate": 2e-05, "loss": 1.5276, "step": 547 }, { "epoch": 0.9829596412556054, "grad_norm": 0.01018478162586689, "learning_rate": 2e-05, "loss": 1.5217, "step": 548 }, { "epoch": 0.9847533632286996, "grad_norm": 0.009828625246882439, "learning_rate": 2e-05, "loss": 1.5194, "step": 549 }, { "epoch": 0.9865470852017937, "grad_norm": 0.010311014950275421, "learning_rate": 2e-05, "loss": 1.5138, "step": 550 }, { "epoch": 0.9883408071748879, "grad_norm": 0.010840130038559437, "learning_rate": 2e-05, "loss": 1.5044, "step": 551 }, { "epoch": 0.9901345291479821, "grad_norm": 0.009595104493200779, "learning_rate": 2e-05, "loss": 1.5165, "step": 552 }, { "epoch": 0.9919282511210762, "grad_norm": 0.01027593482285738, "learning_rate": 2e-05, "loss": 1.5291, "step": 553 }, { "epoch": 0.9937219730941704, "grad_norm": 0.010394555516541004, "learning_rate": 2e-05, "loss": 1.5109, "step": 554 }, { "epoch": 0.9955156950672646, "grad_norm": 0.00996735692024231, "learning_rate": 2e-05, "loss": 1.5212, "step": 555 }, { "epoch": 0.9973094170403587, "grad_norm": 0.010095257312059402, "learning_rate": 2e-05, "loss": 1.5106, "step": 556 }, { "epoch": 0.9991031390134529, "grad_norm": 0.01082176435738802, "learning_rate": 2e-05, "loss": 1.5099, "step": 557 }, { "epoch": 0.9991031390134529, "step": 557, "total_flos": 7841700554735616.0, "train_loss": 0.0, "train_runtime": 1.3623, "train_samples_per_second": 104762.356, "train_steps_per_second": 408.87 } ], "logging_steps": 1, "max_steps": 557, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7841700554735616.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }