qwen_14b_binary / trainer_state.json
jinqij's picture
Upload folder using huggingface_hub
6e196dc verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 3289,
"global_step": 16443,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 24.78779624251123,
"learning_rate": 6.079027355623101e-10,
"loss": 2.6939,
"step": 1
},
{
"epoch": 0.01,
"grad_norm": 35.873109917628135,
"learning_rate": 1.9452887537993922e-08,
"loss": 3.4616,
"step": 32
},
{
"epoch": 0.01,
"grad_norm": 54.36320326793996,
"learning_rate": 3.8905775075987844e-08,
"loss": 3.544,
"step": 64
},
{
"epoch": 0.02,
"grad_norm": 26.680119386381374,
"learning_rate": 5.8358662613981756e-08,
"loss": 3.3751,
"step": 96
},
{
"epoch": 0.02,
"grad_norm": 21.12845447174909,
"learning_rate": 7.781155015197569e-08,
"loss": 3.5534,
"step": 128
},
{
"epoch": 0.03,
"grad_norm": 14.326825936063585,
"learning_rate": 9.72644376899696e-08,
"loss": 3.2669,
"step": 160
},
{
"epoch": 0.04,
"grad_norm": 25.878315201522145,
"learning_rate": 1.1671732522796351e-07,
"loss": 2.9778,
"step": 192
},
{
"epoch": 0.04,
"grad_norm": 36.45201708374618,
"learning_rate": 1.3617021276595745e-07,
"loss": 2.5704,
"step": 224
},
{
"epoch": 0.05,
"grad_norm": 23.711566820577637,
"learning_rate": 1.5562310030395138e-07,
"loss": 1.3894,
"step": 256
},
{
"epoch": 0.05,
"grad_norm": 4.875938314628448,
"learning_rate": 1.7507598784194527e-07,
"loss": 0.6738,
"step": 288
},
{
"epoch": 0.06,
"grad_norm": 6.241944245269492,
"learning_rate": 1.945288753799392e-07,
"loss": 0.5629,
"step": 320
},
{
"epoch": 0.06,
"grad_norm": 2.8778836099761524,
"learning_rate": 2.1398176291793313e-07,
"loss": 0.4872,
"step": 352
},
{
"epoch": 0.07,
"grad_norm": 1.5436867519527993,
"learning_rate": 2.3343465045592702e-07,
"loss": 0.4714,
"step": 384
},
{
"epoch": 0.08,
"grad_norm": 4.3096231574697645,
"learning_rate": 2.52887537993921e-07,
"loss": 0.4378,
"step": 416
},
{
"epoch": 0.08,
"grad_norm": 1.9666937909371047,
"learning_rate": 2.723404255319149e-07,
"loss": 0.4291,
"step": 448
},
{
"epoch": 0.09,
"grad_norm": 2.2115158499878724,
"learning_rate": 2.917933130699088e-07,
"loss": 0.3954,
"step": 480
},
{
"epoch": 0.09,
"grad_norm": 1.7998977018189959,
"learning_rate": 3.1124620060790275e-07,
"loss": 0.4086,
"step": 512
},
{
"epoch": 0.1,
"grad_norm": 1.8295831973352625,
"learning_rate": 3.3069908814589665e-07,
"loss": 0.3926,
"step": 544
},
{
"epoch": 0.11,
"grad_norm": 2.14598569285712,
"learning_rate": 3.5015197568389055e-07,
"loss": 0.4026,
"step": 576
},
{
"epoch": 0.11,
"grad_norm": 1.5109663276292344,
"learning_rate": 3.696048632218845e-07,
"loss": 0.3688,
"step": 608
},
{
"epoch": 0.12,
"grad_norm": 2.453154442640789,
"learning_rate": 3.890577507598784e-07,
"loss": 0.3919,
"step": 640
},
{
"epoch": 0.12,
"grad_norm": 1.6423681780897064,
"learning_rate": 4.085106382978723e-07,
"loss": 0.364,
"step": 672
},
{
"epoch": 0.13,
"grad_norm": 1.8702893489598191,
"learning_rate": 4.2796352583586625e-07,
"loss": 0.3543,
"step": 704
},
{
"epoch": 0.13,
"grad_norm": 1.8401881440614818,
"learning_rate": 4.4741641337386015e-07,
"loss": 0.3601,
"step": 736
},
{
"epoch": 0.14,
"grad_norm": 1.850566734809788,
"learning_rate": 4.6686930091185405e-07,
"loss": 0.3566,
"step": 768
},
{
"epoch": 0.15,
"grad_norm": 1.6994281914890326,
"learning_rate": 4.86322188449848e-07,
"loss": 0.3505,
"step": 800
},
{
"epoch": 0.15,
"grad_norm": 1.8680937366461803,
"learning_rate": 5.05775075987842e-07,
"loss": 0.3358,
"step": 832
},
{
"epoch": 0.16,
"grad_norm": 1.660364495604376,
"learning_rate": 5.252279635258359e-07,
"loss": 0.3312,
"step": 864
},
{
"epoch": 0.16,
"grad_norm": 2.562680158872079,
"learning_rate": 5.446808510638298e-07,
"loss": 0.3346,
"step": 896
},
{
"epoch": 0.17,
"grad_norm": 1.6948446357514406,
"learning_rate": 5.641337386018237e-07,
"loss": 0.3356,
"step": 928
},
{
"epoch": 0.18,
"grad_norm": 1.209363365641442,
"learning_rate": 5.835866261398176e-07,
"loss": 0.3279,
"step": 960
},
{
"epoch": 0.18,
"grad_norm": 1.6925812671610025,
"learning_rate": 6.030395136778115e-07,
"loss": 0.3367,
"step": 992
},
{
"epoch": 0.19,
"grad_norm": 1.5391435373855369,
"learning_rate": 6.224924012158055e-07,
"loss": 0.3356,
"step": 1024
},
{
"epoch": 0.19,
"grad_norm": 1.4744316930455266,
"learning_rate": 6.419452887537994e-07,
"loss": 0.3277,
"step": 1056
},
{
"epoch": 0.2,
"grad_norm": 1.7620056653926688,
"learning_rate": 6.613981762917933e-07,
"loss": 0.3307,
"step": 1088
},
{
"epoch": 0.2,
"grad_norm": 1.4025350626099162,
"learning_rate": 6.808510638297872e-07,
"loss": 0.3197,
"step": 1120
},
{
"epoch": 0.21,
"grad_norm": 1.3405530061214819,
"learning_rate": 7.003039513677811e-07,
"loss": 0.3218,
"step": 1152
},
{
"epoch": 0.22,
"grad_norm": 1.4535529320190042,
"learning_rate": 7.197568389057751e-07,
"loss": 0.3067,
"step": 1184
},
{
"epoch": 0.22,
"grad_norm": 1.445157014748904,
"learning_rate": 7.39209726443769e-07,
"loss": 0.3087,
"step": 1216
},
{
"epoch": 0.23,
"grad_norm": 1.5871727851863946,
"learning_rate": 7.586626139817629e-07,
"loss": 0.3252,
"step": 1248
},
{
"epoch": 0.23,
"grad_norm": 1.3848383016325827,
"learning_rate": 7.781155015197568e-07,
"loss": 0.3113,
"step": 1280
},
{
"epoch": 0.24,
"grad_norm": 1.7455453453163468,
"learning_rate": 7.975683890577507e-07,
"loss": 0.297,
"step": 1312
},
{
"epoch": 0.25,
"grad_norm": 1.3001668041428838,
"learning_rate": 8.170212765957446e-07,
"loss": 0.3223,
"step": 1344
},
{
"epoch": 0.25,
"grad_norm": 1.2132377152476534,
"learning_rate": 8.364741641337386e-07,
"loss": 0.315,
"step": 1376
},
{
"epoch": 0.26,
"grad_norm": 1.4384605625322182,
"learning_rate": 8.559270516717325e-07,
"loss": 0.3105,
"step": 1408
},
{
"epoch": 0.26,
"grad_norm": 1.1957556077083222,
"learning_rate": 8.753799392097264e-07,
"loss": 0.3007,
"step": 1440
},
{
"epoch": 0.27,
"grad_norm": 1.466257939308454,
"learning_rate": 8.948328267477203e-07,
"loss": 0.3034,
"step": 1472
},
{
"epoch": 0.27,
"grad_norm": 1.348037306034132,
"learning_rate": 9.142857142857142e-07,
"loss": 0.3044,
"step": 1504
},
{
"epoch": 0.28,
"grad_norm": 1.3312009756781122,
"learning_rate": 9.337386018237081e-07,
"loss": 0.3124,
"step": 1536
},
{
"epoch": 0.29,
"grad_norm": 1.5165497951579907,
"learning_rate": 9.531914893617021e-07,
"loss": 0.3064,
"step": 1568
},
{
"epoch": 0.29,
"grad_norm": 1.2308464212993413,
"learning_rate": 9.72644376899696e-07,
"loss": 0.3096,
"step": 1600
},
{
"epoch": 0.3,
"grad_norm": 1.4477927464029368,
"learning_rate": 9.9209726443769e-07,
"loss": 0.2967,
"step": 1632
},
{
"epoch": 0.3,
"grad_norm": 1.6161236994381232,
"learning_rate": 9.999959323804955e-07,
"loss": 0.307,
"step": 1664
},
{
"epoch": 0.31,
"grad_norm": 1.408597466262752,
"learning_rate": 9.999706931043734e-07,
"loss": 0.3045,
"step": 1696
},
{
"epoch": 0.32,
"grad_norm": 1.2262804987989535,
"learning_rate": 9.999223791032621e-07,
"loss": 0.2982,
"step": 1728
},
{
"epoch": 0.32,
"grad_norm": 1.4772857240369206,
"learning_rate": 9.998509926069568e-07,
"loss": 0.3085,
"step": 1760
},
{
"epoch": 0.33,
"grad_norm": 1.1705510022471537,
"learning_rate": 9.997565369100982e-07,
"loss": 0.3102,
"step": 1792
},
{
"epoch": 0.33,
"grad_norm": 1.354193914272892,
"learning_rate": 9.996390163720203e-07,
"loss": 0.3079,
"step": 1824
},
{
"epoch": 0.34,
"grad_norm": 1.6499740915478518,
"learning_rate": 9.99498436416549e-07,
"loss": 0.3058,
"step": 1856
},
{
"epoch": 0.34,
"grad_norm": 1.4176797632580183,
"learning_rate": 9.993348035317523e-07,
"loss": 0.2898,
"step": 1888
},
{
"epoch": 0.35,
"grad_norm": 1.221569587599335,
"learning_rate": 9.991481252696405e-07,
"loss": 0.2971,
"step": 1920
},
{
"epoch": 0.36,
"grad_norm": 1.6858523944684565,
"learning_rate": 9.98938410245818e-07,
"loss": 0.3015,
"step": 1952
},
{
"epoch": 0.36,
"grad_norm": 1.3071072285598695,
"learning_rate": 9.987056681390846e-07,
"loss": 0.3017,
"step": 1984
},
{
"epoch": 0.37,
"grad_norm": 1.3559736389856838,
"learning_rate": 9.984499096909904e-07,
"loss": 0.2996,
"step": 2016
},
{
"epoch": 0.37,
"grad_norm": 1.6577354687629797,
"learning_rate": 9.981711467053391e-07,
"loss": 0.2899,
"step": 2048
},
{
"epoch": 0.38,
"grad_norm": 1.3911817831183464,
"learning_rate": 9.97869392047643e-07,
"loss": 0.2916,
"step": 2080
},
{
"epoch": 0.39,
"grad_norm": 1.4068725629362402,
"learning_rate": 9.9754465964453e-07,
"loss": 0.3042,
"step": 2112
},
{
"epoch": 0.39,
"grad_norm": 1.2845003819775453,
"learning_rate": 9.971969644831001e-07,
"loss": 0.2953,
"step": 2144
},
{
"epoch": 0.4,
"grad_norm": 1.3034821415736415,
"learning_rate": 9.968263226102348e-07,
"loss": 0.2947,
"step": 2176
},
{
"epoch": 0.4,
"grad_norm": 1.297243001736353,
"learning_rate": 9.96432751131855e-07,
"loss": 0.2849,
"step": 2208
},
{
"epoch": 0.41,
"grad_norm": 1.411186926138334,
"learning_rate": 9.960162682121326e-07,
"loss": 0.2974,
"step": 2240
},
{
"epoch": 0.41,
"grad_norm": 1.2374941457414885,
"learning_rate": 9.955768930726523e-07,
"loss": 0.2824,
"step": 2272
},
{
"epoch": 0.42,
"grad_norm": 1.2627652146626807,
"learning_rate": 9.95114645991523e-07,
"loss": 0.2865,
"step": 2304
},
{
"epoch": 0.43,
"grad_norm": 1.3134029095546937,
"learning_rate": 9.946295483024446e-07,
"loss": 0.2867,
"step": 2336
},
{
"epoch": 0.43,
"grad_norm": 1.125719093465591,
"learning_rate": 9.9412162239372e-07,
"loss": 0.292,
"step": 2368
},
{
"epoch": 0.44,
"grad_norm": 1.4224011133495338,
"learning_rate": 9.93590891707225e-07,
"loss": 0.309,
"step": 2400
},
{
"epoch": 0.44,
"grad_norm": 1.558277250619359,
"learning_rate": 9.930373807373245e-07,
"loss": 0.289,
"step": 2432
},
{
"epoch": 0.45,
"grad_norm": 1.3208871305799181,
"learning_rate": 9.924611150297428e-07,
"loss": 0.2883,
"step": 2464
},
{
"epoch": 0.46,
"grad_norm": 1.3373963931205746,
"learning_rate": 9.91862121180384e-07,
"loss": 0.2808,
"step": 2496
},
{
"epoch": 0.46,
"grad_norm": 1.3910055395902359,
"learning_rate": 9.912404268341051e-07,
"loss": 0.2932,
"step": 2528
},
{
"epoch": 0.47,
"grad_norm": 1.4574028412197861,
"learning_rate": 9.905960606834404e-07,
"loss": 0.2947,
"step": 2560
},
{
"epoch": 0.47,
"grad_norm": 1.203878457595175,
"learning_rate": 9.899290524672763e-07,
"loss": 0.2867,
"step": 2592
},
{
"epoch": 0.48,
"grad_norm": 1.4673267738670204,
"learning_rate": 9.892394329694793e-07,
"loss": 0.2837,
"step": 2624
},
{
"epoch": 0.48,
"grad_norm": 1.6281520652734605,
"learning_rate": 9.885272340174754e-07,
"loss": 0.3032,
"step": 2656
},
{
"epoch": 0.49,
"grad_norm": 1.2522157670744074,
"learning_rate": 9.877924884807814e-07,
"loss": 0.2963,
"step": 2688
},
{
"epoch": 0.5,
"grad_norm": 1.4632043067121088,
"learning_rate": 9.870352302694869e-07,
"loss": 0.2898,
"step": 2720
},
{
"epoch": 0.5,
"grad_norm": 1.398722594166713,
"learning_rate": 9.86255494332691e-07,
"loss": 0.2896,
"step": 2752
},
{
"epoch": 0.51,
"grad_norm": 1.3916529397644415,
"learning_rate": 9.854533166568867e-07,
"loss": 0.2967,
"step": 2784
},
{
"epoch": 0.51,
"grad_norm": 1.3208718268032766,
"learning_rate": 9.846287342643032e-07,
"loss": 0.2838,
"step": 2816
},
{
"epoch": 0.52,
"grad_norm": 1.0718990960806063,
"learning_rate": 9.837817852111949e-07,
"loss": 0.2942,
"step": 2848
},
{
"epoch": 0.53,
"grad_norm": 1.3984255388696125,
"learning_rate": 9.829125085860858e-07,
"loss": 0.2832,
"step": 2880
},
{
"epoch": 0.53,
"grad_norm": 1.258839102452435,
"learning_rate": 9.820209445079655e-07,
"loss": 0.2964,
"step": 2912
},
{
"epoch": 0.54,
"grad_norm": 1.1619142104791713,
"learning_rate": 9.811071341244379e-07,
"loss": 0.2727,
"step": 2944
},
{
"epoch": 0.54,
"grad_norm": 1.3494984294429655,
"learning_rate": 9.801711196098213e-07,
"loss": 0.2848,
"step": 2976
},
{
"epoch": 0.55,
"grad_norm": 1.2501831577892621,
"learning_rate": 9.792129441632027e-07,
"loss": 0.2749,
"step": 3008
},
{
"epoch": 0.55,
"grad_norm": 1.2580708434458154,
"learning_rate": 9.782326520064443e-07,
"loss": 0.2868,
"step": 3040
},
{
"epoch": 0.56,
"grad_norm": 1.5601918050638766,
"learning_rate": 9.772302883821418e-07,
"loss": 0.292,
"step": 3072
},
{
"epoch": 0.57,
"grad_norm": 1.3545732270149806,
"learning_rate": 9.762058995515362e-07,
"loss": 0.2895,
"step": 3104
},
{
"epoch": 0.57,
"grad_norm": 1.4796071501521761,
"learning_rate": 9.751595327923802e-07,
"loss": 0.291,
"step": 3136
},
{
"epoch": 0.58,
"grad_norm": 1.3020677021992657,
"learning_rate": 9.740912363967546e-07,
"loss": 0.294,
"step": 3168
},
{
"epoch": 0.58,
"grad_norm": 1.2903648809439663,
"learning_rate": 9.730010596688405e-07,
"loss": 0.2971,
"step": 3200
},
{
"epoch": 0.59,
"grad_norm": 1.4322652904369149,
"learning_rate": 9.718890529226432e-07,
"loss": 0.2842,
"step": 3232
},
{
"epoch": 0.6,
"grad_norm": 1.4918014229559349,
"learning_rate": 9.707552674796704e-07,
"loss": 0.2855,
"step": 3264
},
{
"epoch": 0.6,
"grad_norm": 1.495160093783148,
"learning_rate": 9.695997556665642e-07,
"loss": 0.2961,
"step": 3296
},
{
"epoch": 0.61,
"grad_norm": 1.2299397480418435,
"learning_rate": 9.68422570812685e-07,
"loss": 0.275,
"step": 3328
},
{
"epoch": 0.61,
"grad_norm": 1.4595570985088122,
"learning_rate": 9.672237672476504e-07,
"loss": 0.2816,
"step": 3360
},
{
"epoch": 0.62,
"grad_norm": 1.2523764393563719,
"learning_rate": 9.660034002988288e-07,
"loss": 0.2826,
"step": 3392
},
{
"epoch": 0.62,
"grad_norm": 1.1556114989117734,
"learning_rate": 9.64761526288785e-07,
"loss": 0.2979,
"step": 3424
},
{
"epoch": 0.63,
"grad_norm": 1.2760487310976472,
"learning_rate": 9.634982025326808e-07,
"loss": 0.2703,
"step": 3456
},
{
"epoch": 0.64,
"grad_norm": 1.1268494754892229,
"learning_rate": 9.622134873356302e-07,
"loss": 0.2888,
"step": 3488
},
{
"epoch": 0.64,
"grad_norm": 1.5177691734350403,
"learning_rate": 9.60907439990008e-07,
"loss": 0.2854,
"step": 3520
},
{
"epoch": 0.65,
"grad_norm": 1.4479572539864909,
"learning_rate": 9.595801207727145e-07,
"loss": 0.2871,
"step": 3552
},
{
"epoch": 0.65,
"grad_norm": 1.3320405553461194,
"learning_rate": 9.58231590942392e-07,
"loss": 0.2722,
"step": 3584
},
{
"epoch": 0.66,
"grad_norm": 1.2684704306064236,
"learning_rate": 9.568619127365979e-07,
"loss": 0.2918,
"step": 3616
},
{
"epoch": 0.67,
"grad_norm": 1.153157224662708,
"learning_rate": 9.554711493689337e-07,
"loss": 0.2727,
"step": 3648
},
{
"epoch": 0.67,
"grad_norm": 1.2053424871803355,
"learning_rate": 9.540593650261259e-07,
"loss": 0.2727,
"step": 3680
},
{
"epoch": 0.68,
"grad_norm": 1.5070136801134912,
"learning_rate": 9.526266248650647e-07,
"loss": 0.2829,
"step": 3712
},
{
"epoch": 0.68,
"grad_norm": 1.3626256868915094,
"learning_rate": 9.511729950097962e-07,
"loss": 0.2948,
"step": 3744
},
{
"epoch": 0.69,
"grad_norm": 1.3413698467022634,
"learning_rate": 9.496985425484708e-07,
"loss": 0.2804,
"step": 3776
},
{
"epoch": 0.69,
"grad_norm": 1.1666360657022152,
"learning_rate": 9.482033355302474e-07,
"loss": 0.2779,
"step": 3808
},
{
"epoch": 0.7,
"grad_norm": 1.2681212059953142,
"learning_rate": 9.466874429621523e-07,
"loss": 0.2766,
"step": 3840
},
{
"epoch": 0.71,
"grad_norm": 1.570024716133246,
"learning_rate": 9.451509348058943e-07,
"loss": 0.278,
"step": 3872
},
{
"epoch": 0.71,
"grad_norm": 1.348864102054253,
"learning_rate": 9.435938819746363e-07,
"loss": 0.2814,
"step": 3904
},
{
"epoch": 0.72,
"grad_norm": 1.479630195353394,
"learning_rate": 9.420163563297221e-07,
"loss": 0.2768,
"step": 3936
},
{
"epoch": 0.72,
"grad_norm": 1.3437829697550214,
"learning_rate": 9.4041843067736e-07,
"loss": 0.2796,
"step": 3968
},
{
"epoch": 0.73,
"grad_norm": 1.7505808749964085,
"learning_rate": 9.388001787652626e-07,
"loss": 0.2804,
"step": 4000
},
{
"epoch": 0.74,
"grad_norm": 1.2804732688993112,
"learning_rate": 9.37161675279243e-07,
"loss": 0.2671,
"step": 4032
},
{
"epoch": 0.74,
"grad_norm": 1.5599945996293794,
"learning_rate": 9.355029958397686e-07,
"loss": 0.2778,
"step": 4064
},
{
"epoch": 0.75,
"grad_norm": 1.140562030900848,
"learning_rate": 9.338242169984701e-07,
"loss": 0.2779,
"step": 4096
},
{
"epoch": 0.75,
"grad_norm": 1.1891602382042474,
"learning_rate": 9.321254162346089e-07,
"loss": 0.2779,
"step": 4128
},
{
"epoch": 0.76,
"grad_norm": 1.6066027745264384,
"learning_rate": 9.304066719515013e-07,
"loss": 0.2831,
"step": 4160
},
{
"epoch": 0.76,
"grad_norm": 1.3535261095585958,
"learning_rate": 9.286680634729005e-07,
"loss": 0.2869,
"step": 4192
},
{
"epoch": 0.77,
"grad_norm": 1.449372671449822,
"learning_rate": 9.269096710393343e-07,
"loss": 0.2859,
"step": 4224
},
{
"epoch": 0.78,
"grad_norm": 1.2209167474794183,
"learning_rate": 9.251315758044032e-07,
"loss": 0.2908,
"step": 4256
},
{
"epoch": 0.78,
"grad_norm": 1.4925119088076193,
"learning_rate": 9.233338598310343e-07,
"loss": 0.2821,
"step": 4288
},
{
"epoch": 0.79,
"grad_norm": 1.2663637068625269,
"learning_rate": 9.215166060876943e-07,
"loss": 0.2839,
"step": 4320
},
{
"epoch": 0.79,
"grad_norm": 1.335605475288019,
"learning_rate": 9.196798984445595e-07,
"loss": 0.2767,
"step": 4352
},
{
"epoch": 0.8,
"grad_norm": 1.1999064453898192,
"learning_rate": 9.178238216696463e-07,
"loss": 0.2813,
"step": 4384
},
{
"epoch": 0.81,
"grad_norm": 1.2005680882302527,
"learning_rate": 9.159484614248978e-07,
"loss": 0.2776,
"step": 4416
},
{
"epoch": 0.81,
"grad_norm": 1.2022881962586156,
"learning_rate": 9.140539042622311e-07,
"loss": 0.2812,
"step": 4448
},
{
"epoch": 0.82,
"grad_norm": 1.4918926856382582,
"learning_rate": 9.121402376195421e-07,
"loss": 0.2763,
"step": 4480
},
{
"epoch": 0.82,
"grad_norm": 1.452212648334464,
"learning_rate": 9.102075498166705e-07,
"loss": 0.2756,
"step": 4512
},
{
"epoch": 0.83,
"grad_norm": 1.4610868927687688,
"learning_rate": 9.082559300513237e-07,
"loss": 0.2755,
"step": 4544
},
{
"epoch": 0.83,
"grad_norm": 1.2140564103776683,
"learning_rate": 9.062854683949602e-07,
"loss": 0.266,
"step": 4576
},
{
"epoch": 0.84,
"grad_norm": 1.23274371490023,
"learning_rate": 9.042962557886313e-07,
"loss": 0.2856,
"step": 4608
},
{
"epoch": 0.85,
"grad_norm": 1.3898063562012308,
"learning_rate": 9.022883840387865e-07,
"loss": 0.2815,
"step": 4640
},
{
"epoch": 0.85,
"grad_norm": 1.330139097581169,
"learning_rate": 9.002619458130339e-07,
"loss": 0.28,
"step": 4672
},
{
"epoch": 0.86,
"grad_norm": 1.2874922030319198,
"learning_rate": 8.982170346358651e-07,
"loss": 0.2669,
"step": 4704
},
{
"epoch": 0.86,
"grad_norm": 1.107586259057668,
"learning_rate": 8.961537448843377e-07,
"loss": 0.2853,
"step": 4736
},
{
"epoch": 0.87,
"grad_norm": 1.448957134616539,
"learning_rate": 8.940721717837205e-07,
"loss": 0.2652,
"step": 4768
},
{
"epoch": 0.88,
"grad_norm": 1.210724298652212,
"learning_rate": 8.919724114030984e-07,
"loss": 0.2795,
"step": 4800
},
{
"epoch": 0.88,
"grad_norm": 1.3698632505550181,
"learning_rate": 8.898545606509378e-07,
"loss": 0.284,
"step": 4832
},
{
"epoch": 0.89,
"grad_norm": 1.1911981203106947,
"learning_rate": 8.877187172706155e-07,
"loss": 0.2746,
"step": 4864
},
{
"epoch": 0.89,
"grad_norm": 1.3333516374957606,
"learning_rate": 8.855649798359064e-07,
"loss": 0.2888,
"step": 4896
},
{
"epoch": 0.9,
"grad_norm": 1.2351888608354114,
"learning_rate": 8.833934477464347e-07,
"loss": 0.2831,
"step": 4928
},
{
"epoch": 0.9,
"grad_norm": 1.3038819344536112,
"learning_rate": 8.812042212230862e-07,
"loss": 0.2753,
"step": 4960
},
{
"epoch": 0.91,
"grad_norm": 1.5707271650867898,
"learning_rate": 8.789974013033831e-07,
"loss": 0.2759,
"step": 4992
},
{
"epoch": 0.92,
"grad_norm": 1.2820047239265864,
"learning_rate": 8.767730898368208e-07,
"loss": 0.2746,
"step": 5024
},
{
"epoch": 0.92,
"grad_norm": 1.1744722963290781,
"learning_rate": 8.745313894801669e-07,
"loss": 0.291,
"step": 5056
},
{
"epoch": 0.93,
"grad_norm": 1.0913528023957022,
"learning_rate": 8.722724036927242e-07,
"loss": 0.2837,
"step": 5088
},
{
"epoch": 0.93,
"grad_norm": 1.3404827415629068,
"learning_rate": 8.699962367315552e-07,
"loss": 0.2676,
"step": 5120
},
{
"epoch": 0.94,
"grad_norm": 1.163041143181648,
"learning_rate": 8.677029936466706e-07,
"loss": 0.2832,
"step": 5152
},
{
"epoch": 0.95,
"grad_norm": 1.436779107568984,
"learning_rate": 8.653927802761809e-07,
"loss": 0.2728,
"step": 5184
},
{
"epoch": 0.95,
"grad_norm": 1.0080451953033644,
"learning_rate": 8.630657032414121e-07,
"loss": 0.2617,
"step": 5216
},
{
"epoch": 0.96,
"grad_norm": 1.1100819749235487,
"learning_rate": 8.607218699419843e-07,
"loss": 0.2942,
"step": 5248
},
{
"epoch": 0.96,
"grad_norm": 1.2992077116418177,
"learning_rate": 8.583613885508556e-07,
"loss": 0.2672,
"step": 5280
},
{
"epoch": 0.97,
"grad_norm": 1.223919929869006,
"learning_rate": 8.559843680093296e-07,
"loss": 0.277,
"step": 5312
},
{
"epoch": 0.98,
"grad_norm": 1.1391386556761285,
"learning_rate": 8.535909180220268e-07,
"loss": 0.2806,
"step": 5344
},
{
"epoch": 0.98,
"grad_norm": 1.2127293368764673,
"learning_rate": 8.511811490518225e-07,
"loss": 0.2619,
"step": 5376
},
{
"epoch": 0.99,
"grad_norm": 1.140029182839281,
"learning_rate": 8.48755172314748e-07,
"loss": 0.2718,
"step": 5408
},
{
"epoch": 0.99,
"grad_norm": 1.4671327877043736,
"learning_rate": 8.463130997748578e-07,
"loss": 0.272,
"step": 5440
},
{
"epoch": 1.0,
"grad_norm": 1.1390017949538402,
"learning_rate": 8.43855044139063e-07,
"loss": 0.2737,
"step": 5472
},
{
"epoch": 1.0,
"grad_norm": 1.2142876905398077,
"learning_rate": 8.413811188519282e-07,
"loss": 0.2599,
"step": 5504
},
{
"epoch": 1.01,
"grad_norm": 1.266246650515794,
"learning_rate": 8.388914380904373e-07,
"loss": 0.2465,
"step": 5536
},
{
"epoch": 1.02,
"grad_norm": 1.173280963516161,
"learning_rate": 8.36386116758723e-07,
"loss": 0.2483,
"step": 5568
},
{
"epoch": 1.02,
"grad_norm": 1.1617564369184084,
"learning_rate": 8.33865270482764e-07,
"loss": 0.2568,
"step": 5600
},
{
"epoch": 1.03,
"grad_norm": 1.420351934994977,
"learning_rate": 8.313290156050487e-07,
"loss": 0.2502,
"step": 5632
},
{
"epoch": 1.03,
"grad_norm": 1.3547439002353128,
"learning_rate": 8.287774691792057e-07,
"loss": 0.246,
"step": 5664
},
{
"epoch": 1.04,
"grad_norm": 1.1455753710552021,
"learning_rate": 8.262107489646014e-07,
"loss": 0.2554,
"step": 5696
},
{
"epoch": 1.05,
"grad_norm": 1.3859247572490137,
"learning_rate": 8.236289734209054e-07,
"loss": 0.2466,
"step": 5728
},
{
"epoch": 1.05,
"grad_norm": 1.1343048895326515,
"learning_rate": 8.210322617026232e-07,
"loss": 0.2467,
"step": 5760
},
{
"epoch": 1.06,
"grad_norm": 1.2217575468560562,
"learning_rate": 8.184207336535967e-07,
"loss": 0.2574,
"step": 5792
},
{
"epoch": 1.06,
"grad_norm": 1.3780997994944977,
"learning_rate": 8.157945098014733e-07,
"loss": 0.2481,
"step": 5824
},
{
"epoch": 1.07,
"grad_norm": 1.0870131072717084,
"learning_rate": 8.131537113521443e-07,
"loss": 0.2588,
"step": 5856
},
{
"epoch": 1.07,
"grad_norm": 1.0490087402219708,
"learning_rate": 8.104984601841489e-07,
"loss": 0.246,
"step": 5888
},
{
"epoch": 1.08,
"grad_norm": 1.2965280317141281,
"learning_rate": 8.07828878843051e-07,
"loss": 0.249,
"step": 5920
},
{
"epoch": 1.09,
"grad_norm": 1.2861279001372838,
"learning_rate": 8.051450905357829e-07,
"loss": 0.2412,
"step": 5952
},
{
"epoch": 1.09,
"grad_norm": 1.409445153252456,
"learning_rate": 8.024472191249588e-07,
"loss": 0.2431,
"step": 5984
},
{
"epoch": 1.1,
"grad_norm": 1.3691683733505722,
"learning_rate": 7.997353891231584e-07,
"loss": 0.2517,
"step": 6016
},
{
"epoch": 1.1,
"grad_norm": 1.4906649500848184,
"learning_rate": 7.970097256871811e-07,
"loss": 0.2517,
"step": 6048
},
{
"epoch": 1.11,
"grad_norm": 1.4438858223954047,
"learning_rate": 7.942703546122682e-07,
"loss": 0.2584,
"step": 6080
},
{
"epoch": 1.12,
"grad_norm": 1.268958450317541,
"learning_rate": 7.915174023262988e-07,
"loss": 0.2522,
"step": 6112
},
{
"epoch": 1.12,
"grad_norm": 1.4781129605275483,
"learning_rate": 7.887509958839538e-07,
"loss": 0.243,
"step": 6144
},
{
"epoch": 1.13,
"grad_norm": 1.331949493045608,
"learning_rate": 7.859712629608524e-07,
"loss": 0.2429,
"step": 6176
},
{
"epoch": 1.13,
"grad_norm": 1.306062446109076,
"learning_rate": 7.831783318476596e-07,
"loss": 0.2446,
"step": 6208
},
{
"epoch": 1.14,
"grad_norm": 1.242511236714611,
"learning_rate": 7.803723314441656e-07,
"loss": 0.2451,
"step": 6240
},
{
"epoch": 1.14,
"grad_norm": 1.4482758408678764,
"learning_rate": 7.775533912533363e-07,
"loss": 0.2412,
"step": 6272
},
{
"epoch": 1.15,
"grad_norm": 1.3719989834400954,
"learning_rate": 7.747216413753366e-07,
"loss": 0.2421,
"step": 6304
},
{
"epoch": 1.16,
"grad_norm": 1.1916507119961282,
"learning_rate": 7.718772125015263e-07,
"loss": 0.2536,
"step": 6336
},
{
"epoch": 1.16,
"grad_norm": 1.2009787273487618,
"learning_rate": 7.690202359084278e-07,
"loss": 0.2501,
"step": 6368
},
{
"epoch": 1.17,
"grad_norm": 1.2823116236203336,
"learning_rate": 7.661508434516682e-07,
"loss": 0.2524,
"step": 6400
},
{
"epoch": 1.17,
"grad_norm": 1.3430035950312693,
"learning_rate": 7.632691675598935e-07,
"loss": 0.2577,
"step": 6432
},
{
"epoch": 1.18,
"grad_norm": 1.4985924089749454,
"learning_rate": 7.60375341228656e-07,
"loss": 0.2444,
"step": 6464
},
{
"epoch": 1.19,
"grad_norm": 1.2984272436604025,
"learning_rate": 7.574694980142779e-07,
"loss": 0.2615,
"step": 6496
},
{
"epoch": 1.19,
"grad_norm": 1.3354550623218928,
"learning_rate": 7.545517720276857e-07,
"loss": 0.2389,
"step": 6528
},
{
"epoch": 1.2,
"grad_norm": 1.4301768294716657,
"learning_rate": 7.51622297928222e-07,
"loss": 0.2449,
"step": 6560
},
{
"epoch": 1.2,
"grad_norm": 1.1719113606920948,
"learning_rate": 7.4868121091743e-07,
"loss": 0.2423,
"step": 6592
},
{
"epoch": 1.21,
"grad_norm": 1.4204627356681534,
"learning_rate": 7.457286467328135e-07,
"loss": 0.2445,
"step": 6624
},
{
"epoch": 1.21,
"grad_norm": 1.4900172188399152,
"learning_rate": 7.427647416415725e-07,
"loss": 0.2483,
"step": 6656
},
{
"epoch": 1.22,
"grad_norm": 1.3961859599185324,
"learning_rate": 7.397896324343151e-07,
"loss": 0.2359,
"step": 6688
},
{
"epoch": 1.23,
"grad_norm": 1.3110300825940397,
"learning_rate": 7.368034564187424e-07,
"loss": 0.2512,
"step": 6720
},
{
"epoch": 1.23,
"grad_norm": 1.2757069881397953,
"learning_rate": 7.338063514133136e-07,
"loss": 0.2547,
"step": 6752
},
{
"epoch": 1.24,
"grad_norm": 1.4074893254437457,
"learning_rate": 7.307984557408837e-07,
"loss": 0.2496,
"step": 6784
},
{
"epoch": 1.24,
"grad_norm": 1.351839197713545,
"learning_rate": 7.277799082223204e-07,
"loss": 0.2353,
"step": 6816
},
{
"epoch": 1.25,
"grad_norm": 1.2030842443572731,
"learning_rate": 7.24750848170097e-07,
"loss": 0.2559,
"step": 6848
},
{
"epoch": 1.26,
"grad_norm": 1.324374188137303,
"learning_rate": 7.217114153818629e-07,
"loss": 0.2448,
"step": 6880
},
{
"epoch": 1.26,
"grad_norm": 1.2343518311233508,
"learning_rate": 7.186617501339917e-07,
"loss": 0.2475,
"step": 6912
},
{
"epoch": 1.27,
"grad_norm": 1.257477506253048,
"learning_rate": 7.156019931751071e-07,
"loss": 0.2529,
"step": 6944
},
{
"epoch": 1.27,
"grad_norm": 1.3065578571620988,
"learning_rate": 7.125322857195867e-07,
"loss": 0.2439,
"step": 6976
},
{
"epoch": 1.28,
"grad_norm": 1.491754680548748,
"learning_rate": 7.094527694410455e-07,
"loss": 0.2581,
"step": 7008
},
{
"epoch": 1.28,
"grad_norm": 1.2313699652632502,
"learning_rate": 7.063635864657964e-07,
"loss": 0.2462,
"step": 7040
},
{
"epoch": 1.29,
"grad_norm": 1.200487097478514,
"learning_rate": 7.032648793662912e-07,
"loss": 0.2422,
"step": 7072
},
{
"epoch": 1.3,
"grad_norm": 1.3322942461823077,
"learning_rate": 7.001567911545406e-07,
"loss": 0.2456,
"step": 7104
},
{
"epoch": 1.3,
"grad_norm": 1.4431169758144065,
"learning_rate": 6.970394652755144e-07,
"loss": 0.2482,
"step": 7136
},
{
"epoch": 1.31,
"grad_norm": 1.261334251262187,
"learning_rate": 6.939130456005196e-07,
"loss": 0.2456,
"step": 7168
},
{
"epoch": 1.31,
"grad_norm": 1.4803616635583337,
"learning_rate": 6.907776764205622e-07,
"loss": 0.2628,
"step": 7200
},
{
"epoch": 1.32,
"grad_norm": 1.318319361538514,
"learning_rate": 6.876335024396871e-07,
"loss": 0.2408,
"step": 7232
},
{
"epoch": 1.33,
"grad_norm": 1.2799133546143286,
"learning_rate": 6.844806687682996e-07,
"loss": 0.2445,
"step": 7264
},
{
"epoch": 1.33,
"grad_norm": 1.2877655780208916,
"learning_rate": 6.813193209164683e-07,
"loss": 0.2406,
"step": 7296
},
{
"epoch": 1.34,
"grad_norm": 1.344111445915826,
"learning_rate": 6.781496047872098e-07,
"loss": 0.2408,
"step": 7328
},
{
"epoch": 1.34,
"grad_norm": 1.5226804349220626,
"learning_rate": 6.749716666697545e-07,
"loss": 0.2469,
"step": 7360
},
{
"epoch": 1.35,
"grad_norm": 1.3170301342228112,
"learning_rate": 6.717856532327956e-07,
"loss": 0.2407,
"step": 7392
},
{
"epoch": 1.35,
"grad_norm": 1.185336007297627,
"learning_rate": 6.685917115177193e-07,
"loss": 0.2445,
"step": 7424
},
{
"epoch": 1.36,
"grad_norm": 1.3491930633425604,
"learning_rate": 6.653899889318192e-07,
"loss": 0.2562,
"step": 7456
},
{
"epoch": 1.37,
"grad_norm": 1.160879383772545,
"learning_rate": 6.621806332414925e-07,
"loss": 0.2427,
"step": 7488
},
{
"epoch": 1.37,
"grad_norm": 1.5333857134961049,
"learning_rate": 6.589637925654215e-07,
"loss": 0.251,
"step": 7520
},
{
"epoch": 1.38,
"grad_norm": 1.5263236812468681,
"learning_rate": 6.557396153677356e-07,
"loss": 0.2543,
"step": 7552
},
{
"epoch": 1.38,
"grad_norm": 1.0995345813700304,
"learning_rate": 6.525082504511612e-07,
"loss": 0.2493,
"step": 7584
},
{
"epoch": 1.39,
"grad_norm": 1.4740647676869196,
"learning_rate": 6.492698469501532e-07,
"loss": 0.2414,
"step": 7616
},
{
"epoch": 1.4,
"grad_norm": 1.4036985523166017,
"learning_rate": 6.460245543240123e-07,
"loss": 0.2373,
"step": 7648
},
{
"epoch": 1.4,
"grad_norm": 1.4125143998119496,
"learning_rate": 6.427725223499871e-07,
"loss": 0.2427,
"step": 7680
},
{
"epoch": 1.41,
"grad_norm": 1.195620693310162,
"learning_rate": 6.39513901116362e-07,
"loss": 0.2522,
"step": 7712
},
{
"epoch": 1.41,
"grad_norm": 1.2679060414661505,
"learning_rate": 6.362488410155298e-07,
"loss": 0.2448,
"step": 7744
},
{
"epoch": 1.42,
"grad_norm": 1.4675997129467317,
"learning_rate": 6.329774927370504e-07,
"loss": 0.2459,
"step": 7776
},
{
"epoch": 1.42,
"grad_norm": 1.3511806562135458,
"learning_rate": 6.297000072606972e-07,
"loss": 0.257,
"step": 7808
},
{
"epoch": 1.43,
"grad_norm": 1.272848715104825,
"learning_rate": 6.264165358494884e-07,
"loss": 0.246,
"step": 7840
},
{
"epoch": 1.44,
"grad_norm": 1.4081103274783833,
"learning_rate": 6.231272300427059e-07,
"loss": 0.2485,
"step": 7872
},
{
"epoch": 1.44,
"grad_norm": 1.2073750932804077,
"learning_rate": 6.198322416489016e-07,
"loss": 0.2354,
"step": 7904
},
{
"epoch": 1.45,
"grad_norm": 1.3509938246652688,
"learning_rate": 6.165317227388911e-07,
"loss": 0.2423,
"step": 7936
},
{
"epoch": 1.45,
"grad_norm": 1.8975854309012892,
"learning_rate": 6.132258256387348e-07,
"loss": 0.2422,
"step": 7968
},
{
"epoch": 1.46,
"grad_norm": 1.616263805539082,
"learning_rate": 6.099147029227088e-07,
"loss": 0.2452,
"step": 8000
},
{
"epoch": 1.47,
"grad_norm": 1.4581239370531887,
"learning_rate": 6.065985074062624e-07,
"loss": 0.2564,
"step": 8032
},
{
"epoch": 1.47,
"grad_norm": 1.4107792940118344,
"learning_rate": 6.032773921389654e-07,
"loss": 0.2613,
"step": 8064
},
{
"epoch": 1.48,
"grad_norm": 1.4254738015511745,
"learning_rate": 5.999515103974447e-07,
"loss": 0.2441,
"step": 8096
},
{
"epoch": 1.48,
"grad_norm": 1.2694180719111068,
"learning_rate": 5.966210156783108e-07,
"loss": 0.2559,
"step": 8128
},
{
"epoch": 1.49,
"grad_norm": 1.3375582532247023,
"learning_rate": 5.932860616910721e-07,
"loss": 0.2488,
"step": 8160
},
{
"epoch": 1.49,
"grad_norm": 1.2586037041521065,
"learning_rate": 5.899468023510428e-07,
"loss": 0.2427,
"step": 8192
},
{
"epoch": 1.5,
"grad_norm": 1.2896167720703278,
"learning_rate": 5.866033917722379e-07,
"loss": 0.2567,
"step": 8224
},
{
"epoch": 1.51,
"grad_norm": 1.101067530433089,
"learning_rate": 5.832559842602608e-07,
"loss": 0.2342,
"step": 8256
},
{
"epoch": 1.51,
"grad_norm": 1.1591181439961709,
"learning_rate": 5.799047343051827e-07,
"loss": 0.2411,
"step": 8288
},
{
"epoch": 1.52,
"grad_norm": 1.2887740481533372,
"learning_rate": 5.765497965744111e-07,
"loss": 0.2433,
"step": 8320
},
{
"epoch": 1.52,
"grad_norm": 1.1116096883871758,
"learning_rate": 5.731913259055527e-07,
"loss": 0.2435,
"step": 8352
},
{
"epoch": 1.53,
"grad_norm": 1.3978304499588927,
"learning_rate": 5.698294772992669e-07,
"loss": 0.2485,
"step": 8384
},
{
"epoch": 1.54,
"grad_norm": 1.1353060264434358,
"learning_rate": 5.664644059121121e-07,
"loss": 0.2508,
"step": 8416
},
{
"epoch": 1.54,
"grad_norm": 1.337326494552789,
"learning_rate": 5.630962670493848e-07,
"loss": 0.2446,
"step": 8448
},
{
"epoch": 1.55,
"grad_norm": 1.5497857023189168,
"learning_rate": 5.597252161579525e-07,
"loss": 0.2498,
"step": 8480
},
{
"epoch": 1.55,
"grad_norm": 1.6735926212275043,
"learning_rate": 5.563514088190788e-07,
"loss": 0.2382,
"step": 8512
},
{
"epoch": 1.56,
"grad_norm": 1.5802941502096313,
"learning_rate": 5.529750007412435e-07,
"loss": 0.24,
"step": 8544
},
{
"epoch": 1.56,
"grad_norm": 1.434775976642666,
"learning_rate": 5.495961477529559e-07,
"loss": 0.2508,
"step": 8576
},
{
"epoch": 1.57,
"grad_norm": 1.3871703235640929,
"learning_rate": 5.462150057955633e-07,
"loss": 0.24,
"step": 8608
},
{
"epoch": 1.58,
"grad_norm": 1.3448291807441222,
"learning_rate": 5.428317309160538e-07,
"loss": 0.2484,
"step": 8640
},
{
"epoch": 1.58,
"grad_norm": 1.3157038644549268,
"learning_rate": 5.394464792598545e-07,
"loss": 0.2409,
"step": 8672
},
{
"epoch": 1.59,
"grad_norm": 1.2008640772377304,
"learning_rate": 5.360594070636248e-07,
"loss": 0.2454,
"step": 8704
},
{
"epoch": 1.59,
"grad_norm": 1.3073489027432128,
"learning_rate": 5.326706706480467e-07,
"loss": 0.2446,
"step": 8736
},
{
"epoch": 1.6,
"grad_norm": 1.5955373808274311,
"learning_rate": 5.292804264106083e-07,
"loss": 0.2433,
"step": 8768
},
{
"epoch": 1.61,
"grad_norm": 1.2901907823242207,
"learning_rate": 5.25888830818388e-07,
"loss": 0.2426,
"step": 8800
},
{
"epoch": 1.61,
"grad_norm": 1.4454540634426136,
"learning_rate": 5.224960404008319e-07,
"loss": 0.2487,
"step": 8832
},
{
"epoch": 1.62,
"grad_norm": 1.3252857725269103,
"learning_rate": 5.19102211742529e-07,
"loss": 0.242,
"step": 8864
},
{
"epoch": 1.62,
"grad_norm": 1.3963560963709287,
"learning_rate": 5.157075014759866e-07,
"loss": 0.2333,
"step": 8896
},
{
"epoch": 1.63,
"grad_norm": 1.7474842902023573,
"learning_rate": 5.123120662743993e-07,
"loss": 0.2397,
"step": 8928
},
{
"epoch": 1.63,
"grad_norm": 1.463552241271405,
"learning_rate": 5.089160628444193e-07,
"loss": 0.2475,
"step": 8960
},
{
"epoch": 1.64,
"grad_norm": 1.4711904387458636,
"learning_rate": 5.055196479189237e-07,
"loss": 0.2468,
"step": 8992
},
{
"epoch": 1.65,
"grad_norm": 1.1486741329450094,
"learning_rate": 5.021229782497811e-07,
"loss": 0.2474,
"step": 9024
},
{
"epoch": 1.65,
"grad_norm": 1.465014768579321,
"learning_rate": 4.987262106006171e-07,
"loss": 0.2437,
"step": 9056
},
{
"epoch": 1.66,
"grad_norm": 1.6572146980799467,
"learning_rate": 4.953295017395788e-07,
"loss": 0.2358,
"step": 9088
},
{
"epoch": 1.66,
"grad_norm": 1.1497400773353195,
"learning_rate": 4.919330084321009e-07,
"loss": 0.2459,
"step": 9120
},
{
"epoch": 1.67,
"grad_norm": 1.4973289494569486,
"learning_rate": 4.885368874336694e-07,
"loss": 0.2497,
"step": 9152
},
{
"epoch": 1.68,
"grad_norm": 1.375218554438502,
"learning_rate": 4.851412954825874e-07,
"loss": 0.2508,
"step": 9184
},
{
"epoch": 1.68,
"grad_norm": 1.3780517501267755,
"learning_rate": 4.817463892927411e-07,
"loss": 0.2567,
"step": 9216
},
{
"epoch": 1.69,
"grad_norm": 1.4504412670024611,
"learning_rate": 4.783523255463679e-07,
"loss": 0.2409,
"step": 9248
},
{
"epoch": 1.69,
"grad_norm": 1.2870854505712077,
"learning_rate": 4.7495926088682436e-07,
"loss": 0.2423,
"step": 9280
},
{
"epoch": 1.7,
"grad_norm": 1.3985153319949002,
"learning_rate": 4.71567351911357e-07,
"loss": 0.251,
"step": 9312
},
{
"epoch": 1.7,
"grad_norm": 1.0349716694291753,
"learning_rate": 4.681767551638751e-07,
"loss": 0.2404,
"step": 9344
},
{
"epoch": 1.71,
"grad_norm": 1.428114124678113,
"learning_rate": 4.647876271277257e-07,
"loss": 0.2391,
"step": 9376
},
{
"epoch": 1.72,
"grad_norm": 1.3200951440851763,
"learning_rate": 4.6140012421847136e-07,
"loss": 0.2535,
"step": 9408
},
{
"epoch": 1.72,
"grad_norm": 1.3868492152964986,
"learning_rate": 4.5801440277667235e-07,
"loss": 0.2483,
"step": 9440
},
{
"epoch": 1.73,
"grad_norm": 1.2729900579595428,
"learning_rate": 4.5463061906066965e-07,
"loss": 0.2512,
"step": 9472
},
{
"epoch": 1.73,
"grad_norm": 1.315198823460812,
"learning_rate": 4.5124892923937416e-07,
"loss": 0.2417,
"step": 9504
},
{
"epoch": 1.74,
"grad_norm": 1.7120982404617975,
"learning_rate": 4.478694893850593e-07,
"loss": 0.2362,
"step": 9536
},
{
"epoch": 1.75,
"grad_norm": 1.412290074923546,
"learning_rate": 4.4449245546615753e-07,
"loss": 0.2483,
"step": 9568
},
{
"epoch": 1.75,
"grad_norm": 1.2470698026793279,
"learning_rate": 4.411179833400619e-07,
"loss": 0.2432,
"step": 9600
},
{
"epoch": 1.76,
"grad_norm": 1.268429893461802,
"learning_rate": 4.377462287459337e-07,
"loss": 0.2428,
"step": 9632
},
{
"epoch": 1.76,
"grad_norm": 1.2455579800684395,
"learning_rate": 4.343773472975139e-07,
"loss": 0.2508,
"step": 9664
},
{
"epoch": 1.77,
"grad_norm": 1.228328713009981,
"learning_rate": 4.3101149447594163e-07,
"loss": 0.2551,
"step": 9696
},
{
"epoch": 1.77,
"grad_norm": 1.2042503922389745,
"learning_rate": 4.276488256225787e-07,
"loss": 0.2541,
"step": 9728
},
{
"epoch": 1.78,
"grad_norm": 1.2346699498129847,
"learning_rate": 4.242894959318395e-07,
"loss": 0.2461,
"step": 9760
},
{
"epoch": 1.79,
"grad_norm": 1.3489561185627454,
"learning_rate": 4.209336604440294e-07,
"loss": 0.2477,
"step": 9792
},
{
"epoch": 1.79,
"grad_norm": 1.6713827739937024,
"learning_rate": 4.175814740381879e-07,
"loss": 0.2458,
"step": 9824
},
{
"epoch": 1.8,
"grad_norm": 1.3796544902936816,
"learning_rate": 4.1423309142494236e-07,
"loss": 0.2517,
"step": 9856
},
{
"epoch": 1.8,
"grad_norm": 1.6289046211602507,
"learning_rate": 4.1088866713936633e-07,
"loss": 0.2411,
"step": 9888
},
{
"epoch": 1.81,
"grad_norm": 1.4906278025262827,
"learning_rate": 4.0754835553384864e-07,
"loss": 0.2425,
"step": 9920
},
{
"epoch": 1.82,
"grad_norm": 1.1884900865492198,
"learning_rate": 4.0421231077096844e-07,
"loss": 0.2587,
"step": 9952
},
{
"epoch": 1.82,
"grad_norm": 1.329339394262938,
"learning_rate": 4.008806868163814e-07,
"loss": 0.2391,
"step": 9984
},
{
"epoch": 1.83,
"grad_norm": 1.2186411235918218,
"learning_rate": 3.9755363743171265e-07,
"loss": 0.253,
"step": 10016
},
{
"epoch": 1.83,
"grad_norm": 2.0559660633265047,
"learning_rate": 3.9423131616746187e-07,
"loss": 0.2566,
"step": 10048
},
{
"epoch": 1.84,
"grad_norm": 1.255138693515569,
"learning_rate": 3.9091387635591533e-07,
"loss": 0.2404,
"step": 10080
},
{
"epoch": 1.84,
"grad_norm": 1.5524707973995229,
"learning_rate": 3.8760147110406995e-07,
"loss": 0.2519,
"step": 10112
},
{
"epoch": 1.85,
"grad_norm": 1.2244273291645265,
"learning_rate": 3.8429425328656674e-07,
"loss": 0.2388,
"step": 10144
},
{
"epoch": 1.86,
"grad_norm": 1.438386617275495,
"learning_rate": 3.809923755386355e-07,
"loss": 0.25,
"step": 10176
},
{
"epoch": 1.86,
"grad_norm": 1.5262712808405177,
"learning_rate": 3.7769599024905004e-07,
"loss": 0.2299,
"step": 10208
},
{
"epoch": 1.87,
"grad_norm": 1.0721875435698476,
"learning_rate": 3.744052495530959e-07,
"loss": 0.2375,
"step": 10240
},
{
"epoch": 1.87,
"grad_norm": 1.1626282358875817,
"learning_rate": 3.7112030532554806e-07,
"loss": 0.2497,
"step": 10272
},
{
"epoch": 1.88,
"grad_norm": 1.3954184625846942,
"learning_rate": 3.6784130917366195e-07,
"loss": 0.242,
"step": 10304
},
{
"epoch": 1.89,
"grad_norm": 1.3259939974154864,
"learning_rate": 3.64568412430177e-07,
"loss": 0.2516,
"step": 10336
},
{
"epoch": 1.89,
"grad_norm": 1.73086784272557,
"learning_rate": 3.613017661463313e-07,
"loss": 0.2483,
"step": 10368
},
{
"epoch": 1.9,
"grad_norm": 1.188961288223144,
"learning_rate": 3.5804152108489065e-07,
"loss": 0.2391,
"step": 10400
},
{
"epoch": 1.9,
"grad_norm": 1.1246493355621157,
"learning_rate": 3.547878277131913e-07,
"loss": 0.2383,
"step": 10432
},
{
"epoch": 1.91,
"grad_norm": 1.059964023169643,
"learning_rate": 3.515408361961941e-07,
"loss": 0.2404,
"step": 10464
},
{
"epoch": 1.91,
"grad_norm": 1.70561149747635,
"learning_rate": 3.483006963895555e-07,
"loss": 0.2452,
"step": 10496
},
{
"epoch": 1.92,
"grad_norm": 1.2969563700400484,
"learning_rate": 3.4506755783271045e-07,
"loss": 0.2324,
"step": 10528
},
{
"epoch": 1.93,
"grad_norm": 1.2846755096320104,
"learning_rate": 3.418415697419712e-07,
"loss": 0.244,
"step": 10560
},
{
"epoch": 1.93,
"grad_norm": 1.7188757741977305,
"learning_rate": 3.386228810036408e-07,
"loss": 0.2369,
"step": 10592
},
{
"epoch": 1.94,
"grad_norm": 1.362077199933714,
"learning_rate": 3.3541164016714065e-07,
"loss": 0.2458,
"step": 10624
},
{
"epoch": 1.94,
"grad_norm": 1.354641295840275,
"learning_rate": 3.3220799543815634e-07,
"loss": 0.2397,
"step": 10656
},
{
"epoch": 1.95,
"grad_norm": 1.3277536016021265,
"learning_rate": 3.2901209467179637e-07,
"loss": 0.2458,
"step": 10688
},
{
"epoch": 1.96,
"grad_norm": 1.3886060047726545,
"learning_rate": 3.2582408536576877e-07,
"loss": 0.2344,
"step": 10720
},
{
"epoch": 1.96,
"grad_norm": 1.3601591408784948,
"learning_rate": 3.2264411465357335e-07,
"loss": 0.2343,
"step": 10752
},
{
"epoch": 1.97,
"grad_norm": 1.1017573737076742,
"learning_rate": 3.194723292977123e-07,
"loss": 0.2442,
"step": 10784
},
{
"epoch": 1.97,
"grad_norm": 1.4213916209193176,
"learning_rate": 3.1630887568291465e-07,
"loss": 0.2364,
"step": 10816
},
{
"epoch": 1.98,
"grad_norm": 1.3035008616357522,
"learning_rate": 3.131538998093828e-07,
"loss": 0.2458,
"step": 10848
},
{
"epoch": 1.99,
"grad_norm": 1.218412032577656,
"learning_rate": 3.1000754728605256e-07,
"loss": 0.2413,
"step": 10880
},
{
"epoch": 1.99,
"grad_norm": 1.2325820939927188,
"learning_rate": 3.068699633238738e-07,
"loss": 0.2511,
"step": 10912
},
{
"epoch": 2.0,
"grad_norm": 1.5401850018534633,
"learning_rate": 3.037412927291083e-07,
"loss": 0.2461,
"step": 10944
},
{
"epoch": 2.0,
"grad_norm": 1.3586152389713673,
"learning_rate": 3.006216798966468e-07,
"loss": 0.2358,
"step": 10976
},
{
"epoch": 2.01,
"grad_norm": 1.4011774529220307,
"learning_rate": 2.975112688033442e-07,
"loss": 0.2169,
"step": 11008
},
{
"epoch": 2.01,
"grad_norm": 1.2241897731555507,
"learning_rate": 2.944102030013763e-07,
"loss": 0.2108,
"step": 11040
},
{
"epoch": 2.02,
"grad_norm": 1.4027474158874935,
"learning_rate": 2.9131862561161293e-07,
"loss": 0.2293,
"step": 11072
},
{
"epoch": 2.03,
"grad_norm": 1.4670817911133367,
"learning_rate": 2.882366793170133e-07,
"loss": 0.2244,
"step": 11104
},
{
"epoch": 2.03,
"grad_norm": 1.3153219602513937,
"learning_rate": 2.8516450635604086e-07,
"loss": 0.22,
"step": 11136
},
{
"epoch": 2.04,
"grad_norm": 1.2976761125367386,
"learning_rate": 2.821022485160991e-07,
"loss": 0.2097,
"step": 11168
},
{
"epoch": 2.04,
"grad_norm": 1.3625162799781199,
"learning_rate": 2.7905004712698645e-07,
"loss": 0.2199,
"step": 11200
},
{
"epoch": 2.05,
"grad_norm": 1.232550115651448,
"learning_rate": 2.7600804305437474e-07,
"loss": 0.2239,
"step": 11232
},
{
"epoch": 2.06,
"grad_norm": 1.4796643786771864,
"learning_rate": 2.7297637669330847e-07,
"loss": 0.222,
"step": 11264
},
{
"epoch": 2.06,
"grad_norm": 1.2946736681294901,
"learning_rate": 2.699551879617235e-07,
"loss": 0.2246,
"step": 11296
},
{
"epoch": 2.07,
"grad_norm": 1.2068473533560813,
"learning_rate": 2.669446162939909e-07,
"loss": 0.2229,
"step": 11328
},
{
"epoch": 2.07,
"grad_norm": 1.4805315280685087,
"learning_rate": 2.639448006344812e-07,
"loss": 0.2131,
"step": 11360
},
{
"epoch": 2.08,
"grad_norm": 1.1871137311489093,
"learning_rate": 2.6095587943115227e-07,
"loss": 0.2235,
"step": 11392
},
{
"epoch": 2.08,
"grad_norm": 1.088896036136479,
"learning_rate": 2.57977990629159e-07,
"loss": 0.2073,
"step": 11424
},
{
"epoch": 2.09,
"grad_norm": 1.2960712344910867,
"learning_rate": 2.550112716644875e-07,
"loss": 0.2197,
"step": 11456
},
{
"epoch": 2.1,
"grad_norm": 1.2204092718307542,
"learning_rate": 2.520558594576115e-07,
"loss": 0.2282,
"step": 11488
},
{
"epoch": 2.1,
"grad_norm": 1.2693877027082432,
"learning_rate": 2.4911189040717357e-07,
"loss": 0.225,
"step": 11520
},
{
"epoch": 2.11,
"grad_norm": 1.2846435895566266,
"learning_rate": 2.4617950038369035e-07,
"loss": 0.2234,
"step": 11552
},
{
"epoch": 2.11,
"grad_norm": 1.2777271226205853,
"learning_rate": 2.432588247232811e-07,
"loss": 0.2211,
"step": 11584
},
{
"epoch": 2.12,
"grad_norm": 1.528372242529096,
"learning_rate": 2.4034999822142166e-07,
"loss": 0.22,
"step": 11616
},
{
"epoch": 2.13,
"grad_norm": 1.3339829276713877,
"learning_rate": 2.3745315512672397e-07,
"loss": 0.2195,
"step": 11648
},
{
"epoch": 2.13,
"grad_norm": 1.3101509979598953,
"learning_rate": 2.345684291347403e-07,
"loss": 0.2159,
"step": 11680
},
{
"epoch": 2.14,
"grad_norm": 1.3071577674707573,
"learning_rate": 2.3169595338179192e-07,
"loss": 0.2295,
"step": 11712
},
{
"epoch": 2.14,
"grad_norm": 1.3186809416500118,
"learning_rate": 2.2883586043882548e-07,
"loss": 0.2133,
"step": 11744
},
{
"epoch": 2.15,
"grad_norm": 1.5867314938269408,
"learning_rate": 2.259882823052941e-07,
"loss": 0.2316,
"step": 11776
},
{
"epoch": 2.15,
"grad_norm": 1.1557431014242034,
"learning_rate": 2.2315335040306554e-07,
"loss": 0.2189,
"step": 11808
},
{
"epoch": 2.16,
"grad_norm": 1.2736258434323458,
"learning_rate": 2.203311955703568e-07,
"loss": 0.2199,
"step": 11840
},
{
"epoch": 2.17,
"grad_norm": 1.302216300423092,
"learning_rate": 2.1752194805569552e-07,
"loss": 0.229,
"step": 11872
},
{
"epoch": 2.17,
"grad_norm": 1.3128201457087052,
"learning_rate": 2.147257375119087e-07,
"loss": 0.236,
"step": 11904
},
{
"epoch": 2.18,
"grad_norm": 1.3748881212985564,
"learning_rate": 2.119426929901391e-07,
"loss": 0.2206,
"step": 11936
},
{
"epoch": 2.18,
"grad_norm": 2.0921207372413435,
"learning_rate": 2.0917294293388966e-07,
"loss": 0.2117,
"step": 11968
},
{
"epoch": 2.19,
"grad_norm": 1.2949118526726222,
"learning_rate": 2.0641661517309417e-07,
"loss": 0.2135,
"step": 12000
},
{
"epoch": 2.2,
"grad_norm": 1.3415735498836403,
"learning_rate": 2.03673836918219e-07,
"loss": 0.2259,
"step": 12032
},
{
"epoch": 2.2,
"grad_norm": 1.55429350265861,
"learning_rate": 2.00944734754392e-07,
"loss": 0.2279,
"step": 12064
},
{
"epoch": 2.21,
"grad_norm": 1.2935837103909433,
"learning_rate": 1.9822943463555948e-07,
"loss": 0.2308,
"step": 12096
},
{
"epoch": 2.21,
"grad_norm": 1.642310611966749,
"learning_rate": 1.9552806187867372e-07,
"loss": 0.2144,
"step": 12128
},
{
"epoch": 2.22,
"grad_norm": 1.2465970815408216,
"learning_rate": 1.9284074115790932e-07,
"loss": 0.2198,
"step": 12160
},
{
"epoch": 2.22,
"grad_norm": 1.6894912323452993,
"learning_rate": 1.9016759649890895e-07,
"loss": 0.2147,
"step": 12192
},
{
"epoch": 2.23,
"grad_norm": 1.1571001402696899,
"learning_rate": 1.8750875127305977e-07,
"loss": 0.2149,
"step": 12224
},
{
"epoch": 2.24,
"grad_norm": 1.311070700271489,
"learning_rate": 1.8486432819179892e-07,
"loss": 0.2135,
"step": 12256
},
{
"epoch": 2.24,
"grad_norm": 1.358941411906666,
"learning_rate": 1.822344493009509e-07,
"loss": 0.2291,
"step": 12288
},
{
"epoch": 2.25,
"grad_norm": 1.2572126595123219,
"learning_rate": 1.7961923597509387e-07,
"loss": 0.2253,
"step": 12320
},
{
"epoch": 2.25,
"grad_norm": 1.4204877616087515,
"learning_rate": 1.7701880891195942e-07,
"loss": 0.212,
"step": 12352
},
{
"epoch": 2.26,
"grad_norm": 1.3082239260119037,
"learning_rate": 1.744332881268607e-07,
"loss": 0.2258,
"step": 12384
},
{
"epoch": 2.27,
"grad_norm": 1.3818606595114133,
"learning_rate": 1.7186279294715382e-07,
"loss": 0.2186,
"step": 12416
},
{
"epoch": 2.27,
"grad_norm": 1.344313917268325,
"learning_rate": 1.6930744200673104e-07,
"loss": 0.2309,
"step": 12448
},
{
"epoch": 2.28,
"grad_norm": 1.2849788562306779,
"learning_rate": 1.667673532405457e-07,
"loss": 0.2119,
"step": 12480
},
{
"epoch": 2.28,
"grad_norm": 1.048703241818117,
"learning_rate": 1.6424264387916842e-07,
"loss": 0.2175,
"step": 12512
},
{
"epoch": 2.29,
"grad_norm": 1.643116827546219,
"learning_rate": 1.6173343044337734e-07,
"loss": 0.2264,
"step": 12544
},
{
"epoch": 2.29,
"grad_norm": 1.2754388516106783,
"learning_rate": 1.5923982873878012e-07,
"loss": 0.2296,
"step": 12576
},
{
"epoch": 2.3,
"grad_norm": 1.2217192268295975,
"learning_rate": 1.567619538504696e-07,
"loss": 0.2281,
"step": 12608
},
{
"epoch": 2.31,
"grad_norm": 1.4770776525365248,
"learning_rate": 1.542999201377119e-07,
"loss": 0.2277,
"step": 12640
},
{
"epoch": 2.31,
"grad_norm": 1.642064071179986,
"learning_rate": 1.5185384122866897e-07,
"loss": 0.218,
"step": 12672
},
{
"epoch": 2.32,
"grad_norm": 1.237204280837322,
"learning_rate": 1.4942383001515386e-07,
"loss": 0.2079,
"step": 12704
},
{
"epoch": 2.32,
"grad_norm": 1.1715182674707414,
"learning_rate": 1.4700999864742136e-07,
"loss": 0.2124,
"step": 12736
},
{
"epoch": 2.33,
"grad_norm": 1.770494939875621,
"learning_rate": 1.4461245852899128e-07,
"loss": 0.2046,
"step": 12768
},
{
"epoch": 2.34,
"grad_norm": 1.2181121499687988,
"learning_rate": 1.4223132031150676e-07,
"loss": 0.2179,
"step": 12800
},
{
"epoch": 2.34,
"grad_norm": 1.2327845797316905,
"learning_rate": 1.3986669388962819e-07,
"loss": 0.2202,
"step": 12832
},
{
"epoch": 2.35,
"grad_norm": 1.5192364095540325,
"learning_rate": 1.375186883959613e-07,
"loss": 0.2187,
"step": 12864
},
{
"epoch": 2.35,
"grad_norm": 1.1447367117093976,
"learning_rate": 1.3518741219601988e-07,
"loss": 0.2267,
"step": 12896
},
{
"epoch": 2.36,
"grad_norm": 1.374708110513197,
"learning_rate": 1.3287297288322458e-07,
"loss": 0.2192,
"step": 12928
},
{
"epoch": 2.36,
"grad_norm": 1.1547706075025737,
"learning_rate": 1.305754772739377e-07,
"loss": 0.2101,
"step": 12960
},
{
"epoch": 2.37,
"grad_norm": 1.350330342693555,
"learning_rate": 1.2829503140253296e-07,
"loss": 0.2169,
"step": 12992
},
{
"epoch": 2.38,
"grad_norm": 1.416668233872815,
"learning_rate": 1.2603174051650207e-07,
"loss": 0.2181,
"step": 13024
},
{
"epoch": 2.38,
"grad_norm": 1.2796831245124296,
"learning_rate": 1.2378570907159697e-07,
"loss": 0.2264,
"step": 13056
},
{
"epoch": 2.39,
"grad_norm": 1.5438983104227264,
"learning_rate": 1.215570407270095e-07,
"loss": 0.2233,
"step": 13088
},
{
"epoch": 2.39,
"grad_norm": 1.7803269270190754,
"learning_rate": 1.1934583834058658e-07,
"loss": 0.2241,
"step": 13120
},
{
"epoch": 2.4,
"grad_norm": 1.5471141725466007,
"learning_rate": 1.1715220396408426e-07,
"loss": 0.2124,
"step": 13152
},
{
"epoch": 2.41,
"grad_norm": 1.1206628903783873,
"learning_rate": 1.1497623883845614e-07,
"loss": 0.2127,
"step": 13184
},
{
"epoch": 2.41,
"grad_norm": 2.5095477739496967,
"learning_rate": 1.1281804338918239e-07,
"loss": 0.2282,
"step": 13216
},
{
"epoch": 2.42,
"grad_norm": 1.3128197458740387,
"learning_rate": 1.1067771722163404e-07,
"loss": 0.2185,
"step": 13248
},
{
"epoch": 2.42,
"grad_norm": 1.2471858850247612,
"learning_rate": 1.0855535911647673e-07,
"loss": 0.2185,
"step": 13280
},
{
"epoch": 2.43,
"grad_norm": 1.655671236852773,
"learning_rate": 1.064510670251108e-07,
"loss": 0.2246,
"step": 13312
},
{
"epoch": 2.43,
"grad_norm": 1.2165519973679901,
"learning_rate": 1.0436493806515128e-07,
"loss": 0.223,
"step": 13344
},
{
"epoch": 2.44,
"grad_norm": 1.1834223459074134,
"learning_rate": 1.0229706851594561e-07,
"loss": 0.2188,
"step": 13376
},
{
"epoch": 2.45,
"grad_norm": 1.5035939672617133,
"learning_rate": 1.0024755381412997e-07,
"loss": 0.2204,
"step": 13408
},
{
"epoch": 2.45,
"grad_norm": 1.362765323113097,
"learning_rate": 9.821648854922481e-08,
"loss": 0.2165,
"step": 13440
},
{
"epoch": 2.46,
"grad_norm": 1.4329165569522946,
"learning_rate": 9.620396645926937e-08,
"loss": 0.2104,
"step": 13472
},
{
"epoch": 2.46,
"grad_norm": 1.2794493917175387,
"learning_rate": 9.421008042649514e-08,
"loss": 0.2225,
"step": 13504
},
{
"epoch": 2.47,
"grad_norm": 1.365469705607083,
"learning_rate": 9.223492247303999e-08,
"loss": 0.2166,
"step": 13536
},
{
"epoch": 2.48,
"grad_norm": 1.3658393916273586,
"learning_rate": 9.027858375670005e-08,
"loss": 0.2244,
"step": 13568
},
{
"epoch": 2.48,
"grad_norm": 1.6134388977737084,
"learning_rate": 8.834115456672314e-08,
"loss": 0.2238,
"step": 13600
},
{
"epoch": 2.49,
"grad_norm": 1.2747001650429752,
"learning_rate": 8.642272431964171e-08,
"loss": 0.2117,
"step": 13632
},
{
"epoch": 2.49,
"grad_norm": 1.4495915191793898,
"learning_rate": 8.452338155514644e-08,
"loss": 0.2153,
"step": 13664
},
{
"epoch": 2.5,
"grad_norm": 1.704239796998396,
"learning_rate": 8.264321393199902e-08,
"loss": 0.214,
"step": 13696
},
{
"epoch": 2.5,
"grad_norm": 1.1540455435637955,
"learning_rate": 8.078230822398735e-08,
"loss": 0.2189,
"step": 13728
},
{
"epoch": 2.51,
"grad_norm": 1.4264481560916344,
"learning_rate": 7.894075031592029e-08,
"loss": 0.2142,
"step": 13760
},
{
"epoch": 2.52,
"grad_norm": 1.2771622178109003,
"learning_rate": 7.711862519966416e-08,
"loss": 0.2238,
"step": 13792
},
{
"epoch": 2.52,
"grad_norm": 1.4176109439678235,
"learning_rate": 7.531601697022e-08,
"loss": 0.2141,
"step": 13824
},
{
"epoch": 2.53,
"grad_norm": 1.211872373307708,
"learning_rate": 7.353300882184238e-08,
"loss": 0.2094,
"step": 13856
},
{
"epoch": 2.53,
"grad_norm": 1.6208090587491066,
"learning_rate": 7.176968304420006e-08,
"loss": 0.213,
"step": 13888
},
{
"epoch": 2.54,
"grad_norm": 1.4294182508163742,
"learning_rate": 7.002612101857769e-08,
"loss": 0.2206,
"step": 13920
},
{
"epoch": 2.55,
"grad_norm": 1.5238939549044208,
"learning_rate": 6.830240321412067e-08,
"loss": 0.2187,
"step": 13952
},
{
"epoch": 2.55,
"grad_norm": 1.4817796496161193,
"learning_rate": 6.65986091841203e-08,
"loss": 0.223,
"step": 13984
},
{
"epoch": 2.56,
"grad_norm": 1.4428543645745906,
"learning_rate": 6.491481756234296e-08,
"loss": 0.2178,
"step": 14016
},
{
"epoch": 2.56,
"grad_norm": 1.3986608111205634,
"learning_rate": 6.32511060594007e-08,
"loss": 0.2244,
"step": 14048
},
{
"epoch": 2.57,
"grad_norm": 1.1779478103336172,
"learning_rate": 6.160755145916518e-08,
"loss": 0.2326,
"step": 14080
},
{
"epoch": 2.57,
"grad_norm": 1.4661784823721757,
"learning_rate": 5.99842296152231e-08,
"loss": 0.2257,
"step": 14112
},
{
"epoch": 2.58,
"grad_norm": 1.3997989298044267,
"learning_rate": 5.8381215447376006e-08,
"loss": 0.218,
"step": 14144
},
{
"epoch": 2.59,
"grad_norm": 1.1966150734513934,
"learning_rate": 5.679858293818257e-08,
"loss": 0.2139,
"step": 14176
},
{
"epoch": 2.59,
"grad_norm": 1.2975301799355876,
"learning_rate": 5.523640512954375e-08,
"loss": 0.2298,
"step": 14208
},
{
"epoch": 2.6,
"grad_norm": 1.1520772784392557,
"learning_rate": 5.369475411933222e-08,
"loss": 0.2086,
"step": 14240
},
{
"epoch": 2.6,
"grad_norm": 1.4920063917973507,
"learning_rate": 5.2173701058064454e-08,
"loss": 0.215,
"step": 14272
},
{
"epoch": 2.61,
"grad_norm": 1.4418059613129284,
"learning_rate": 5.067331614561726e-08,
"loss": 0.2108,
"step": 14304
},
{
"epoch": 2.62,
"grad_norm": 1.122573143556837,
"learning_rate": 4.9193668627988073e-08,
"loss": 0.2156,
"step": 14336
},
{
"epoch": 2.62,
"grad_norm": 1.169213395405146,
"learning_rate": 4.7734826794098624e-08,
"loss": 0.2155,
"step": 14368
},
{
"epoch": 2.63,
"grad_norm": 1.4739891927117594,
"learning_rate": 4.629685797264338e-08,
"loss": 0.2197,
"step": 14400
},
{
"epoch": 2.63,
"grad_norm": 1.1061529994285666,
"learning_rate": 4.487982852898248e-08,
"loss": 0.2161,
"step": 14432
},
{
"epoch": 2.64,
"grad_norm": 1.2665429377042383,
"learning_rate": 4.348380386207884e-08,
"loss": 0.2293,
"step": 14464
},
{
"epoch": 2.64,
"grad_norm": 1.4251977400136069,
"learning_rate": 4.210884840147927e-08,
"loss": 0.219,
"step": 14496
},
{
"epoch": 2.65,
"grad_norm": 1.3125295629014255,
"learning_rate": 4.075502560434158e-08,
"loss": 0.2215,
"step": 14528
},
{
"epoch": 2.66,
"grad_norm": 1.6114888474605462,
"learning_rate": 3.942239795250546e-08,
"loss": 0.2141,
"step": 14560
},
{
"epoch": 2.66,
"grad_norm": 1.3415946209746852,
"learning_rate": 3.8111026949609045e-08,
"loss": 0.2145,
"step": 14592
},
{
"epoch": 2.67,
"grad_norm": 1.5169187816349006,
"learning_rate": 3.682097311825033e-08,
"loss": 0.2241,
"step": 14624
},
{
"epoch": 2.67,
"grad_norm": 1.3176284198519965,
"learning_rate": 3.5552295997193834e-08,
"loss": 0.2238,
"step": 14656
},
{
"epoch": 2.68,
"grad_norm": 1.5611137018278904,
"learning_rate": 3.430505413862284e-08,
"loss": 0.2029,
"step": 14688
},
{
"epoch": 2.69,
"grad_norm": 1.4814939062116772,
"learning_rate": 3.307930510543705e-08,
"loss": 0.2073,
"step": 14720
},
{
"epoch": 2.69,
"grad_norm": 1.4585161010377092,
"learning_rate": 3.187510546859623e-08,
"loss": 0.2158,
"step": 14752
},
{
"epoch": 2.7,
"grad_norm": 1.3111001801067392,
"learning_rate": 3.069251080450863e-08,
"loss": 0.2114,
"step": 14784
},
{
"epoch": 2.7,
"grad_norm": 1.3748879882896838,
"learning_rate": 2.953157569246656e-08,
"loss": 0.2191,
"step": 14816
},
{
"epoch": 2.71,
"grad_norm": 1.2458399054987248,
"learning_rate": 2.8392353712127615e-08,
"loss": 0.2226,
"step": 14848
},
{
"epoch": 2.71,
"grad_norm": 1.183457902413645,
"learning_rate": 2.7274897441041223e-08,
"loss": 0.2276,
"step": 14880
},
{
"epoch": 2.72,
"grad_norm": 1.3996675489557702,
"learning_rate": 2.6179258452222687e-08,
"loss": 0.2223,
"step": 14912
},
{
"epoch": 2.73,
"grad_norm": 1.3797514575768821,
"learning_rate": 2.5105487311772478e-08,
"loss": 0.2346,
"step": 14944
},
{
"epoch": 2.73,
"grad_norm": 1.3453274139513625,
"learning_rate": 2.4053633576542833e-08,
"loss": 0.2148,
"step": 14976
},
{
"epoch": 2.74,
"grad_norm": 1.4296834231597417,
"learning_rate": 2.3023745791850623e-08,
"loss": 0.2252,
"step": 15008
},
{
"epoch": 2.74,
"grad_norm": 1.5820604044938296,
"learning_rate": 2.201587148923656e-08,
"loss": 0.2203,
"step": 15040
},
{
"epoch": 2.75,
"grad_norm": 1.3679161695840207,
"learning_rate": 2.103005718427181e-08,
"loss": 0.2096,
"step": 15072
},
{
"epoch": 2.76,
"grad_norm": 1.2454498587173481,
"learning_rate": 2.0066348374411122e-08,
"loss": 0.228,
"step": 15104
},
{
"epoch": 2.76,
"grad_norm": 1.094395180196629,
"learning_rate": 1.912478953689306e-08,
"loss": 0.2183,
"step": 15136
},
{
"epoch": 2.77,
"grad_norm": 1.2797588189588773,
"learning_rate": 1.8205424126687187e-08,
"loss": 0.2218,
"step": 15168
},
{
"epoch": 2.77,
"grad_norm": 1.4533835000058788,
"learning_rate": 1.7308294574488347e-08,
"loss": 0.2226,
"step": 15200
},
{
"epoch": 2.78,
"grad_norm": 1.4492805126695452,
"learning_rate": 1.6433442284758902e-08,
"loss": 0.2274,
"step": 15232
},
{
"epoch": 2.78,
"grad_norm": 1.4350039162599688,
"learning_rate": 1.5580907633817475e-08,
"loss": 0.2282,
"step": 15264
},
{
"epoch": 2.79,
"grad_norm": 1.1810113931037964,
"learning_rate": 1.475072996797555e-08,
"loss": 0.2224,
"step": 15296
},
{
"epoch": 2.8,
"grad_norm": 1.1286561070757017,
"learning_rate": 1.3942947601721488e-08,
"loss": 0.215,
"step": 15328
},
{
"epoch": 2.8,
"grad_norm": 1.1871769686911617,
"learning_rate": 1.3157597815952382e-08,
"loss": 0.2214,
"step": 15360
},
{
"epoch": 2.81,
"grad_norm": 1.2452370435442608,
"learning_rate": 1.2394716856253374e-08,
"loss": 0.2122,
"step": 15392
},
{
"epoch": 2.81,
"grad_norm": 1.2518679025330413,
"learning_rate": 1.1654339931224887e-08,
"loss": 0.217,
"step": 15424
},
{
"epoch": 2.82,
"grad_norm": 1.252731324308369,
"learning_rate": 1.0936501210857651e-08,
"loss": 0.2317,
"step": 15456
},
{
"epoch": 2.83,
"grad_norm": 1.3386257390126015,
"learning_rate": 1.0241233824955509e-08,
"loss": 0.2256,
"step": 15488
},
{
"epoch": 2.83,
"grad_norm": 1.1313351017583362,
"learning_rate": 9.568569861606868e-09,
"loss": 0.2341,
"step": 15520
},
{
"epoch": 2.84,
"grad_norm": 1.4449940727667887,
"learning_rate": 8.918540365703331e-09,
"loss": 0.2166,
"step": 15552
},
{
"epoch": 2.84,
"grad_norm": 1.4530241900991612,
"learning_rate": 8.291175337506894e-09,
"loss": 0.2408,
"step": 15584
},
{
"epoch": 2.85,
"grad_norm": 1.2609793071770357,
"learning_rate": 7.68650373126567e-09,
"loss": 0.2046,
"step": 15616
},
{
"epoch": 2.85,
"grad_norm": 1.1344728672758062,
"learning_rate": 7.10455345387756e-09,
"loss": 0.2196,
"step": 15648
},
{
"epoch": 2.86,
"grad_norm": 1.5048112445610096,
"learning_rate": 6.545351363601959e-09,
"loss": 0.2214,
"step": 15680
},
{
"epoch": 2.87,
"grad_norm": 1.4050748695357576,
"learning_rate": 6.00892326882052e-09,
"loss": 0.2265,
"step": 15712
},
{
"epoch": 2.87,
"grad_norm": 1.1997477163250476,
"learning_rate": 5.495293926845834e-09,
"loss": 0.2163,
"step": 15744
},
{
"epoch": 2.88,
"grad_norm": 1.3144891186853807,
"learning_rate": 5.00448704277906e-09,
"loss": 0.224,
"step": 15776
},
{
"epoch": 2.88,
"grad_norm": 1.353851004486482,
"learning_rate": 4.536525268415581e-09,
"loss": 0.2127,
"step": 15808
},
{
"epoch": 2.89,
"grad_norm": 1.4192328708826423,
"learning_rate": 4.091430201199841e-09,
"loss": 0.221,
"step": 15840
},
{
"epoch": 2.9,
"grad_norm": 1.3274920629279923,
"learning_rate": 3.6692223832283653e-09,
"loss": 0.2137,
"step": 15872
},
{
"epoch": 2.9,
"grad_norm": 1.2096820045248131,
"learning_rate": 3.2699213003019588e-09,
"loss": 0.2164,
"step": 15904
},
{
"epoch": 2.91,
"grad_norm": 1.4949483678866275,
"learning_rate": 2.8935453810260413e-09,
"loss": 0.211,
"step": 15936
},
{
"epoch": 2.91,
"grad_norm": 1.4480264990764071,
"learning_rate": 2.5401119959606033e-09,
"loss": 0.2219,
"step": 15968
},
{
"epoch": 2.92,
"grad_norm": 1.3984591809802007,
"learning_rate": 2.209637456817848e-09,
"loss": 0.2187,
"step": 16000
},
{
"epoch": 2.93,
"grad_norm": 1.2138235012481149,
"learning_rate": 1.9021370157100703e-09,
"loss": 0.2044,
"step": 16032
},
{
"epoch": 2.93,
"grad_norm": 1.4196275071259965,
"learning_rate": 1.6176248644453883e-09,
"loss": 0.2206,
"step": 16064
},
{
"epoch": 2.94,
"grad_norm": 1.4494340877185368,
"learning_rate": 1.3561141338726545e-09,
"loss": 0.2155,
"step": 16096
},
{
"epoch": 2.94,
"grad_norm": 1.3684818624338524,
"learning_rate": 1.117616893275719e-09,
"loss": 0.228,
"step": 16128
},
{
"epoch": 2.95,
"grad_norm": 4.300044658083074,
"learning_rate": 9.02144149816153e-10,
"loss": 0.2138,
"step": 16160
},
{
"epoch": 2.95,
"grad_norm": 1.4879340606023383,
"learning_rate": 7.097058480255991e-10,
"loss": 0.2135,
"step": 16192
},
{
"epoch": 2.96,
"grad_norm": 1.628757008031087,
"learning_rate": 5.403108693462499e-10,
"loss": 0.2169,
"step": 16224
},
{
"epoch": 2.97,
"grad_norm": 1.355596318247962,
"learning_rate": 3.9396703172150936e-10,
"loss": 0.212,
"step": 16256
},
{
"epoch": 2.97,
"grad_norm": 1.1638938575418796,
"learning_rate": 2.706810892348921e-10,
"loss": 0.2177,
"step": 16288
},
{
"epoch": 2.98,
"grad_norm": 1.2545697493468788,
"learning_rate": 1.704587317983286e-10,
"loss": 0.2226,
"step": 16320
},
{
"epoch": 2.98,
"grad_norm": 1.2958637026668016,
"learning_rate": 9.330458488959746e-11,
"loss": 0.2206,
"step": 16352
},
{
"epoch": 2.99,
"grad_norm": 1.1646489691129969,
"learning_rate": 3.922220933882947e-11,
"loss": 0.2219,
"step": 16384
},
{
"epoch": 3.0,
"grad_norm": 1.126290495132723,
"learning_rate": 8.214101164305543e-12,
"loss": 0.2393,
"step": 16416
}
],
"logging_steps": 32,
"max_steps": 16443,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 3289,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}