sedrickkeh's picture
End of training
2e9a856 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.0,
"eval_steps": 500,
"global_step": 621,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004830917874396135,
"grad_norm": 6.679134368896484,
"learning_rate": 1.5873015873015874e-07,
"loss": 1.2509,
"step": 1
},
{
"epoch": 0.00966183574879227,
"grad_norm": 6.96745491027832,
"learning_rate": 3.174603174603175e-07,
"loss": 1.294,
"step": 2
},
{
"epoch": 0.014492753623188406,
"grad_norm": 6.5353102684021,
"learning_rate": 4.7619047619047623e-07,
"loss": 1.2325,
"step": 3
},
{
"epoch": 0.01932367149758454,
"grad_norm": 6.902965068817139,
"learning_rate": 6.34920634920635e-07,
"loss": 1.2676,
"step": 4
},
{
"epoch": 0.024154589371980676,
"grad_norm": 6.7968244552612305,
"learning_rate": 7.936507936507937e-07,
"loss": 1.2375,
"step": 5
},
{
"epoch": 0.028985507246376812,
"grad_norm": 6.779526233673096,
"learning_rate": 9.523809523809525e-07,
"loss": 1.2592,
"step": 6
},
{
"epoch": 0.033816425120772944,
"grad_norm": 6.198831558227539,
"learning_rate": 1.111111111111111e-06,
"loss": 1.2144,
"step": 7
},
{
"epoch": 0.03864734299516908,
"grad_norm": 5.981875419616699,
"learning_rate": 1.26984126984127e-06,
"loss": 1.1995,
"step": 8
},
{
"epoch": 0.043478260869565216,
"grad_norm": 4.696550369262695,
"learning_rate": 1.4285714285714286e-06,
"loss": 1.1667,
"step": 9
},
{
"epoch": 0.04830917874396135,
"grad_norm": 4.801881790161133,
"learning_rate": 1.5873015873015873e-06,
"loss": 1.2129,
"step": 10
},
{
"epoch": 0.05314009661835749,
"grad_norm": 4.528290748596191,
"learning_rate": 1.746031746031746e-06,
"loss": 1.2057,
"step": 11
},
{
"epoch": 0.057971014492753624,
"grad_norm": 2.914287805557251,
"learning_rate": 1.904761904761905e-06,
"loss": 1.1337,
"step": 12
},
{
"epoch": 0.06280193236714976,
"grad_norm": 2.909637212753296,
"learning_rate": 2.0634920634920634e-06,
"loss": 1.1577,
"step": 13
},
{
"epoch": 0.06763285024154589,
"grad_norm": 2.6452553272247314,
"learning_rate": 2.222222222222222e-06,
"loss": 1.1333,
"step": 14
},
{
"epoch": 0.07246376811594203,
"grad_norm": 2.5321853160858154,
"learning_rate": 2.380952380952381e-06,
"loss": 1.1074,
"step": 15
},
{
"epoch": 0.07729468599033816,
"grad_norm": 3.3950793743133545,
"learning_rate": 2.53968253968254e-06,
"loss": 1.1057,
"step": 16
},
{
"epoch": 0.0821256038647343,
"grad_norm": 3.9431092739105225,
"learning_rate": 2.6984126984126986e-06,
"loss": 1.1045,
"step": 17
},
{
"epoch": 0.08695652173913043,
"grad_norm": 3.998680830001831,
"learning_rate": 2.8571428571428573e-06,
"loss": 1.1112,
"step": 18
},
{
"epoch": 0.09178743961352658,
"grad_norm": 3.75687837600708,
"learning_rate": 3.015873015873016e-06,
"loss": 1.0995,
"step": 19
},
{
"epoch": 0.0966183574879227,
"grad_norm": 3.270338535308838,
"learning_rate": 3.1746031746031746e-06,
"loss": 1.0742,
"step": 20
},
{
"epoch": 0.10144927536231885,
"grad_norm": 2.4752681255340576,
"learning_rate": 3.3333333333333333e-06,
"loss": 1.0497,
"step": 21
},
{
"epoch": 0.10628019323671498,
"grad_norm": 2.0890939235687256,
"learning_rate": 3.492063492063492e-06,
"loss": 1.0556,
"step": 22
},
{
"epoch": 0.1111111111111111,
"grad_norm": 1.8987175226211548,
"learning_rate": 3.6507936507936507e-06,
"loss": 1.0459,
"step": 23
},
{
"epoch": 0.11594202898550725,
"grad_norm": 1.8444652557373047,
"learning_rate": 3.80952380952381e-06,
"loss": 1.0059,
"step": 24
},
{
"epoch": 0.12077294685990338,
"grad_norm": 1.6351381540298462,
"learning_rate": 3.968253968253968e-06,
"loss": 0.9934,
"step": 25
},
{
"epoch": 0.12560386473429952,
"grad_norm": 1.5129081010818481,
"learning_rate": 4.126984126984127e-06,
"loss": 1.0023,
"step": 26
},
{
"epoch": 0.13043478260869565,
"grad_norm": 1.260694146156311,
"learning_rate": 4.2857142857142855e-06,
"loss": 1.0085,
"step": 27
},
{
"epoch": 0.13526570048309178,
"grad_norm": 1.1098384857177734,
"learning_rate": 4.444444444444444e-06,
"loss": 1.0117,
"step": 28
},
{
"epoch": 0.14009661835748793,
"grad_norm": 1.0046857595443726,
"learning_rate": 4.603174603174604e-06,
"loss": 1.022,
"step": 29
},
{
"epoch": 0.14492753623188406,
"grad_norm": 1.019243836402893,
"learning_rate": 4.761904761904762e-06,
"loss": 0.9697,
"step": 30
},
{
"epoch": 0.1497584541062802,
"grad_norm": 1.1034119129180908,
"learning_rate": 4.920634920634921e-06,
"loss": 0.9374,
"step": 31
},
{
"epoch": 0.15458937198067632,
"grad_norm": 1.0104053020477295,
"learning_rate": 5.07936507936508e-06,
"loss": 0.9662,
"step": 32
},
{
"epoch": 0.15942028985507245,
"grad_norm": 0.8873758316040039,
"learning_rate": 5.2380952380952384e-06,
"loss": 0.9419,
"step": 33
},
{
"epoch": 0.1642512077294686,
"grad_norm": 0.892684817314148,
"learning_rate": 5.396825396825397e-06,
"loss": 0.9036,
"step": 34
},
{
"epoch": 0.16908212560386474,
"grad_norm": 0.867564857006073,
"learning_rate": 5.555555555555557e-06,
"loss": 0.9402,
"step": 35
},
{
"epoch": 0.17391304347826086,
"grad_norm": 0.9065471887588501,
"learning_rate": 5.7142857142857145e-06,
"loss": 0.9335,
"step": 36
},
{
"epoch": 0.178743961352657,
"grad_norm": 0.8308513760566711,
"learning_rate": 5.873015873015874e-06,
"loss": 0.9084,
"step": 37
},
{
"epoch": 0.18357487922705315,
"grad_norm": 0.8211532831192017,
"learning_rate": 6.031746031746032e-06,
"loss": 0.9343,
"step": 38
},
{
"epoch": 0.18840579710144928,
"grad_norm": 0.9129012227058411,
"learning_rate": 6.1904761904761914e-06,
"loss": 0.9141,
"step": 39
},
{
"epoch": 0.1932367149758454,
"grad_norm": 0.9120469093322754,
"learning_rate": 6.349206349206349e-06,
"loss": 0.9409,
"step": 40
},
{
"epoch": 0.19806763285024154,
"grad_norm": 0.8367810845375061,
"learning_rate": 6.507936507936509e-06,
"loss": 0.9085,
"step": 41
},
{
"epoch": 0.2028985507246377,
"grad_norm": 0.7510740756988525,
"learning_rate": 6.666666666666667e-06,
"loss": 0.9286,
"step": 42
},
{
"epoch": 0.20772946859903382,
"grad_norm": 0.6897188425064087,
"learning_rate": 6.825396825396826e-06,
"loss": 0.9239,
"step": 43
},
{
"epoch": 0.21256038647342995,
"grad_norm": 0.7415544986724854,
"learning_rate": 6.984126984126984e-06,
"loss": 0.9135,
"step": 44
},
{
"epoch": 0.21739130434782608,
"grad_norm": 0.8625911474227905,
"learning_rate": 7.1428571428571436e-06,
"loss": 0.9253,
"step": 45
},
{
"epoch": 0.2222222222222222,
"grad_norm": 0.745945930480957,
"learning_rate": 7.301587301587301e-06,
"loss": 0.9366,
"step": 46
},
{
"epoch": 0.22705314009661837,
"grad_norm": 0.8231242895126343,
"learning_rate": 7.460317460317461e-06,
"loss": 0.9105,
"step": 47
},
{
"epoch": 0.2318840579710145,
"grad_norm": 0.6277362704277039,
"learning_rate": 7.61904761904762e-06,
"loss": 0.8868,
"step": 48
},
{
"epoch": 0.23671497584541062,
"grad_norm": 0.6948869228363037,
"learning_rate": 7.77777777777778e-06,
"loss": 0.887,
"step": 49
},
{
"epoch": 0.24154589371980675,
"grad_norm": 0.5923482775688171,
"learning_rate": 7.936507936507936e-06,
"loss": 0.906,
"step": 50
},
{
"epoch": 0.2463768115942029,
"grad_norm": 0.6197386980056763,
"learning_rate": 8.095238095238097e-06,
"loss": 0.9321,
"step": 51
},
{
"epoch": 0.25120772946859904,
"grad_norm": 0.8314595222473145,
"learning_rate": 8.253968253968254e-06,
"loss": 0.9195,
"step": 52
},
{
"epoch": 0.2560386473429952,
"grad_norm": 0.7024510502815247,
"learning_rate": 8.412698412698414e-06,
"loss": 0.932,
"step": 53
},
{
"epoch": 0.2608695652173913,
"grad_norm": 0.8203377723693848,
"learning_rate": 8.571428571428571e-06,
"loss": 0.9115,
"step": 54
},
{
"epoch": 0.26570048309178745,
"grad_norm": 0.7045034170150757,
"learning_rate": 8.730158730158731e-06,
"loss": 0.917,
"step": 55
},
{
"epoch": 0.27053140096618356,
"grad_norm": 0.6258231401443481,
"learning_rate": 8.888888888888888e-06,
"loss": 0.8916,
"step": 56
},
{
"epoch": 0.2753623188405797,
"grad_norm": 0.7873698472976685,
"learning_rate": 9.047619047619049e-06,
"loss": 0.8831,
"step": 57
},
{
"epoch": 0.28019323671497587,
"grad_norm": 0.9555951952934265,
"learning_rate": 9.206349206349207e-06,
"loss": 0.901,
"step": 58
},
{
"epoch": 0.28502415458937197,
"grad_norm": 0.6864539384841919,
"learning_rate": 9.365079365079366e-06,
"loss": 0.9124,
"step": 59
},
{
"epoch": 0.2898550724637681,
"grad_norm": 0.9573421478271484,
"learning_rate": 9.523809523809525e-06,
"loss": 0.8774,
"step": 60
},
{
"epoch": 0.2946859903381642,
"grad_norm": 0.775773286819458,
"learning_rate": 9.682539682539683e-06,
"loss": 0.9277,
"step": 61
},
{
"epoch": 0.2995169082125604,
"grad_norm": 0.639119029045105,
"learning_rate": 9.841269841269842e-06,
"loss": 0.8504,
"step": 62
},
{
"epoch": 0.30434782608695654,
"grad_norm": 0.912753164768219,
"learning_rate": 1e-05,
"loss": 0.9119,
"step": 63
},
{
"epoch": 0.30917874396135264,
"grad_norm": 0.695167064666748,
"learning_rate": 9.999920755303033e-06,
"loss": 0.8696,
"step": 64
},
{
"epoch": 0.3140096618357488,
"grad_norm": 0.8863600492477417,
"learning_rate": 9.999683023724021e-06,
"loss": 0.8948,
"step": 65
},
{
"epoch": 0.3188405797101449,
"grad_norm": 0.7954296469688416,
"learning_rate": 9.99928681279855e-06,
"loss": 0.9085,
"step": 66
},
{
"epoch": 0.32367149758454106,
"grad_norm": 0.7403387427330017,
"learning_rate": 9.998732135085665e-06,
"loss": 0.9368,
"step": 67
},
{
"epoch": 0.3285024154589372,
"grad_norm": 0.7288274168968201,
"learning_rate": 9.998019008167476e-06,
"loss": 0.8966,
"step": 68
},
{
"epoch": 0.3333333333333333,
"grad_norm": 0.836815595626831,
"learning_rate": 9.99714745464859e-06,
"loss": 0.8806,
"step": 69
},
{
"epoch": 0.33816425120772947,
"grad_norm": 0.8034535646438599,
"learning_rate": 9.99611750215541e-06,
"loss": 0.9018,
"step": 70
},
{
"epoch": 0.34299516908212563,
"grad_norm": 0.7000673413276672,
"learning_rate": 9.994929183335237e-06,
"loss": 0.8668,
"step": 71
},
{
"epoch": 0.34782608695652173,
"grad_norm": 0.7168377637863159,
"learning_rate": 9.993582535855265e-06,
"loss": 0.8809,
"step": 72
},
{
"epoch": 0.3526570048309179,
"grad_norm": 0.8486096858978271,
"learning_rate": 9.992077602401358e-06,
"loss": 0.8878,
"step": 73
},
{
"epoch": 0.357487922705314,
"grad_norm": 0.6581878662109375,
"learning_rate": 9.990414430676716e-06,
"loss": 0.8555,
"step": 74
},
{
"epoch": 0.36231884057971014,
"grad_norm": 0.8234553933143616,
"learning_rate": 9.988593073400354e-06,
"loss": 0.8852,
"step": 75
},
{
"epoch": 0.3671497584541063,
"grad_norm": 0.7650076746940613,
"learning_rate": 9.986613588305435e-06,
"loss": 0.8572,
"step": 76
},
{
"epoch": 0.3719806763285024,
"grad_norm": 0.6619766354560852,
"learning_rate": 9.984476038137437e-06,
"loss": 0.9015,
"step": 77
},
{
"epoch": 0.37681159420289856,
"grad_norm": 0.9842355847358704,
"learning_rate": 9.982180490652165e-06,
"loss": 0.8912,
"step": 78
},
{
"epoch": 0.38164251207729466,
"grad_norm": 0.8628110289573669,
"learning_rate": 9.979727018613607e-06,
"loss": 0.9033,
"step": 79
},
{
"epoch": 0.3864734299516908,
"grad_norm": 0.8167057633399963,
"learning_rate": 9.977115699791622e-06,
"loss": 0.8583,
"step": 80
},
{
"epoch": 0.391304347826087,
"grad_norm": 0.8511431813240051,
"learning_rate": 9.974346616959476e-06,
"loss": 0.8625,
"step": 81
},
{
"epoch": 0.3961352657004831,
"grad_norm": 0.8892733454704285,
"learning_rate": 9.971419857891223e-06,
"loss": 0.8967,
"step": 82
},
{
"epoch": 0.40096618357487923,
"grad_norm": 0.6108592748641968,
"learning_rate": 9.968335515358916e-06,
"loss": 0.8727,
"step": 83
},
{
"epoch": 0.4057971014492754,
"grad_norm": 0.9057552218437195,
"learning_rate": 9.965093687129669e-06,
"loss": 0.8489,
"step": 84
},
{
"epoch": 0.4106280193236715,
"grad_norm": 0.7670313715934753,
"learning_rate": 9.961694475962562e-06,
"loss": 0.8862,
"step": 85
},
{
"epoch": 0.41545893719806765,
"grad_norm": 0.7009760141372681,
"learning_rate": 9.95813798960538e-06,
"loss": 0.8481,
"step": 86
},
{
"epoch": 0.42028985507246375,
"grad_norm": 0.8547497391700745,
"learning_rate": 9.954424340791195e-06,
"loss": 0.8574,
"step": 87
},
{
"epoch": 0.4251207729468599,
"grad_norm": 0.6579576730728149,
"learning_rate": 9.950553647234798e-06,
"loss": 0.8876,
"step": 88
},
{
"epoch": 0.42995169082125606,
"grad_norm": 0.7313647270202637,
"learning_rate": 9.94652603162896e-06,
"loss": 0.8833,
"step": 89
},
{
"epoch": 0.43478260869565216,
"grad_norm": 0.8301870226860046,
"learning_rate": 9.942341621640558e-06,
"loss": 0.842,
"step": 90
},
{
"epoch": 0.4396135265700483,
"grad_norm": 0.634363055229187,
"learning_rate": 9.938000549906509e-06,
"loss": 0.8724,
"step": 91
},
{
"epoch": 0.4444444444444444,
"grad_norm": 0.6919889450073242,
"learning_rate": 9.93350295402958e-06,
"loss": 0.8459,
"step": 92
},
{
"epoch": 0.4492753623188406,
"grad_norm": 0.7910676598548889,
"learning_rate": 9.92884897657402e-06,
"loss": 0.8793,
"step": 93
},
{
"epoch": 0.45410628019323673,
"grad_norm": 0.7085094451904297,
"learning_rate": 9.924038765061042e-06,
"loss": 0.8618,
"step": 94
},
{
"epoch": 0.45893719806763283,
"grad_norm": 0.8088275194168091,
"learning_rate": 9.919072471964146e-06,
"loss": 0.8547,
"step": 95
},
{
"epoch": 0.463768115942029,
"grad_norm": 0.6322200298309326,
"learning_rate": 9.913950254704291e-06,
"loss": 0.8545,
"step": 96
},
{
"epoch": 0.46859903381642515,
"grad_norm": 0.7928032279014587,
"learning_rate": 9.908672275644898e-06,
"loss": 0.8642,
"step": 97
},
{
"epoch": 0.47342995169082125,
"grad_norm": 0.7358760237693787,
"learning_rate": 9.903238702086707e-06,
"loss": 0.8677,
"step": 98
},
{
"epoch": 0.4782608695652174,
"grad_norm": 0.7627845406532288,
"learning_rate": 9.897649706262474e-06,
"loss": 0.8784,
"step": 99
},
{
"epoch": 0.4830917874396135,
"grad_norm": 0.7139146327972412,
"learning_rate": 9.89190546533151e-06,
"loss": 0.8929,
"step": 100
},
{
"epoch": 0.48792270531400966,
"grad_norm": 0.7086196541786194,
"learning_rate": 9.88600616137407e-06,
"loss": 0.8687,
"step": 101
},
{
"epoch": 0.4927536231884058,
"grad_norm": 0.7507730722427368,
"learning_rate": 9.879951981385577e-06,
"loss": 0.8278,
"step": 102
},
{
"epoch": 0.4975845410628019,
"grad_norm": 0.7131170630455017,
"learning_rate": 9.873743117270691e-06,
"loss": 0.8751,
"step": 103
},
{
"epoch": 0.5024154589371981,
"grad_norm": 0.9851748943328857,
"learning_rate": 9.867379765837237e-06,
"loss": 0.8885,
"step": 104
},
{
"epoch": 0.5072463768115942,
"grad_norm": 0.7180512547492981,
"learning_rate": 9.860862128789954e-06,
"loss": 0.8686,
"step": 105
},
{
"epoch": 0.5120772946859904,
"grad_norm": 0.8217784762382507,
"learning_rate": 9.854190412724114e-06,
"loss": 0.8665,
"step": 106
},
{
"epoch": 0.5169082125603864,
"grad_norm": 0.863905668258667,
"learning_rate": 9.847364829118963e-06,
"loss": 0.86,
"step": 107
},
{
"epoch": 0.5217391304347826,
"grad_norm": 0.9537051916122437,
"learning_rate": 9.840385594331022e-06,
"loss": 0.8251,
"step": 108
},
{
"epoch": 0.5265700483091788,
"grad_norm": 0.6802311539649963,
"learning_rate": 9.833252929587231e-06,
"loss": 0.8688,
"step": 109
},
{
"epoch": 0.5314009661835749,
"grad_norm": 0.7750963568687439,
"learning_rate": 9.825967060977933e-06,
"loss": 0.8599,
"step": 110
},
{
"epoch": 0.5362318840579711,
"grad_norm": 0.7233647108078003,
"learning_rate": 9.818528219449705e-06,
"loss": 0.8659,
"step": 111
},
{
"epoch": 0.5410628019323671,
"grad_norm": 0.6166982054710388,
"learning_rate": 9.810936640798046e-06,
"loss": 0.853,
"step": 112
},
{
"epoch": 0.5458937198067633,
"grad_norm": 0.6997689008712769,
"learning_rate": 9.803192565659898e-06,
"loss": 0.8404,
"step": 113
},
{
"epoch": 0.5507246376811594,
"grad_norm": 0.6236562728881836,
"learning_rate": 9.795296239506011e-06,
"loss": 0.8722,
"step": 114
},
{
"epoch": 0.5555555555555556,
"grad_norm": 0.7246277332305908,
"learning_rate": 9.78724791263318e-06,
"loss": 0.8656,
"step": 115
},
{
"epoch": 0.5603864734299517,
"grad_norm": 0.6707713603973389,
"learning_rate": 9.779047840156288e-06,
"loss": 0.8821,
"step": 116
},
{
"epoch": 0.5652173913043478,
"grad_norm": 0.596466064453125,
"learning_rate": 9.770696282000245e-06,
"loss": 0.8461,
"step": 117
},
{
"epoch": 0.5700483091787439,
"grad_norm": 0.6369594931602478,
"learning_rate": 9.762193502891726e-06,
"loss": 0.8723,
"step": 118
},
{
"epoch": 0.5748792270531401,
"grad_norm": 0.6298103928565979,
"learning_rate": 9.753539772350792e-06,
"loss": 0.8751,
"step": 119
},
{
"epoch": 0.5797101449275363,
"grad_norm": 0.7989038825035095,
"learning_rate": 9.744735364682347e-06,
"loss": 0.8829,
"step": 120
},
{
"epoch": 0.5845410628019324,
"grad_norm": 0.6647093296051025,
"learning_rate": 9.735780558967434e-06,
"loss": 0.8725,
"step": 121
},
{
"epoch": 0.5893719806763285,
"grad_norm": 0.6525388360023499,
"learning_rate": 9.726675639054403e-06,
"loss": 0.8398,
"step": 122
},
{
"epoch": 0.5942028985507246,
"grad_norm": 0.6918199062347412,
"learning_rate": 9.717420893549902e-06,
"loss": 0.8523,
"step": 123
},
{
"epoch": 0.5990338164251208,
"grad_norm": 0.6548693776130676,
"learning_rate": 9.70801661580973e-06,
"loss": 0.8584,
"step": 124
},
{
"epoch": 0.6038647342995169,
"grad_norm": 0.7120753526687622,
"learning_rate": 9.698463103929542e-06,
"loss": 0.8821,
"step": 125
},
{
"epoch": 0.6086956521739131,
"grad_norm": 0.6639739871025085,
"learning_rate": 9.688760660735403e-06,
"loss": 0.8791,
"step": 126
},
{
"epoch": 0.6135265700483091,
"grad_norm": 0.5813497304916382,
"learning_rate": 9.67890959377418e-06,
"loss": 0.8738,
"step": 127
},
{
"epoch": 0.6183574879227053,
"grad_norm": 0.7182924747467041,
"learning_rate": 9.668910215303797e-06,
"loss": 0.8889,
"step": 128
},
{
"epoch": 0.6231884057971014,
"grad_norm": 0.6352912783622742,
"learning_rate": 9.658762842283343e-06,
"loss": 0.8625,
"step": 129
},
{
"epoch": 0.6280193236714976,
"grad_norm": 0.7149222493171692,
"learning_rate": 9.648467796363019e-06,
"loss": 0.8458,
"step": 130
},
{
"epoch": 0.6328502415458938,
"grad_norm": 0.591304361820221,
"learning_rate": 9.638025403873939e-06,
"loss": 0.8241,
"step": 131
},
{
"epoch": 0.6376811594202898,
"grad_norm": 0.6942040920257568,
"learning_rate": 9.627435995817799e-06,
"loss": 0.8459,
"step": 132
},
{
"epoch": 0.642512077294686,
"grad_norm": 0.6825373768806458,
"learning_rate": 9.616699907856368e-06,
"loss": 0.881,
"step": 133
},
{
"epoch": 0.6473429951690821,
"grad_norm": 0.7252668738365173,
"learning_rate": 9.605817480300863e-06,
"loss": 0.822,
"step": 134
},
{
"epoch": 0.6521739130434783,
"grad_norm": 0.702181875705719,
"learning_rate": 9.594789058101154e-06,
"loss": 0.8841,
"step": 135
},
{
"epoch": 0.6570048309178744,
"grad_norm": 0.6287996768951416,
"learning_rate": 9.58361499083483e-06,
"loss": 0.8266,
"step": 136
},
{
"epoch": 0.6618357487922706,
"grad_norm": 0.7371240258216858,
"learning_rate": 9.57229563269612e-06,
"loss": 0.8691,
"step": 137
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.8119018077850342,
"learning_rate": 9.560831342484668e-06,
"loss": 0.8842,
"step": 138
},
{
"epoch": 0.6714975845410628,
"grad_norm": 0.6123917102813721,
"learning_rate": 9.549222483594154e-06,
"loss": 0.82,
"step": 139
},
{
"epoch": 0.6763285024154589,
"grad_norm": 0.7341845631599426,
"learning_rate": 9.53746942400078e-06,
"loss": 0.8919,
"step": 140
},
{
"epoch": 0.6811594202898551,
"grad_norm": 0.6273525357246399,
"learning_rate": 9.525572536251608e-06,
"loss": 0.8453,
"step": 141
},
{
"epoch": 0.6859903381642513,
"grad_norm": 0.6209537386894226,
"learning_rate": 9.513532197452737e-06,
"loss": 0.833,
"step": 142
},
{
"epoch": 0.6908212560386473,
"grad_norm": 0.7676011919975281,
"learning_rate": 9.501348789257373e-06,
"loss": 0.8634,
"step": 143
},
{
"epoch": 0.6956521739130435,
"grad_norm": 0.5898526906967163,
"learning_rate": 9.48902269785371e-06,
"loss": 0.8844,
"step": 144
},
{
"epoch": 0.7004830917874396,
"grad_norm": 0.7219299077987671,
"learning_rate": 9.476554313952697e-06,
"loss": 0.8543,
"step": 145
},
{
"epoch": 0.7053140096618358,
"grad_norm": 0.6188173890113831,
"learning_rate": 9.46394403277566e-06,
"loss": 0.8325,
"step": 146
},
{
"epoch": 0.7101449275362319,
"grad_norm": 0.6188879609107971,
"learning_rate": 9.451192254041759e-06,
"loss": 0.8591,
"step": 147
},
{
"epoch": 0.714975845410628,
"grad_norm": 0.648301899433136,
"learning_rate": 9.438299381955333e-06,
"loss": 0.8646,
"step": 148
},
{
"epoch": 0.7198067632850241,
"grad_norm": 0.6337832808494568,
"learning_rate": 9.425265825193077e-06,
"loss": 0.8394,
"step": 149
},
{
"epoch": 0.7246376811594203,
"grad_norm": 0.5935664772987366,
"learning_rate": 9.412091996891097e-06,
"loss": 0.8865,
"step": 150
},
{
"epoch": 0.7294685990338164,
"grad_norm": 0.759688675403595,
"learning_rate": 9.398778314631801e-06,
"loss": 0.8644,
"step": 151
},
{
"epoch": 0.7342995169082126,
"grad_norm": 0.65163654088974,
"learning_rate": 9.385325200430679e-06,
"loss": 0.8421,
"step": 152
},
{
"epoch": 0.7391304347826086,
"grad_norm": 0.7104676365852356,
"learning_rate": 9.371733080722911e-06,
"loss": 0.8562,
"step": 153
},
{
"epoch": 0.7439613526570048,
"grad_norm": 0.6826310753822327,
"learning_rate": 9.358002386349862e-06,
"loss": 0.8867,
"step": 154
},
{
"epoch": 0.748792270531401,
"grad_norm": 0.6869804263114929,
"learning_rate": 9.34413355254542e-06,
"loss": 0.8403,
"step": 155
},
{
"epoch": 0.7536231884057971,
"grad_norm": 0.6028906106948853,
"learning_rate": 9.330127018922195e-06,
"loss": 0.8489,
"step": 156
},
{
"epoch": 0.7584541062801933,
"grad_norm": 0.8058885335922241,
"learning_rate": 9.31598322945759e-06,
"loss": 0.8623,
"step": 157
},
{
"epoch": 0.7632850241545893,
"grad_norm": 0.5606011748313904,
"learning_rate": 9.301702632479734e-06,
"loss": 0.8434,
"step": 158
},
{
"epoch": 0.7681159420289855,
"grad_norm": 0.7772120237350464,
"learning_rate": 9.287285680653254e-06,
"loss": 0.8924,
"step": 159
},
{
"epoch": 0.7729468599033816,
"grad_norm": 0.6535057425498962,
"learning_rate": 9.272732830964948e-06,
"loss": 0.8694,
"step": 160
},
{
"epoch": 0.7777777777777778,
"grad_norm": 0.6197230219841003,
"learning_rate": 9.258044544709276e-06,
"loss": 0.8541,
"step": 161
},
{
"epoch": 0.782608695652174,
"grad_norm": 0.6519168019294739,
"learning_rate": 9.243221287473755e-06,
"loss": 0.8679,
"step": 162
},
{
"epoch": 0.7874396135265701,
"grad_norm": 0.8101065158843994,
"learning_rate": 9.228263529124199e-06,
"loss": 0.8318,
"step": 163
},
{
"epoch": 0.7922705314009661,
"grad_norm": 0.6714143753051758,
"learning_rate": 9.21317174378982e-06,
"loss": 0.8526,
"step": 164
},
{
"epoch": 0.7971014492753623,
"grad_norm": 0.8356069922447205,
"learning_rate": 9.197946409848196e-06,
"loss": 0.8707,
"step": 165
},
{
"epoch": 0.8019323671497585,
"grad_norm": 0.6714097857475281,
"learning_rate": 9.182588009910119e-06,
"loss": 0.8601,
"step": 166
},
{
"epoch": 0.8067632850241546,
"grad_norm": 0.6646215319633484,
"learning_rate": 9.167097030804289e-06,
"loss": 0.8398,
"step": 167
},
{
"epoch": 0.8115942028985508,
"grad_norm": 0.6809929609298706,
"learning_rate": 9.151473963561884e-06,
"loss": 0.7931,
"step": 168
},
{
"epoch": 0.8164251207729468,
"grad_norm": 0.7758853435516357,
"learning_rate": 9.135719303400995e-06,
"loss": 0.8542,
"step": 169
},
{
"epoch": 0.821256038647343,
"grad_norm": 0.743437647819519,
"learning_rate": 9.119833549710927e-06,
"loss": 0.8251,
"step": 170
},
{
"epoch": 0.8260869565217391,
"grad_norm": 0.6809033155441284,
"learning_rate": 9.103817206036383e-06,
"loss": 0.8496,
"step": 171
},
{
"epoch": 0.8309178743961353,
"grad_norm": 0.6170421838760376,
"learning_rate": 9.087670780061477e-06,
"loss": 0.8558,
"step": 172
},
{
"epoch": 0.8357487922705314,
"grad_norm": 0.5970268845558167,
"learning_rate": 9.071394783593664e-06,
"loss": 0.8295,
"step": 173
},
{
"epoch": 0.8405797101449275,
"grad_norm": 0.6291173100471497,
"learning_rate": 9.054989732547507e-06,
"loss": 0.859,
"step": 174
},
{
"epoch": 0.8454106280193237,
"grad_norm": 0.6694471836090088,
"learning_rate": 9.038456146928325e-06,
"loss": 0.8654,
"step": 175
},
{
"epoch": 0.8502415458937198,
"grad_norm": 0.6172890663146973,
"learning_rate": 9.021794550815713e-06,
"loss": 0.8752,
"step": 176
},
{
"epoch": 0.855072463768116,
"grad_norm": 0.58245849609375,
"learning_rate": 9.005005472346923e-06,
"loss": 0.8326,
"step": 177
},
{
"epoch": 0.8599033816425121,
"grad_norm": 0.5857954621315002,
"learning_rate": 8.988089443700131e-06,
"loss": 0.8185,
"step": 178
},
{
"epoch": 0.8647342995169082,
"grad_norm": 0.5742454528808594,
"learning_rate": 8.971047001077561e-06,
"loss": 0.8957,
"step": 179
},
{
"epoch": 0.8695652173913043,
"grad_norm": 0.696999728679657,
"learning_rate": 8.953878684688492e-06,
"loss": 0.8769,
"step": 180
},
{
"epoch": 0.8743961352657005,
"grad_norm": 0.6432103514671326,
"learning_rate": 8.936585038732143e-06,
"loss": 0.863,
"step": 181
},
{
"epoch": 0.8792270531400966,
"grad_norm": 0.7360765933990479,
"learning_rate": 8.919166611380397e-06,
"loss": 0.8449,
"step": 182
},
{
"epoch": 0.8840579710144928,
"grad_norm": 0.6782159805297852,
"learning_rate": 8.90162395476046e-06,
"loss": 0.8473,
"step": 183
},
{
"epoch": 0.8888888888888888,
"grad_norm": 0.6303816437721252,
"learning_rate": 8.883957624937333e-06,
"loss": 0.839,
"step": 184
},
{
"epoch": 0.893719806763285,
"grad_norm": 0.6260365843772888,
"learning_rate": 8.866168181896198e-06,
"loss": 0.876,
"step": 185
},
{
"epoch": 0.8985507246376812,
"grad_norm": 0.5933541059494019,
"learning_rate": 8.848256189524661e-06,
"loss": 0.8496,
"step": 186
},
{
"epoch": 0.9033816425120773,
"grad_norm": 0.5709136724472046,
"learning_rate": 8.83022221559489e-06,
"loss": 0.8522,
"step": 187
},
{
"epoch": 0.9082125603864735,
"grad_norm": 0.5638396143913269,
"learning_rate": 8.812066831745602e-06,
"loss": 0.8649,
"step": 188
},
{
"epoch": 0.9130434782608695,
"grad_norm": 0.6326948404312134,
"learning_rate": 8.793790613463956e-06,
"loss": 0.8464,
"step": 189
},
{
"epoch": 0.9178743961352657,
"grad_norm": 0.6702078580856323,
"learning_rate": 8.775394140067299e-06,
"loss": 0.8051,
"step": 190
},
{
"epoch": 0.9227053140096618,
"grad_norm": 0.6150868535041809,
"learning_rate": 8.756877994684818e-06,
"loss": 0.8543,
"step": 191
},
{
"epoch": 0.927536231884058,
"grad_norm": 0.6802798509597778,
"learning_rate": 8.738242764239046e-06,
"loss": 0.8202,
"step": 192
},
{
"epoch": 0.9323671497584541,
"grad_norm": 0.6473823189735413,
"learning_rate": 8.719489039427256e-06,
"loss": 0.8058,
"step": 193
},
{
"epoch": 0.9371980676328503,
"grad_norm": 0.6301891207695007,
"learning_rate": 8.700617414702746e-06,
"loss": 0.8518,
"step": 194
},
{
"epoch": 0.9420289855072463,
"grad_norm": 0.5725292563438416,
"learning_rate": 8.681628488255986e-06,
"loss": 0.8444,
"step": 195
},
{
"epoch": 0.9468599033816425,
"grad_norm": 0.6120401620864868,
"learning_rate": 8.66252286199567e-06,
"loss": 0.8505,
"step": 196
},
{
"epoch": 0.9516908212560387,
"grad_norm": 0.6610070466995239,
"learning_rate": 8.643301141529619e-06,
"loss": 0.8238,
"step": 197
},
{
"epoch": 0.9565217391304348,
"grad_norm": 0.5605831146240234,
"learning_rate": 8.6239639361456e-06,
"loss": 0.8455,
"step": 198
},
{
"epoch": 0.961352657004831,
"grad_norm": 0.5961610674858093,
"learning_rate": 8.604511858792006e-06,
"loss": 0.8275,
"step": 199
},
{
"epoch": 0.966183574879227,
"grad_norm": 0.5695908069610596,
"learning_rate": 8.584945526058426e-06,
"loss": 0.8218,
"step": 200
},
{
"epoch": 0.9710144927536232,
"grad_norm": 0.6564455032348633,
"learning_rate": 8.565265558156101e-06,
"loss": 0.8291,
"step": 201
},
{
"epoch": 0.9758454106280193,
"grad_norm": 0.5902118682861328,
"learning_rate": 8.545472578898276e-06,
"loss": 0.87,
"step": 202
},
{
"epoch": 0.9806763285024155,
"grad_norm": 0.5982496738433838,
"learning_rate": 8.525567215680397e-06,
"loss": 0.8411,
"step": 203
},
{
"epoch": 0.9855072463768116,
"grad_norm": 0.5730376243591309,
"learning_rate": 8.505550099460264e-06,
"loss": 0.806,
"step": 204
},
{
"epoch": 0.9903381642512077,
"grad_norm": 0.6329107284545898,
"learning_rate": 8.485421864737997e-06,
"loss": 0.8181,
"step": 205
},
{
"epoch": 0.9951690821256038,
"grad_norm": 0.5955251455307007,
"learning_rate": 8.465183149535939e-06,
"loss": 0.8617,
"step": 206
},
{
"epoch": 1.0,
"grad_norm": 0.7374234199523926,
"learning_rate": 8.444834595378434e-06,
"loss": 0.8778,
"step": 207
},
{
"epoch": 1.0048309178743962,
"grad_norm": 0.5692369937896729,
"learning_rate": 8.424376847271483e-06,
"loss": 0.7686,
"step": 208
},
{
"epoch": 1.0096618357487923,
"grad_norm": 0.5868251919746399,
"learning_rate": 8.403810553682307e-06,
"loss": 0.7638,
"step": 209
},
{
"epoch": 1.0144927536231885,
"grad_norm": 0.6814491152763367,
"learning_rate": 8.383136366518788e-06,
"loss": 0.7903,
"step": 210
},
{
"epoch": 1.0193236714975846,
"grad_norm": 0.5972342491149902,
"learning_rate": 8.362354941108803e-06,
"loss": 0.7882,
"step": 211
},
{
"epoch": 1.0241545893719808,
"grad_norm": 0.709455668926239,
"learning_rate": 8.341466936179457e-06,
"loss": 0.7901,
"step": 212
},
{
"epoch": 1.0289855072463767,
"grad_norm": 0.6567758321762085,
"learning_rate": 8.320473013836197e-06,
"loss": 0.8157,
"step": 213
},
{
"epoch": 1.0338164251207729,
"grad_norm": 0.8612772226333618,
"learning_rate": 8.299373839541829e-06,
"loss": 0.7975,
"step": 214
},
{
"epoch": 1.038647342995169,
"grad_norm": 0.7711191177368164,
"learning_rate": 8.278170082095422e-06,
"loss": 0.7865,
"step": 215
},
{
"epoch": 1.0434782608695652,
"grad_norm": 0.6870383024215698,
"learning_rate": 8.256862413611113e-06,
"loss": 0.8089,
"step": 216
},
{
"epoch": 1.0483091787439613,
"grad_norm": 0.6059009432792664,
"learning_rate": 8.23545150949679e-06,
"loss": 0.7751,
"step": 217
},
{
"epoch": 1.0531400966183575,
"grad_norm": 0.7190768122673035,
"learning_rate": 8.213938048432697e-06,
"loss": 0.7912,
"step": 218
},
{
"epoch": 1.0579710144927537,
"grad_norm": 0.7017462253570557,
"learning_rate": 8.192322712349917e-06,
"loss": 0.779,
"step": 219
},
{
"epoch": 1.0628019323671498,
"grad_norm": 0.5871282815933228,
"learning_rate": 8.170606186408756e-06,
"loss": 0.8335,
"step": 220
},
{
"epoch": 1.067632850241546,
"grad_norm": 0.8193174004554749,
"learning_rate": 8.148789158977012e-06,
"loss": 0.8206,
"step": 221
},
{
"epoch": 1.0724637681159421,
"grad_norm": 0.6280666589736938,
"learning_rate": 8.126872321608185e-06,
"loss": 0.7918,
"step": 222
},
{
"epoch": 1.077294685990338,
"grad_norm": 0.593765139579773,
"learning_rate": 8.104856369019525e-06,
"loss": 0.8393,
"step": 223
},
{
"epoch": 1.0821256038647342,
"grad_norm": 0.6945616602897644,
"learning_rate": 8.08274199907003e-06,
"loss": 0.8082,
"step": 224
},
{
"epoch": 1.0869565217391304,
"grad_norm": 0.6425365209579468,
"learning_rate": 8.060529912738316e-06,
"loss": 0.7877,
"step": 225
},
{
"epoch": 1.0917874396135265,
"grad_norm": 0.602546215057373,
"learning_rate": 8.038220814100403e-06,
"loss": 0.7981,
"step": 226
},
{
"epoch": 1.0966183574879227,
"grad_norm": 0.5716975331306458,
"learning_rate": 8.0158154103074e-06,
"loss": 0.7971,
"step": 227
},
{
"epoch": 1.1014492753623188,
"grad_norm": 0.5490554571151733,
"learning_rate": 7.993314411563075e-06,
"loss": 0.7832,
"step": 228
},
{
"epoch": 1.106280193236715,
"grad_norm": 0.6210470795631409,
"learning_rate": 7.970718531101365e-06,
"loss": 0.8262,
"step": 229
},
{
"epoch": 1.1111111111111112,
"grad_norm": 0.5391402840614319,
"learning_rate": 7.948028485163744e-06,
"loss": 0.7775,
"step": 230
},
{
"epoch": 1.1159420289855073,
"grad_norm": 0.5497602224349976,
"learning_rate": 7.925244992976538e-06,
"loss": 0.7948,
"step": 231
},
{
"epoch": 1.1207729468599035,
"grad_norm": 0.6025332808494568,
"learning_rate": 7.902368776728125e-06,
"loss": 0.8187,
"step": 232
},
{
"epoch": 1.1256038647342996,
"grad_norm": 0.569249153137207,
"learning_rate": 7.879400561546033e-06,
"loss": 0.793,
"step": 233
},
{
"epoch": 1.1304347826086956,
"grad_norm": 0.6411007642745972,
"learning_rate": 7.856341075473963e-06,
"loss": 0.7905,
"step": 234
},
{
"epoch": 1.1352657004830917,
"grad_norm": 0.5739008188247681,
"learning_rate": 7.833191049448706e-06,
"loss": 0.7941,
"step": 235
},
{
"epoch": 1.1400966183574879,
"grad_norm": 0.641865074634552,
"learning_rate": 7.809951217276986e-06,
"loss": 0.8057,
"step": 236
},
{
"epoch": 1.144927536231884,
"grad_norm": 0.6288030743598938,
"learning_rate": 7.786622315612182e-06,
"loss": 0.7624,
"step": 237
},
{
"epoch": 1.1497584541062802,
"grad_norm": 0.6016022562980652,
"learning_rate": 7.763205083930995e-06,
"loss": 0.8071,
"step": 238
},
{
"epoch": 1.1545893719806763,
"grad_norm": 0.517977237701416,
"learning_rate": 7.739700264509993e-06,
"loss": 0.7639,
"step": 239
},
{
"epoch": 1.1594202898550725,
"grad_norm": 0.5750455856323242,
"learning_rate": 7.716108602402094e-06,
"loss": 0.7915,
"step": 240
},
{
"epoch": 1.1642512077294687,
"grad_norm": 0.5806918144226074,
"learning_rate": 7.692430845412946e-06,
"loss": 0.7944,
"step": 241
},
{
"epoch": 1.1690821256038648,
"grad_norm": 0.5185478925704956,
"learning_rate": 7.668667744077215e-06,
"loss": 0.8127,
"step": 242
},
{
"epoch": 1.1739130434782608,
"grad_norm": 0.5086768269538879,
"learning_rate": 7.644820051634813e-06,
"loss": 0.7772,
"step": 243
},
{
"epoch": 1.178743961352657,
"grad_norm": 0.6286478042602539,
"learning_rate": 7.6208885240069995e-06,
"loss": 0.7506,
"step": 244
},
{
"epoch": 1.183574879227053,
"grad_norm": 0.6119205355644226,
"learning_rate": 7.596873919772438e-06,
"loss": 0.7967,
"step": 245
},
{
"epoch": 1.1884057971014492,
"grad_norm": 0.5589543581008911,
"learning_rate": 7.572777000143145e-06,
"loss": 0.8363,
"step": 246
},
{
"epoch": 1.1932367149758454,
"grad_norm": 0.6248103380203247,
"learning_rate": 7.548598528940354e-06,
"loss": 0.8014,
"step": 247
},
{
"epoch": 1.1980676328502415,
"grad_norm": 0.5842779874801636,
"learning_rate": 7.524339272570317e-06,
"loss": 0.7897,
"step": 248
},
{
"epoch": 1.2028985507246377,
"grad_norm": 0.6430383324623108,
"learning_rate": 7.500000000000001e-06,
"loss": 0.7779,
"step": 249
},
{
"epoch": 1.2077294685990339,
"grad_norm": 0.5114961862564087,
"learning_rate": 7.475581482732717e-06,
"loss": 0.7671,
"step": 250
},
{
"epoch": 1.21256038647343,
"grad_norm": 0.5602841973304749,
"learning_rate": 7.451084494783668e-06,
"loss": 0.791,
"step": 251
},
{
"epoch": 1.2173913043478262,
"grad_norm": 0.6211389303207397,
"learning_rate": 7.4265098126554065e-06,
"loss": 0.8377,
"step": 252
},
{
"epoch": 1.2222222222222223,
"grad_norm": 0.6181166768074036,
"learning_rate": 7.401858215313228e-06,
"loss": 0.7738,
"step": 253
},
{
"epoch": 1.2270531400966185,
"grad_norm": 0.5928196907043457,
"learning_rate": 7.3771304841604764e-06,
"loss": 0.776,
"step": 254
},
{
"epoch": 1.2318840579710144,
"grad_norm": 0.5380995273590088,
"learning_rate": 7.352327403013779e-06,
"loss": 0.7791,
"step": 255
},
{
"epoch": 1.2367149758454106,
"grad_norm": 0.6451078057289124,
"learning_rate": 7.327449758078194e-06,
"loss": 0.816,
"step": 256
},
{
"epoch": 1.2415458937198067,
"grad_norm": 0.5372149348258972,
"learning_rate": 7.302498337922293e-06,
"loss": 0.8001,
"step": 257
},
{
"epoch": 1.2463768115942029,
"grad_norm": 0.5098853707313538,
"learning_rate": 7.27747393345317e-06,
"loss": 0.7566,
"step": 258
},
{
"epoch": 1.251207729468599,
"grad_norm": 0.6665354371070862,
"learning_rate": 7.2523773378913655e-06,
"loss": 0.7506,
"step": 259
},
{
"epoch": 1.2560386473429952,
"grad_norm": 0.654194712638855,
"learning_rate": 7.2272093467457226e-06,
"loss": 0.8192,
"step": 260
},
{
"epoch": 1.2608695652173914,
"grad_norm": 0.6488723754882812,
"learning_rate": 7.201970757788172e-06,
"loss": 0.8196,
"step": 261
},
{
"epoch": 1.2657004830917875,
"grad_norm": 0.6150570511817932,
"learning_rate": 7.17666237102845e-06,
"loss": 0.8084,
"step": 262
},
{
"epoch": 1.2705314009661834,
"grad_norm": 0.5813282132148743,
"learning_rate": 7.151284988688731e-06,
"loss": 0.7909,
"step": 263
},
{
"epoch": 1.2753623188405796,
"grad_norm": 0.6258924603462219,
"learning_rate": 7.125839415178204e-06,
"loss": 0.8144,
"step": 264
},
{
"epoch": 1.2801932367149758,
"grad_norm": 0.5735957622528076,
"learning_rate": 7.100326457067576e-06,
"loss": 0.7568,
"step": 265
},
{
"epoch": 1.285024154589372,
"grad_norm": 0.6329940557479858,
"learning_rate": 7.074746923063497e-06,
"loss": 0.7718,
"step": 266
},
{
"epoch": 1.289855072463768,
"grad_norm": 0.5378438830375671,
"learning_rate": 7.049101623982938e-06,
"loss": 0.8011,
"step": 267
},
{
"epoch": 1.2946859903381642,
"grad_norm": 0.6654214859008789,
"learning_rate": 7.02339137272748e-06,
"loss": 0.8156,
"step": 268
},
{
"epoch": 1.2995169082125604,
"grad_norm": 0.7466998100280762,
"learning_rate": 6.9976169842575526e-06,
"loss": 0.7883,
"step": 269
},
{
"epoch": 1.3043478260869565,
"grad_norm": 0.527633547782898,
"learning_rate": 6.971779275566593e-06,
"loss": 0.7967,
"step": 270
},
{
"epoch": 1.3091787439613527,
"grad_norm": 0.7060922384262085,
"learning_rate": 6.945879065655164e-06,
"loss": 0.8033,
"step": 271
},
{
"epoch": 1.3140096618357489,
"grad_norm": 0.6210879683494568,
"learning_rate": 6.919917175504978e-06,
"loss": 0.7613,
"step": 272
},
{
"epoch": 1.318840579710145,
"grad_norm": 0.5505712032318115,
"learning_rate": 6.893894428052881e-06,
"loss": 0.7572,
"step": 273
},
{
"epoch": 1.3236714975845412,
"grad_norm": 0.5760694742202759,
"learning_rate": 6.867811648164769e-06,
"loss": 0.794,
"step": 274
},
{
"epoch": 1.3285024154589373,
"grad_norm": 0.5254794955253601,
"learning_rate": 6.841669662609437e-06,
"loss": 0.7707,
"step": 275
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.5995458960533142,
"learning_rate": 6.815469300032374e-06,
"loss": 0.7778,
"step": 276
},
{
"epoch": 1.3381642512077294,
"grad_norm": 0.6530429124832153,
"learning_rate": 6.789211390929497e-06,
"loss": 0.7817,
"step": 277
},
{
"epoch": 1.3429951690821256,
"grad_norm": 0.6288939118385315,
"learning_rate": 6.762896767620827e-06,
"loss": 0.8001,
"step": 278
},
{
"epoch": 1.3478260869565217,
"grad_norm": 0.5950453281402588,
"learning_rate": 6.736526264224101e-06,
"loss": 0.8142,
"step": 279
},
{
"epoch": 1.3526570048309179,
"grad_norm": 0.6037265062332153,
"learning_rate": 6.710100716628345e-06,
"loss": 0.7762,
"step": 280
},
{
"epoch": 1.357487922705314,
"grad_norm": 0.784972071647644,
"learning_rate": 6.6836209624673575e-06,
"loss": 0.7972,
"step": 281
},
{
"epoch": 1.3623188405797102,
"grad_norm": 0.6714766621589661,
"learning_rate": 6.657087841093179e-06,
"loss": 0.7919,
"step": 282
},
{
"epoch": 1.3671497584541064,
"grad_norm": 0.6096466183662415,
"learning_rate": 6.6305021935494755e-06,
"loss": 0.79,
"step": 283
},
{
"epoch": 1.3719806763285023,
"grad_norm": 0.5656675696372986,
"learning_rate": 6.603864862544879e-06,
"loss": 0.7857,
"step": 284
},
{
"epoch": 1.3768115942028984,
"grad_norm": 0.7064409255981445,
"learning_rate": 6.5771766924262795e-06,
"loss": 0.8041,
"step": 285
},
{
"epoch": 1.3816425120772946,
"grad_norm": 0.6364306211471558,
"learning_rate": 6.5504385291520554e-06,
"loss": 0.8144,
"step": 286
},
{
"epoch": 1.3864734299516908,
"grad_norm": 0.5433293581008911,
"learning_rate": 6.523651220265269e-06,
"loss": 0.8049,
"step": 287
},
{
"epoch": 1.391304347826087,
"grad_norm": 0.6152275204658508,
"learning_rate": 6.496815614866792e-06,
"loss": 0.7886,
"step": 288
},
{
"epoch": 1.396135265700483,
"grad_norm": 0.5549913644790649,
"learning_rate": 6.469932563588386e-06,
"loss": 0.7809,
"step": 289
},
{
"epoch": 1.4009661835748792,
"grad_norm": 0.48793748021125793,
"learning_rate": 6.443002918565754e-06,
"loss": 0.7887,
"step": 290
},
{
"epoch": 1.4057971014492754,
"grad_norm": 0.6952255964279175,
"learning_rate": 6.41602753341152e-06,
"loss": 0.7703,
"step": 291
},
{
"epoch": 1.4106280193236715,
"grad_norm": 0.5199931859970093,
"learning_rate": 6.389007263188176e-06,
"loss": 0.8129,
"step": 292
},
{
"epoch": 1.4154589371980677,
"grad_norm": 0.5502813458442688,
"learning_rate": 6.361942964380967e-06,
"loss": 0.814,
"step": 293
},
{
"epoch": 1.4202898550724639,
"grad_norm": 0.6408051252365112,
"learning_rate": 6.334835494870759e-06,
"loss": 0.8181,
"step": 294
},
{
"epoch": 1.42512077294686,
"grad_norm": 0.6277318596839905,
"learning_rate": 6.307685713906835e-06,
"loss": 0.8188,
"step": 295
},
{
"epoch": 1.4299516908212562,
"grad_norm": 0.513602614402771,
"learning_rate": 6.2804944820796596e-06,
"loss": 0.8033,
"step": 296
},
{
"epoch": 1.434782608695652,
"grad_norm": 0.6088364124298096,
"learning_rate": 6.2532626612936035e-06,
"loss": 0.7887,
"step": 297
},
{
"epoch": 1.4396135265700483,
"grad_norm": 0.582086980342865,
"learning_rate": 6.225991114739622e-06,
"loss": 0.8111,
"step": 298
},
{
"epoch": 1.4444444444444444,
"grad_norm": 0.5153341889381409,
"learning_rate": 6.1986807068678926e-06,
"loss": 0.8097,
"step": 299
},
{
"epoch": 1.4492753623188406,
"grad_norm": 0.5599061846733093,
"learning_rate": 6.171332303360411e-06,
"loss": 0.7631,
"step": 300
},
{
"epoch": 1.4541062801932367,
"grad_norm": 0.5631583333015442,
"learning_rate": 6.143946771103561e-06,
"loss": 0.7744,
"step": 301
},
{
"epoch": 1.458937198067633,
"grad_norm": 0.6205547451972961,
"learning_rate": 6.11652497816062e-06,
"loss": 0.8067,
"step": 302
},
{
"epoch": 1.463768115942029,
"grad_norm": 0.5770376920700073,
"learning_rate": 6.089067793744258e-06,
"loss": 0.8072,
"step": 303
},
{
"epoch": 1.4685990338164252,
"grad_norm": 0.5772625803947449,
"learning_rate": 6.061576088188981e-06,
"loss": 0.7799,
"step": 304
},
{
"epoch": 1.4734299516908211,
"grad_norm": 0.6021287441253662,
"learning_rate": 6.034050732923538e-06,
"loss": 0.8046,
"step": 305
},
{
"epoch": 1.4782608695652173,
"grad_norm": 0.5101819634437561,
"learning_rate": 6.006492600443301e-06,
"loss": 0.7904,
"step": 306
},
{
"epoch": 1.4830917874396135,
"grad_norm": 0.5455768704414368,
"learning_rate": 5.978902564282616e-06,
"loss": 0.8112,
"step": 307
},
{
"epoch": 1.4879227053140096,
"grad_norm": 0.5585688352584839,
"learning_rate": 5.951281498987106e-06,
"loss": 0.778,
"step": 308
},
{
"epoch": 1.4927536231884058,
"grad_norm": 0.5855573415756226,
"learning_rate": 5.923630280085948e-06,
"loss": 0.7856,
"step": 309
},
{
"epoch": 1.497584541062802,
"grad_norm": 0.6162655353546143,
"learning_rate": 5.895949784064126e-06,
"loss": 0.7649,
"step": 310
},
{
"epoch": 1.502415458937198,
"grad_norm": 0.5619444847106934,
"learning_rate": 5.8682408883346535e-06,
"loss": 0.7734,
"step": 311
},
{
"epoch": 1.5072463768115942,
"grad_norm": 0.6392289400100708,
"learning_rate": 5.840504471210742e-06,
"loss": 0.8006,
"step": 312
},
{
"epoch": 1.5120772946859904,
"grad_norm": 0.5445345044136047,
"learning_rate": 5.8127414118779825e-06,
"loss": 0.807,
"step": 313
},
{
"epoch": 1.5169082125603865,
"grad_norm": 0.5203850865364075,
"learning_rate": 5.7849525903664636e-06,
"loss": 0.8264,
"step": 314
},
{
"epoch": 1.5217391304347827,
"grad_norm": 0.5838399529457092,
"learning_rate": 5.757138887522884e-06,
"loss": 0.8093,
"step": 315
},
{
"epoch": 1.5265700483091789,
"grad_norm": 0.4624084234237671,
"learning_rate": 5.729301184982622e-06,
"loss": 0.8046,
"step": 316
},
{
"epoch": 1.531400966183575,
"grad_norm": 0.5060942769050598,
"learning_rate": 5.701440365141799e-06,
"loss": 0.805,
"step": 317
},
{
"epoch": 1.5362318840579712,
"grad_norm": 0.5092195272445679,
"learning_rate": 5.673557311129306e-06,
"loss": 0.778,
"step": 318
},
{
"epoch": 1.541062801932367,
"grad_norm": 0.6425656080245972,
"learning_rate": 5.645652906778808e-06,
"loss": 0.7599,
"step": 319
},
{
"epoch": 1.5458937198067633,
"grad_norm": 0.6229380965232849,
"learning_rate": 5.617728036600734e-06,
"loss": 0.8282,
"step": 320
},
{
"epoch": 1.5507246376811594,
"grad_norm": 0.6080556511878967,
"learning_rate": 5.5897835857542315e-06,
"loss": 0.7854,
"step": 321
},
{
"epoch": 1.5555555555555556,
"grad_norm": 0.5633357167243958,
"learning_rate": 5.561820440019117e-06,
"loss": 0.7987,
"step": 322
},
{
"epoch": 1.5603864734299517,
"grad_norm": 0.62427818775177,
"learning_rate": 5.533839485767795e-06,
"loss": 0.8051,
"step": 323
},
{
"epoch": 1.5652173913043477,
"grad_norm": 0.5080029964447021,
"learning_rate": 5.505841609937162e-06,
"loss": 0.8136,
"step": 324
},
{
"epoch": 1.5700483091787438,
"grad_norm": 0.5406042337417603,
"learning_rate": 5.477827700000492e-06,
"loss": 0.8322,
"step": 325
},
{
"epoch": 1.57487922705314,
"grad_norm": 0.6476158499717712,
"learning_rate": 5.449798643939305e-06,
"loss": 0.8038,
"step": 326
},
{
"epoch": 1.5797101449275361,
"grad_norm": 0.6692044734954834,
"learning_rate": 5.421755330215223e-06,
"loss": 0.8115,
"step": 327
},
{
"epoch": 1.5845410628019323,
"grad_norm": 0.4540793299674988,
"learning_rate": 5.39369864774181e-06,
"loss": 0.7707,
"step": 328
},
{
"epoch": 1.5893719806763285,
"grad_norm": 0.683095395565033,
"learning_rate": 5.365629485856381e-06,
"loss": 0.8036,
"step": 329
},
{
"epoch": 1.5942028985507246,
"grad_norm": 0.6176734566688538,
"learning_rate": 5.337548734291827e-06,
"loss": 0.784,
"step": 330
},
{
"epoch": 1.5990338164251208,
"grad_norm": 0.5553250908851624,
"learning_rate": 5.30945728314841e-06,
"loss": 0.766,
"step": 331
},
{
"epoch": 1.603864734299517,
"grad_norm": 0.5356951951980591,
"learning_rate": 5.281356022865542e-06,
"loss": 0.7747,
"step": 332
},
{
"epoch": 1.608695652173913,
"grad_norm": 0.5851414203643799,
"learning_rate": 5.253245844193564e-06,
"loss": 0.7913,
"step": 333
},
{
"epoch": 1.6135265700483092,
"grad_norm": 0.4888221025466919,
"learning_rate": 5.225127638165514e-06,
"loss": 0.8018,
"step": 334
},
{
"epoch": 1.6183574879227054,
"grad_norm": 0.526310384273529,
"learning_rate": 5.197002296068878e-06,
"loss": 0.7714,
"step": 335
},
{
"epoch": 1.6231884057971016,
"grad_norm": 0.5102785229682922,
"learning_rate": 5.168870709417342e-06,
"loss": 0.7642,
"step": 336
},
{
"epoch": 1.6280193236714977,
"grad_norm": 0.602107048034668,
"learning_rate": 5.140733769922525e-06,
"loss": 0.755,
"step": 337
},
{
"epoch": 1.6328502415458939,
"grad_norm": 0.6300762295722961,
"learning_rate": 5.112592369465731e-06,
"loss": 0.8213,
"step": 338
},
{
"epoch": 1.6376811594202898,
"grad_norm": 0.5401408076286316,
"learning_rate": 5.084447400069656e-06,
"loss": 0.8322,
"step": 339
},
{
"epoch": 1.642512077294686,
"grad_norm": 0.6203901767730713,
"learning_rate": 5.0562997538701295e-06,
"loss": 0.7944,
"step": 340
},
{
"epoch": 1.6473429951690821,
"grad_norm": 0.6917441487312317,
"learning_rate": 5.0281503230878304e-06,
"loss": 0.7778,
"step": 341
},
{
"epoch": 1.6521739130434783,
"grad_norm": 0.4909766614437103,
"learning_rate": 5e-06,
"loss": 0.7992,
"step": 342
},
{
"epoch": 1.6570048309178744,
"grad_norm": 0.6045981049537659,
"learning_rate": 4.971849676912172e-06,
"loss": 0.8009,
"step": 343
},
{
"epoch": 1.6618357487922706,
"grad_norm": 0.7382987141609192,
"learning_rate": 4.943700246129871e-06,
"loss": 0.7973,
"step": 344
},
{
"epoch": 1.6666666666666665,
"grad_norm": 0.6006978154182434,
"learning_rate": 4.915552599930345e-06,
"loss": 0.7998,
"step": 345
},
{
"epoch": 1.6714975845410627,
"grad_norm": 0.6009412407875061,
"learning_rate": 4.887407630534271e-06,
"loss": 0.8081,
"step": 346
},
{
"epoch": 1.6763285024154588,
"grad_norm": 0.592892050743103,
"learning_rate": 4.859266230077474e-06,
"loss": 0.7681,
"step": 347
},
{
"epoch": 1.681159420289855,
"grad_norm": 0.6504932641983032,
"learning_rate": 4.83112929058266e-06,
"loss": 0.7812,
"step": 348
},
{
"epoch": 1.6859903381642511,
"grad_norm": 0.7033050656318665,
"learning_rate": 4.802997703931124e-06,
"loss": 0.8025,
"step": 349
},
{
"epoch": 1.6908212560386473,
"grad_norm": 0.6602729558944702,
"learning_rate": 4.7748723618344865e-06,
"loss": 0.7681,
"step": 350
},
{
"epoch": 1.6956521739130435,
"grad_norm": 0.5453357696533203,
"learning_rate": 4.746754155806437e-06,
"loss": 0.7494,
"step": 351
},
{
"epoch": 1.7004830917874396,
"grad_norm": 0.5276068449020386,
"learning_rate": 4.71864397713446e-06,
"loss": 0.7774,
"step": 352
},
{
"epoch": 1.7053140096618358,
"grad_norm": 0.6807137131690979,
"learning_rate": 4.6905427168515914e-06,
"loss": 0.7866,
"step": 353
},
{
"epoch": 1.710144927536232,
"grad_norm": 0.7194874882698059,
"learning_rate": 4.662451265708174e-06,
"loss": 0.7877,
"step": 354
},
{
"epoch": 1.714975845410628,
"grad_norm": 0.5802371501922607,
"learning_rate": 4.63437051414362e-06,
"loss": 0.7838,
"step": 355
},
{
"epoch": 1.7198067632850242,
"grad_norm": 0.5806293487548828,
"learning_rate": 4.606301352258192e-06,
"loss": 0.7698,
"step": 356
},
{
"epoch": 1.7246376811594204,
"grad_norm": 0.6043258309364319,
"learning_rate": 4.5782446697847775e-06,
"loss": 0.7921,
"step": 357
},
{
"epoch": 1.7294685990338166,
"grad_norm": 0.6771955490112305,
"learning_rate": 4.550201356060695e-06,
"loss": 0.7832,
"step": 358
},
{
"epoch": 1.7342995169082127,
"grad_norm": 0.5330072045326233,
"learning_rate": 4.52217229999951e-06,
"loss": 0.7826,
"step": 359
},
{
"epoch": 1.7391304347826086,
"grad_norm": 0.5096263885498047,
"learning_rate": 4.49415839006284e-06,
"loss": 0.7976,
"step": 360
},
{
"epoch": 1.7439613526570048,
"grad_norm": 0.6540196537971497,
"learning_rate": 4.466160514232206e-06,
"loss": 0.8119,
"step": 361
},
{
"epoch": 1.748792270531401,
"grad_norm": 0.6180545687675476,
"learning_rate": 4.438179559980885e-06,
"loss": 0.8178,
"step": 362
},
{
"epoch": 1.7536231884057971,
"grad_norm": 0.4621651768684387,
"learning_rate": 4.410216414245771e-06,
"loss": 0.7464,
"step": 363
},
{
"epoch": 1.7584541062801933,
"grad_norm": 0.543982744216919,
"learning_rate": 4.382271963399268e-06,
"loss": 0.8122,
"step": 364
},
{
"epoch": 1.7632850241545892,
"grad_norm": 0.5957825779914856,
"learning_rate": 4.354347093221194e-06,
"loss": 0.7915,
"step": 365
},
{
"epoch": 1.7681159420289854,
"grad_norm": 0.5544920563697815,
"learning_rate": 4.326442688870697e-06,
"loss": 0.7982,
"step": 366
},
{
"epoch": 1.7729468599033815,
"grad_norm": 0.4633856415748596,
"learning_rate": 4.298559634858202e-06,
"loss": 0.8165,
"step": 367
},
{
"epoch": 1.7777777777777777,
"grad_norm": 0.538975179195404,
"learning_rate": 4.270698815017379e-06,
"loss": 0.7815,
"step": 368
},
{
"epoch": 1.7826086956521738,
"grad_norm": 0.5998566746711731,
"learning_rate": 4.2428611124771184e-06,
"loss": 0.8045,
"step": 369
},
{
"epoch": 1.78743961352657,
"grad_norm": 0.5452315211296082,
"learning_rate": 4.2150474096335356e-06,
"loss": 0.7713,
"step": 370
},
{
"epoch": 1.7922705314009661,
"grad_norm": 0.47987136244773865,
"learning_rate": 4.187258588122019e-06,
"loss": 0.772,
"step": 371
},
{
"epoch": 1.7971014492753623,
"grad_norm": 0.5281115174293518,
"learning_rate": 4.15949552878926e-06,
"loss": 0.7981,
"step": 372
},
{
"epoch": 1.8019323671497585,
"grad_norm": 0.49062639474868774,
"learning_rate": 4.131759111665349e-06,
"loss": 0.763,
"step": 373
},
{
"epoch": 1.8067632850241546,
"grad_norm": 0.512249231338501,
"learning_rate": 4.104050215935875e-06,
"loss": 0.7778,
"step": 374
},
{
"epoch": 1.8115942028985508,
"grad_norm": 0.5328919887542725,
"learning_rate": 4.076369719914055e-06,
"loss": 0.7979,
"step": 375
},
{
"epoch": 1.816425120772947,
"grad_norm": 0.5143810510635376,
"learning_rate": 4.048718501012895e-06,
"loss": 0.7616,
"step": 376
},
{
"epoch": 1.821256038647343,
"grad_norm": 0.5369464159011841,
"learning_rate": 4.021097435717386e-06,
"loss": 0.7626,
"step": 377
},
{
"epoch": 1.8260869565217392,
"grad_norm": 0.5135899186134338,
"learning_rate": 3.993507399556699e-06,
"loss": 0.7478,
"step": 378
},
{
"epoch": 1.8309178743961354,
"grad_norm": 0.5900195240974426,
"learning_rate": 3.965949267076465e-06,
"loss": 0.8072,
"step": 379
},
{
"epoch": 1.8357487922705316,
"grad_norm": 0.6226658821105957,
"learning_rate": 3.938423911811021e-06,
"loss": 0.7713,
"step": 380
},
{
"epoch": 1.8405797101449275,
"grad_norm": 0.5240309238433838,
"learning_rate": 3.910932206255742e-06,
"loss": 0.7827,
"step": 381
},
{
"epoch": 1.8454106280193237,
"grad_norm": 0.5698277354240417,
"learning_rate": 3.883475021839382e-06,
"loss": 0.7783,
"step": 382
},
{
"epoch": 1.8502415458937198,
"grad_norm": 0.6924400925636292,
"learning_rate": 3.856053228896442e-06,
"loss": 0.7967,
"step": 383
},
{
"epoch": 1.855072463768116,
"grad_norm": 0.5029117465019226,
"learning_rate": 3.8286676966395895e-06,
"loss": 0.7897,
"step": 384
},
{
"epoch": 1.8599033816425121,
"grad_norm": 0.5228816866874695,
"learning_rate": 3.8013192931321095e-06,
"loss": 0.811,
"step": 385
},
{
"epoch": 1.864734299516908,
"grad_norm": 0.554703414440155,
"learning_rate": 3.77400888526038e-06,
"loss": 0.8261,
"step": 386
},
{
"epoch": 1.8695652173913042,
"grad_norm": 0.5827821493148804,
"learning_rate": 3.7467373387063973e-06,
"loss": 0.7898,
"step": 387
},
{
"epoch": 1.8743961352657004,
"grad_norm": 0.5374961495399475,
"learning_rate": 3.719505517920342e-06,
"loss": 0.8126,
"step": 388
},
{
"epoch": 1.8792270531400965,
"grad_norm": 0.5197350382804871,
"learning_rate": 3.692314286093167e-06,
"loss": 0.7839,
"step": 389
},
{
"epoch": 1.8840579710144927,
"grad_norm": 0.6360486149787903,
"learning_rate": 3.6651645051292415e-06,
"loss": 0.7879,
"step": 390
},
{
"epoch": 1.8888888888888888,
"grad_norm": 0.4924606680870056,
"learning_rate": 3.6380570356190346e-06,
"loss": 0.7753,
"step": 391
},
{
"epoch": 1.893719806763285,
"grad_norm": 0.5175696015357971,
"learning_rate": 3.610992736811827e-06,
"loss": 0.8057,
"step": 392
},
{
"epoch": 1.8985507246376812,
"grad_norm": 0.5930569171905518,
"learning_rate": 3.58397246658848e-06,
"loss": 0.7971,
"step": 393
},
{
"epoch": 1.9033816425120773,
"grad_norm": 0.4917478561401367,
"learning_rate": 3.556997081434248e-06,
"loss": 0.7796,
"step": 394
},
{
"epoch": 1.9082125603864735,
"grad_norm": 0.5090054869651794,
"learning_rate": 3.5300674364116173e-06,
"loss": 0.785,
"step": 395
},
{
"epoch": 1.9130434782608696,
"grad_norm": 0.5114875435829163,
"learning_rate": 3.5031843851332105e-06,
"loss": 0.805,
"step": 396
},
{
"epoch": 1.9178743961352658,
"grad_norm": 0.5346699953079224,
"learning_rate": 3.476348779734732e-06,
"loss": 0.8004,
"step": 397
},
{
"epoch": 1.922705314009662,
"grad_norm": 0.6212336421012878,
"learning_rate": 3.449561470847947e-06,
"loss": 0.8092,
"step": 398
},
{
"epoch": 1.927536231884058,
"grad_norm": 0.5195980668067932,
"learning_rate": 3.4228233075737225e-06,
"loss": 0.7694,
"step": 399
},
{
"epoch": 1.9323671497584543,
"grad_norm": 0.5516073703765869,
"learning_rate": 3.3961351374551234e-06,
"loss": 0.8211,
"step": 400
},
{
"epoch": 1.9371980676328504,
"grad_norm": 0.5031560659408569,
"learning_rate": 3.3694978064505258e-06,
"loss": 0.808,
"step": 401
},
{
"epoch": 1.9420289855072463,
"grad_norm": 0.6618154644966125,
"learning_rate": 3.3429121589068213e-06,
"loss": 0.814,
"step": 402
},
{
"epoch": 1.9468599033816425,
"grad_norm": 0.5210357904434204,
"learning_rate": 3.316379037532644e-06,
"loss": 0.8077,
"step": 403
},
{
"epoch": 1.9516908212560387,
"grad_norm": 0.5135445594787598,
"learning_rate": 3.289899283371657e-06,
"loss": 0.7805,
"step": 404
},
{
"epoch": 1.9565217391304348,
"grad_norm": 0.5435628294944763,
"learning_rate": 3.2634737357758994e-06,
"loss": 0.7839,
"step": 405
},
{
"epoch": 1.961352657004831,
"grad_norm": 0.5716672539710999,
"learning_rate": 3.2371032323791757e-06,
"loss": 0.7948,
"step": 406
},
{
"epoch": 1.966183574879227,
"grad_norm": 0.5514191389083862,
"learning_rate": 3.2107886090705035e-06,
"loss": 0.7613,
"step": 407
},
{
"epoch": 1.971014492753623,
"grad_norm": 0.5380987524986267,
"learning_rate": 3.1845306999676274e-06,
"loss": 0.7799,
"step": 408
},
{
"epoch": 1.9758454106280192,
"grad_norm": 0.5445787906646729,
"learning_rate": 3.158330337390565e-06,
"loss": 0.7906,
"step": 409
},
{
"epoch": 1.9806763285024154,
"grad_norm": 0.5519409775733948,
"learning_rate": 3.132188351835232e-06,
"loss": 0.7932,
"step": 410
},
{
"epoch": 1.9855072463768115,
"grad_norm": 0.6107297539710999,
"learning_rate": 3.10610557194712e-06,
"loss": 0.8045,
"step": 411
},
{
"epoch": 1.9903381642512077,
"grad_norm": 0.5743387937545776,
"learning_rate": 3.080082824495024e-06,
"loss": 0.8321,
"step": 412
},
{
"epoch": 1.9951690821256038,
"grad_norm": 0.5688542723655701,
"learning_rate": 3.0541209343448373e-06,
"loss": 0.7812,
"step": 413
},
{
"epoch": 2.0,
"grad_norm": 0.5131503343582153,
"learning_rate": 3.0282207244334084e-06,
"loss": 0.772,
"step": 414
},
{
"epoch": 2.004830917874396,
"grad_norm": 0.5330387949943542,
"learning_rate": 3.0023830157424504e-06,
"loss": 0.7047,
"step": 415
},
{
"epoch": 2.0096618357487923,
"grad_norm": 0.6632731556892395,
"learning_rate": 2.97660862727252e-06,
"loss": 0.7167,
"step": 416
},
{
"epoch": 2.0144927536231885,
"grad_norm": 0.529600203037262,
"learning_rate": 2.950898376017064e-06,
"loss": 0.7295,
"step": 417
},
{
"epoch": 2.0193236714975846,
"grad_norm": 0.6404242515563965,
"learning_rate": 2.9252530769365053e-06,
"loss": 0.7641,
"step": 418
},
{
"epoch": 2.024154589371981,
"grad_norm": 0.5049023628234863,
"learning_rate": 2.8996735429324256e-06,
"loss": 0.7486,
"step": 419
},
{
"epoch": 2.028985507246377,
"grad_norm": 0.5269747376441956,
"learning_rate": 2.874160584821798e-06,
"loss": 0.784,
"step": 420
},
{
"epoch": 2.033816425120773,
"grad_norm": 0.5848202109336853,
"learning_rate": 2.848715011311271e-06,
"loss": 0.7328,
"step": 421
},
{
"epoch": 2.0386473429951693,
"grad_norm": 0.6017028093338013,
"learning_rate": 2.823337628971551e-06,
"loss": 0.81,
"step": 422
},
{
"epoch": 2.0434782608695654,
"grad_norm": 0.5388191938400269,
"learning_rate": 2.7980292422118282e-06,
"loss": 0.7456,
"step": 423
},
{
"epoch": 2.0483091787439616,
"grad_norm": 0.4810311496257782,
"learning_rate": 2.7727906532542783e-06,
"loss": 0.7718,
"step": 424
},
{
"epoch": 2.0531400966183573,
"grad_norm": 0.519432008266449,
"learning_rate": 2.7476226621086354e-06,
"loss": 0.7562,
"step": 425
},
{
"epoch": 2.0579710144927534,
"grad_norm": 0.4953659474849701,
"learning_rate": 2.72252606654683e-06,
"loss": 0.7303,
"step": 426
},
{
"epoch": 2.0628019323671496,
"grad_norm": 0.5011581182479858,
"learning_rate": 2.697501662077707e-06,
"loss": 0.7589,
"step": 427
},
{
"epoch": 2.0676328502415457,
"grad_norm": 0.585135817527771,
"learning_rate": 2.6725502419218084e-06,
"loss": 0.7573,
"step": 428
},
{
"epoch": 2.072463768115942,
"grad_norm": 0.5075230002403259,
"learning_rate": 2.6476725969862227e-06,
"loss": 0.7592,
"step": 429
},
{
"epoch": 2.077294685990338,
"grad_norm": 0.45230966806411743,
"learning_rate": 2.622869515839524e-06,
"loss": 0.7663,
"step": 430
},
{
"epoch": 2.082125603864734,
"grad_norm": 0.4765617549419403,
"learning_rate": 2.5981417846867753e-06,
"loss": 0.726,
"step": 431
},
{
"epoch": 2.0869565217391304,
"grad_norm": 0.5006648898124695,
"learning_rate": 2.573490187344596e-06,
"loss": 0.7582,
"step": 432
},
{
"epoch": 2.0917874396135265,
"grad_norm": 0.48479005694389343,
"learning_rate": 2.548915505216333e-06,
"loss": 0.7271,
"step": 433
},
{
"epoch": 2.0966183574879227,
"grad_norm": 0.4033167064189911,
"learning_rate": 2.524418517267283e-06,
"loss": 0.7936,
"step": 434
},
{
"epoch": 2.101449275362319,
"grad_norm": 0.452640563249588,
"learning_rate": 2.5000000000000015e-06,
"loss": 0.7356,
"step": 435
},
{
"epoch": 2.106280193236715,
"grad_norm": 0.4914138913154602,
"learning_rate": 2.4756607274296844e-06,
"loss": 0.7526,
"step": 436
},
{
"epoch": 2.111111111111111,
"grad_norm": 0.4990180432796478,
"learning_rate": 2.4514014710596467e-06,
"loss": 0.7147,
"step": 437
},
{
"epoch": 2.1159420289855073,
"grad_norm": 0.46856606006622314,
"learning_rate": 2.4272229998568576e-06,
"loss": 0.7559,
"step": 438
},
{
"epoch": 2.1207729468599035,
"grad_norm": 0.48758986592292786,
"learning_rate": 2.4031260802275623e-06,
"loss": 0.7652,
"step": 439
},
{
"epoch": 2.1256038647342996,
"grad_norm": 0.4734245538711548,
"learning_rate": 2.3791114759930013e-06,
"loss": 0.7335,
"step": 440
},
{
"epoch": 2.130434782608696,
"grad_norm": 0.45640215277671814,
"learning_rate": 2.3551799483651894e-06,
"loss": 0.7535,
"step": 441
},
{
"epoch": 2.135265700483092,
"grad_norm": 0.4912705719470978,
"learning_rate": 2.331332255922784e-06,
"loss": 0.7503,
"step": 442
},
{
"epoch": 2.140096618357488,
"grad_norm": 0.46668151021003723,
"learning_rate": 2.307569154587056e-06,
"loss": 0.7335,
"step": 443
},
{
"epoch": 2.1449275362318843,
"grad_norm": 0.5260794162750244,
"learning_rate": 2.283891397597908e-06,
"loss": 0.7048,
"step": 444
},
{
"epoch": 2.14975845410628,
"grad_norm": 0.46279335021972656,
"learning_rate": 2.2602997354900075e-06,
"loss": 0.7288,
"step": 445
},
{
"epoch": 2.154589371980676,
"grad_norm": 0.464158296585083,
"learning_rate": 2.236794916069007e-06,
"loss": 0.7249,
"step": 446
},
{
"epoch": 2.1594202898550723,
"grad_norm": 0.46745479106903076,
"learning_rate": 2.2133776843878185e-06,
"loss": 0.7094,
"step": 447
},
{
"epoch": 2.1642512077294684,
"grad_norm": 0.4427666664123535,
"learning_rate": 2.190048782723015e-06,
"loss": 0.7349,
"step": 448
},
{
"epoch": 2.1690821256038646,
"grad_norm": 0.44838133454322815,
"learning_rate": 2.166808950551296e-06,
"loss": 0.7258,
"step": 449
},
{
"epoch": 2.1739130434782608,
"grad_norm": 0.49869880080223083,
"learning_rate": 2.1436589245260375e-06,
"loss": 0.7153,
"step": 450
},
{
"epoch": 2.178743961352657,
"grad_norm": 0.46117615699768066,
"learning_rate": 2.120599438453968e-06,
"loss": 0.7267,
"step": 451
},
{
"epoch": 2.183574879227053,
"grad_norm": 0.46450984477996826,
"learning_rate": 2.0976312232718763e-06,
"loss": 0.7158,
"step": 452
},
{
"epoch": 2.1884057971014492,
"grad_norm": 0.45363008975982666,
"learning_rate": 2.074755007023461e-06,
"loss": 0.7507,
"step": 453
},
{
"epoch": 2.1932367149758454,
"grad_norm": 0.48838719725608826,
"learning_rate": 2.0519715148362585e-06,
"loss": 0.7438,
"step": 454
},
{
"epoch": 2.1980676328502415,
"grad_norm": 0.5061821341514587,
"learning_rate": 2.0292814688986375e-06,
"loss": 0.7559,
"step": 455
},
{
"epoch": 2.2028985507246377,
"grad_norm": 0.4787469208240509,
"learning_rate": 2.0066855884369246e-06,
"loss": 0.7813,
"step": 456
},
{
"epoch": 2.207729468599034,
"grad_norm": 0.5006588697433472,
"learning_rate": 1.9841845896926022e-06,
"loss": 0.7688,
"step": 457
},
{
"epoch": 2.21256038647343,
"grad_norm": 0.4781847596168518,
"learning_rate": 1.961779185899597e-06,
"loss": 0.7436,
"step": 458
},
{
"epoch": 2.217391304347826,
"grad_norm": 0.4518663287162781,
"learning_rate": 1.9394700872616856e-06,
"loss": 0.736,
"step": 459
},
{
"epoch": 2.2222222222222223,
"grad_norm": 0.4903218746185303,
"learning_rate": 1.9172580009299735e-06,
"loss": 0.7169,
"step": 460
},
{
"epoch": 2.2270531400966185,
"grad_norm": 0.49644899368286133,
"learning_rate": 1.8951436309804766e-06,
"loss": 0.7392,
"step": 461
},
{
"epoch": 2.2318840579710146,
"grad_norm": 0.4554813802242279,
"learning_rate": 1.8731276783918162e-06,
"loss": 0.7329,
"step": 462
},
{
"epoch": 2.236714975845411,
"grad_norm": 0.42116937041282654,
"learning_rate": 1.8512108410229878e-06,
"loss": 0.7632,
"step": 463
},
{
"epoch": 2.241545893719807,
"grad_norm": 0.47068312764167786,
"learning_rate": 1.8293938135912475e-06,
"loss": 0.7631,
"step": 464
},
{
"epoch": 2.246376811594203,
"grad_norm": 0.4414040446281433,
"learning_rate": 1.8076772876500831e-06,
"loss": 0.7638,
"step": 465
},
{
"epoch": 2.2512077294685993,
"grad_norm": 0.44085627794265747,
"learning_rate": 1.7860619515673034e-06,
"loss": 0.7597,
"step": 466
},
{
"epoch": 2.2560386473429954,
"grad_norm": 0.41587820649147034,
"learning_rate": 1.7645484905032129e-06,
"loss": 0.723,
"step": 467
},
{
"epoch": 2.260869565217391,
"grad_norm": 0.4661046266555786,
"learning_rate": 1.74313758638889e-06,
"loss": 0.7493,
"step": 468
},
{
"epoch": 2.2657004830917873,
"grad_norm": 0.4823686182498932,
"learning_rate": 1.7218299179045789e-06,
"loss": 0.7505,
"step": 469
},
{
"epoch": 2.2705314009661834,
"grad_norm": 0.4889514744281769,
"learning_rate": 1.7006261604581725e-06,
"loss": 0.7372,
"step": 470
},
{
"epoch": 2.2753623188405796,
"grad_norm": 0.4301418960094452,
"learning_rate": 1.6795269861638041e-06,
"loss": 0.7674,
"step": 471
},
{
"epoch": 2.2801932367149758,
"grad_norm": 0.45111072063446045,
"learning_rate": 1.6585330638205454e-06,
"loss": 0.7547,
"step": 472
},
{
"epoch": 2.285024154589372,
"grad_norm": 0.46377280354499817,
"learning_rate": 1.6376450588911985e-06,
"loss": 0.7286,
"step": 473
},
{
"epoch": 2.289855072463768,
"grad_norm": 0.4372880458831787,
"learning_rate": 1.6168636334812126e-06,
"loss": 0.7425,
"step": 474
},
{
"epoch": 2.2946859903381642,
"grad_norm": 0.40830719470977783,
"learning_rate": 1.5961894463176942e-06,
"loss": 0.7568,
"step": 475
},
{
"epoch": 2.2995169082125604,
"grad_norm": 0.4759189784526825,
"learning_rate": 1.5756231527285181e-06,
"loss": 0.7517,
"step": 476
},
{
"epoch": 2.3043478260869565,
"grad_norm": 0.4460543990135193,
"learning_rate": 1.555165404621567e-06,
"loss": 0.7401,
"step": 477
},
{
"epoch": 2.3091787439613527,
"grad_norm": 0.43344846367836,
"learning_rate": 1.5348168504640631e-06,
"loss": 0.7509,
"step": 478
},
{
"epoch": 2.314009661835749,
"grad_norm": 0.4556834399700165,
"learning_rate": 1.5145781352620054e-06,
"loss": 0.7514,
"step": 479
},
{
"epoch": 2.318840579710145,
"grad_norm": 0.4389399290084839,
"learning_rate": 1.4944499005397372e-06,
"loss": 0.7228,
"step": 480
},
{
"epoch": 2.323671497584541,
"grad_norm": 0.4609167277812958,
"learning_rate": 1.4744327843196043e-06,
"loss": 0.7593,
"step": 481
},
{
"epoch": 2.3285024154589373,
"grad_norm": 0.494718462228775,
"learning_rate": 1.4545274211017264e-06,
"loss": 0.7569,
"step": 482
},
{
"epoch": 2.3333333333333335,
"grad_norm": 0.4857819974422455,
"learning_rate": 1.434734441843899e-06,
"loss": 0.7721,
"step": 483
},
{
"epoch": 2.3381642512077296,
"grad_norm": 0.4424082040786743,
"learning_rate": 1.4150544739415755e-06,
"loss": 0.7408,
"step": 484
},
{
"epoch": 2.342995169082126,
"grad_norm": 0.48782283067703247,
"learning_rate": 1.3954881412079945e-06,
"loss": 0.7648,
"step": 485
},
{
"epoch": 2.3478260869565215,
"grad_norm": 0.45818960666656494,
"learning_rate": 1.3760360638544012e-06,
"loss": 0.7492,
"step": 486
},
{
"epoch": 2.3526570048309177,
"grad_norm": 0.49244335293769836,
"learning_rate": 1.3566988584703817e-06,
"loss": 0.7357,
"step": 487
},
{
"epoch": 2.357487922705314,
"grad_norm": 0.4363764822483063,
"learning_rate": 1.3374771380043306e-06,
"loss": 0.7283,
"step": 488
},
{
"epoch": 2.36231884057971,
"grad_norm": 0.44438326358795166,
"learning_rate": 1.3183715117440143e-06,
"loss": 0.7493,
"step": 489
},
{
"epoch": 2.367149758454106,
"grad_norm": 0.43266430497169495,
"learning_rate": 1.2993825852972559e-06,
"loss": 0.7782,
"step": 490
},
{
"epoch": 2.3719806763285023,
"grad_norm": 0.4312337338924408,
"learning_rate": 1.280510960572745e-06,
"loss": 0.7685,
"step": 491
},
{
"epoch": 2.3768115942028984,
"grad_norm": 0.4180818498134613,
"learning_rate": 1.2617572357609565e-06,
"loss": 0.7557,
"step": 492
},
{
"epoch": 2.3816425120772946,
"grad_norm": 0.4240153431892395,
"learning_rate": 1.2431220053151832e-06,
"loss": 0.746,
"step": 493
},
{
"epoch": 2.3864734299516908,
"grad_norm": 0.47555652260780334,
"learning_rate": 1.2246058599327021e-06,
"loss": 0.7249,
"step": 494
},
{
"epoch": 2.391304347826087,
"grad_norm": 0.4910667836666107,
"learning_rate": 1.2062093865360458e-06,
"loss": 0.7812,
"step": 495
},
{
"epoch": 2.396135265700483,
"grad_norm": 0.44273853302001953,
"learning_rate": 1.1879331682543972e-06,
"loss": 0.7552,
"step": 496
},
{
"epoch": 2.4009661835748792,
"grad_norm": 0.4478420913219452,
"learning_rate": 1.1697777844051105e-06,
"loss": 0.7505,
"step": 497
},
{
"epoch": 2.4057971014492754,
"grad_norm": 0.4299091398715973,
"learning_rate": 1.1517438104753386e-06,
"loss": 0.7401,
"step": 498
},
{
"epoch": 2.4106280193236715,
"grad_norm": 0.4194658696651459,
"learning_rate": 1.1338318181038037e-06,
"loss": 0.7372,
"step": 499
},
{
"epoch": 2.4154589371980677,
"grad_norm": 0.44260597229003906,
"learning_rate": 1.1160423750626693e-06,
"loss": 0.7408,
"step": 500
},
{
"epoch": 2.420289855072464,
"grad_norm": 0.48269572854042053,
"learning_rate": 1.0983760452395415e-06,
"loss": 0.7874,
"step": 501
},
{
"epoch": 2.42512077294686,
"grad_norm": 0.4460936188697815,
"learning_rate": 1.0808333886196038e-06,
"loss": 0.7575,
"step": 502
},
{
"epoch": 2.429951690821256,
"grad_norm": 0.42938852310180664,
"learning_rate": 1.063414961267859e-06,
"loss": 0.7842,
"step": 503
},
{
"epoch": 2.4347826086956523,
"grad_norm": 0.4390149712562561,
"learning_rate": 1.046121315311508e-06,
"loss": 0.753,
"step": 504
},
{
"epoch": 2.4396135265700485,
"grad_norm": 0.42639121413230896,
"learning_rate": 1.02895299892244e-06,
"loss": 0.7587,
"step": 505
},
{
"epoch": 2.4444444444444446,
"grad_norm": 0.4300375282764435,
"learning_rate": 1.01191055629987e-06,
"loss": 0.7457,
"step": 506
},
{
"epoch": 2.449275362318841,
"grad_norm": 0.45614680647850037,
"learning_rate": 9.949945276530782e-07,
"loss": 0.7336,
"step": 507
},
{
"epoch": 2.454106280193237,
"grad_norm": 0.43812501430511475,
"learning_rate": 9.782054491842879e-07,
"loss": 0.7686,
"step": 508
},
{
"epoch": 2.4589371980676327,
"grad_norm": 0.4528813064098358,
"learning_rate": 9.615438530716753e-07,
"loss": 0.71,
"step": 509
},
{
"epoch": 2.463768115942029,
"grad_norm": 0.4434235394001007,
"learning_rate": 9.450102674524952e-07,
"loss": 0.6996,
"step": 510
},
{
"epoch": 2.468599033816425,
"grad_norm": 0.42420923709869385,
"learning_rate": 9.286052164063369e-07,
"loss": 0.7661,
"step": 511
},
{
"epoch": 2.473429951690821,
"grad_norm": 0.4244982600212097,
"learning_rate": 9.123292199385247e-07,
"loss": 0.7905,
"step": 512
},
{
"epoch": 2.4782608695652173,
"grad_norm": 0.45026880502700806,
"learning_rate": 8.961827939636198e-07,
"loss": 0.7306,
"step": 513
},
{
"epoch": 2.4830917874396135,
"grad_norm": 0.4042074978351593,
"learning_rate": 8.801664502890722e-07,
"loss": 0.757,
"step": 514
},
{
"epoch": 2.4879227053140096,
"grad_norm": 0.41660401225090027,
"learning_rate": 8.64280696599008e-07,
"loss": 0.7432,
"step": 515
},
{
"epoch": 2.4927536231884058,
"grad_norm": 0.42759013175964355,
"learning_rate": 8.485260364381187e-07,
"loss": 0.7641,
"step": 516
},
{
"epoch": 2.497584541062802,
"grad_norm": 0.40827131271362305,
"learning_rate": 8.329029691957124e-07,
"loss": 0.7504,
"step": 517
},
{
"epoch": 2.502415458937198,
"grad_norm": 0.41131073236465454,
"learning_rate": 8.17411990089883e-07,
"loss": 0.7909,
"step": 518
},
{
"epoch": 2.5072463768115942,
"grad_norm": 0.4378702640533447,
"learning_rate": 8.02053590151805e-07,
"loss": 0.753,
"step": 519
},
{
"epoch": 2.5120772946859904,
"grad_norm": 0.4602629542350769,
"learning_rate": 7.868282562101819e-07,
"loss": 0.7367,
"step": 520
},
{
"epoch": 2.5169082125603865,
"grad_norm": 0.4568740129470825,
"learning_rate": 7.717364708758024e-07,
"loss": 0.7604,
"step": 521
},
{
"epoch": 2.5217391304347827,
"grad_norm": 0.46792659163475037,
"learning_rate": 7.567787125262449e-07,
"loss": 0.7315,
"step": 522
},
{
"epoch": 2.526570048309179,
"grad_norm": 0.4383016526699066,
"learning_rate": 7.41955455290726e-07,
"loss": 0.7408,
"step": 523
},
{
"epoch": 2.531400966183575,
"grad_norm": 0.40000349283218384,
"learning_rate": 7.27267169035053e-07,
"loss": 0.732,
"step": 524
},
{
"epoch": 2.536231884057971,
"grad_norm": 0.3945160508155823,
"learning_rate": 7.127143193467445e-07,
"loss": 0.7751,
"step": 525
},
{
"epoch": 2.541062801932367,
"grad_norm": 0.4534020721912384,
"learning_rate": 6.982973675202676e-07,
"loss": 0.7625,
"step": 526
},
{
"epoch": 2.545893719806763,
"grad_norm": 0.3865754008293152,
"learning_rate": 6.840167705424106e-07,
"loss": 0.7565,
"step": 527
},
{
"epoch": 2.550724637681159,
"grad_norm": 0.441019207239151,
"learning_rate": 6.698729810778065e-07,
"loss": 0.7452,
"step": 528
},
{
"epoch": 2.5555555555555554,
"grad_norm": 0.42603588104248047,
"learning_rate": 6.558664474545817e-07,
"loss": 0.7297,
"step": 529
},
{
"epoch": 2.5603864734299515,
"grad_norm": 0.4300025999546051,
"learning_rate": 6.419976136501377e-07,
"loss": 0.7457,
"step": 530
},
{
"epoch": 2.5652173913043477,
"grad_norm": 0.48573917150497437,
"learning_rate": 6.282669192770896e-07,
"loss": 0.7646,
"step": 531
},
{
"epoch": 2.570048309178744,
"grad_norm": 0.4400321841239929,
"learning_rate": 6.146747995693225e-07,
"loss": 0.7085,
"step": 532
},
{
"epoch": 2.57487922705314,
"grad_norm": 0.4423598349094391,
"learning_rate": 6.012216853682001e-07,
"loss": 0.7478,
"step": 533
},
{
"epoch": 2.579710144927536,
"grad_norm": 0.46172380447387695,
"learning_rate": 5.879080031089047e-07,
"loss": 0.7339,
"step": 534
},
{
"epoch": 2.5845410628019323,
"grad_norm": 0.4211689233779907,
"learning_rate": 5.747341748069229e-07,
"loss": 0.774,
"step": 535
},
{
"epoch": 2.5893719806763285,
"grad_norm": 0.4327799379825592,
"learning_rate": 5.617006180446688e-07,
"loss": 0.7417,
"step": 536
},
{
"epoch": 2.5942028985507246,
"grad_norm": 0.4260788559913635,
"learning_rate": 5.488077459582425e-07,
"loss": 0.7593,
"step": 537
},
{
"epoch": 2.5990338164251208,
"grad_norm": 0.44230300188064575,
"learning_rate": 5.360559672243421e-07,
"loss": 0.7384,
"step": 538
},
{
"epoch": 2.603864734299517,
"grad_norm": 0.4378679394721985,
"learning_rate": 5.234456860473042e-07,
"loss": 0.7479,
"step": 539
},
{
"epoch": 2.608695652173913,
"grad_norm": 0.3903830647468567,
"learning_rate": 5.109773021462921e-07,
"loss": 0.7484,
"step": 540
},
{
"epoch": 2.6135265700483092,
"grad_norm": 0.4425307810306549,
"learning_rate": 4.986512107426283e-07,
"loss": 0.7575,
"step": 541
},
{
"epoch": 2.6183574879227054,
"grad_norm": 0.43397703766822815,
"learning_rate": 4.864678025472635e-07,
"loss": 0.7421,
"step": 542
},
{
"epoch": 2.6231884057971016,
"grad_norm": 0.4102035164833069,
"learning_rate": 4.7442746374839363e-07,
"loss": 0.7684,
"step": 543
},
{
"epoch": 2.6280193236714977,
"grad_norm": 0.4425765872001648,
"learning_rate": 4.625305759992205e-07,
"loss": 0.7295,
"step": 544
},
{
"epoch": 2.632850241545894,
"grad_norm": 0.38346314430236816,
"learning_rate": 4.50777516405847e-07,
"loss": 0.767,
"step": 545
},
{
"epoch": 2.63768115942029,
"grad_norm": 0.44841113686561584,
"learning_rate": 4.3916865751533313e-07,
"loss": 0.7601,
"step": 546
},
{
"epoch": 2.642512077294686,
"grad_norm": 0.42210566997528076,
"learning_rate": 4.2770436730388166e-07,
"loss": 0.7117,
"step": 547
},
{
"epoch": 2.6473429951690823,
"grad_norm": 0.40515097975730896,
"learning_rate": 4.163850091651717e-07,
"loss": 0.7199,
"step": 548
},
{
"epoch": 2.6521739130434785,
"grad_norm": 0.43395963311195374,
"learning_rate": 4.05210941898847e-07,
"loss": 0.7549,
"step": 549
},
{
"epoch": 2.6570048309178746,
"grad_norm": 0.41985607147216797,
"learning_rate": 3.941825196991378e-07,
"loss": 0.7058,
"step": 550
},
{
"epoch": 2.661835748792271,
"grad_norm": 0.44412142038345337,
"learning_rate": 3.8330009214363197e-07,
"loss": 0.7635,
"step": 551
},
{
"epoch": 2.6666666666666665,
"grad_norm": 0.44575539231300354,
"learning_rate": 3.725640041822026e-07,
"loss": 0.688,
"step": 552
},
{
"epoch": 2.6714975845410627,
"grad_norm": 0.42022770643234253,
"learning_rate": 3.619745961260623e-07,
"loss": 0.776,
"step": 553
},
{
"epoch": 2.676328502415459,
"grad_norm": 0.4149934947490692,
"learning_rate": 3.5153220363698225e-07,
"loss": 0.7615,
"step": 554
},
{
"epoch": 2.681159420289855,
"grad_norm": 0.4015776216983795,
"learning_rate": 3.4123715771665786e-07,
"loss": 0.774,
"step": 555
},
{
"epoch": 2.685990338164251,
"grad_norm": 0.4558965265750885,
"learning_rate": 3.310897846962041e-07,
"loss": 0.6754,
"step": 556
},
{
"epoch": 2.6908212560386473,
"grad_norm": 0.44629231095314026,
"learning_rate": 3.2109040622582186e-07,
"loss": 0.7403,
"step": 557
},
{
"epoch": 2.6956521739130435,
"grad_norm": 0.42704764008522034,
"learning_rate": 3.112393392645985e-07,
"loss": 0.7617,
"step": 558
},
{
"epoch": 2.7004830917874396,
"grad_norm": 0.4057011604309082,
"learning_rate": 3.015368960704584e-07,
"loss": 0.7681,
"step": 559
},
{
"epoch": 2.7053140096618358,
"grad_norm": 0.4393041431903839,
"learning_rate": 2.919833841902714e-07,
"loss": 0.735,
"step": 560
},
{
"epoch": 2.710144927536232,
"grad_norm": 0.44212228059768677,
"learning_rate": 2.8257910645009935e-07,
"loss": 0.7628,
"step": 561
},
{
"epoch": 2.714975845410628,
"grad_norm": 0.4455677270889282,
"learning_rate": 2.733243609455971e-07,
"loss": 0.7659,
"step": 562
},
{
"epoch": 2.7198067632850242,
"grad_norm": 0.4192866384983063,
"learning_rate": 2.6421944103256657e-07,
"loss": 0.7334,
"step": 563
},
{
"epoch": 2.7246376811594204,
"grad_norm": 0.40188413858413696,
"learning_rate": 2.5526463531765467e-07,
"loss": 0.7454,
"step": 564
},
{
"epoch": 2.7294685990338166,
"grad_norm": 0.46752142906188965,
"learning_rate": 2.4646022764920843e-07,
"loss": 0.7245,
"step": 565
},
{
"epoch": 2.7342995169082127,
"grad_norm": 0.43731552362442017,
"learning_rate": 2.3780649710827552e-07,
"loss": 0.7286,
"step": 566
},
{
"epoch": 2.7391304347826084,
"grad_norm": 0.4043439030647278,
"learning_rate": 2.2930371799975593e-07,
"loss": 0.749,
"step": 567
},
{
"epoch": 2.7439613526570046,
"grad_norm": 0.45707106590270996,
"learning_rate": 2.20952159843712e-07,
"loss": 0.7253,
"step": 568
},
{
"epoch": 2.7487922705314007,
"grad_norm": 0.42659294605255127,
"learning_rate": 2.1275208736682262e-07,
"loss": 0.74,
"step": 569
},
{
"epoch": 2.753623188405797,
"grad_norm": 0.4186418652534485,
"learning_rate": 2.0470376049398944e-07,
"loss": 0.7382,
"step": 570
},
{
"epoch": 2.758454106280193,
"grad_norm": 0.4962987005710602,
"learning_rate": 1.9680743434010385e-07,
"loss": 0.74,
"step": 571
},
{
"epoch": 2.763285024154589,
"grad_norm": 0.45507651567459106,
"learning_rate": 1.8906335920195418e-07,
"loss": 0.7158,
"step": 572
},
{
"epoch": 2.7681159420289854,
"grad_norm": 0.4043714702129364,
"learning_rate": 1.814717805502958e-07,
"loss": 0.7602,
"step": 573
},
{
"epoch": 2.7729468599033815,
"grad_norm": 0.4091343581676483,
"learning_rate": 1.7403293902206851e-07,
"loss": 0.7612,
"step": 574
},
{
"epoch": 2.7777777777777777,
"grad_norm": 0.46215108036994934,
"learning_rate": 1.667470704127694e-07,
"loss": 0.7294,
"step": 575
},
{
"epoch": 2.782608695652174,
"grad_norm": 0.4418387711048126,
"learning_rate": 1.5961440566897913e-07,
"loss": 0.7653,
"step": 576
},
{
"epoch": 2.78743961352657,
"grad_norm": 0.4204854667186737,
"learning_rate": 1.5263517088103862e-07,
"loss": 0.7491,
"step": 577
},
{
"epoch": 2.792270531400966,
"grad_norm": 0.42175278067588806,
"learning_rate": 1.4580958727588746e-07,
"loss": 0.7647,
"step": 578
},
{
"epoch": 2.7971014492753623,
"grad_norm": 0.39392223954200745,
"learning_rate": 1.3913787121004717e-07,
"loss": 0.7367,
"step": 579
},
{
"epoch": 2.8019323671497585,
"grad_norm": 0.44153186678886414,
"learning_rate": 1.3262023416276414e-07,
"loss": 0.7382,
"step": 580
},
{
"epoch": 2.8067632850241546,
"grad_norm": 0.42436960339546204,
"learning_rate": 1.2625688272930925e-07,
"loss": 0.7637,
"step": 581
},
{
"epoch": 2.8115942028985508,
"grad_norm": 0.41522523760795593,
"learning_rate": 1.2004801861442373e-07,
"loss": 0.7504,
"step": 582
},
{
"epoch": 2.816425120772947,
"grad_norm": 0.4469909071922302,
"learning_rate": 1.1399383862592928e-07,
"loss": 0.7511,
"step": 583
},
{
"epoch": 2.821256038647343,
"grad_norm": 0.46624404191970825,
"learning_rate": 1.0809453466849029e-07,
"loss": 0.7385,
"step": 584
},
{
"epoch": 2.8260869565217392,
"grad_norm": 0.41857796907424927,
"learning_rate": 1.0235029373752758e-07,
"loss": 0.7778,
"step": 585
},
{
"epoch": 2.8309178743961354,
"grad_norm": 0.42972105741500854,
"learning_rate": 9.676129791329481e-08,
"loss": 0.7589,
"step": 586
},
{
"epoch": 2.8357487922705316,
"grad_norm": 0.40334904193878174,
"learning_rate": 9.132772435510362e-08,
"loss": 0.7381,
"step": 587
},
{
"epoch": 2.8405797101449277,
"grad_norm": 0.4147365689277649,
"learning_rate": 8.604974529571042e-08,
"loss": 0.7363,
"step": 588
},
{
"epoch": 2.845410628019324,
"grad_norm": 0.3909109830856323,
"learning_rate": 8.092752803585513e-08,
"loss": 0.7541,
"step": 589
},
{
"epoch": 2.85024154589372,
"grad_norm": 0.4555034041404724,
"learning_rate": 7.59612349389599e-08,
"loss": 0.7431,
"step": 590
},
{
"epoch": 2.855072463768116,
"grad_norm": 0.4270079433917999,
"learning_rate": 7.115102342598101e-08,
"loss": 0.7417,
"step": 591
},
{
"epoch": 2.8599033816425123,
"grad_norm": 0.41564372181892395,
"learning_rate": 6.649704597042061e-08,
"loss": 0.78,
"step": 592
},
{
"epoch": 2.864734299516908,
"grad_norm": 0.4386310875415802,
"learning_rate": 6.199945009349173e-08,
"loss": 0.735,
"step": 593
},
{
"epoch": 2.869565217391304,
"grad_norm": 0.47602468729019165,
"learning_rate": 5.7658378359443104e-08,
"loss": 0.7267,
"step": 594
},
{
"epoch": 2.8743961352657004,
"grad_norm": 0.41371598839759827,
"learning_rate": 5.3473968371040575e-08,
"loss": 0.7403,
"step": 595
},
{
"epoch": 2.8792270531400965,
"grad_norm": 0.404881089925766,
"learning_rate": 4.944635276520393e-08,
"loss": 0.7199,
"step": 596
},
{
"epoch": 2.8840579710144927,
"grad_norm": 0.4284209609031677,
"learning_rate": 4.55756592088058e-08,
"loss": 0.7354,
"step": 597
},
{
"epoch": 2.888888888888889,
"grad_norm": 0.40044698119163513,
"learning_rate": 4.186201039462046e-08,
"loss": 0.7433,
"step": 598
},
{
"epoch": 2.893719806763285,
"grad_norm": 0.43060269951820374,
"learning_rate": 3.8305524037438035e-08,
"loss": 0.7518,
"step": 599
},
{
"epoch": 2.898550724637681,
"grad_norm": 0.42064300179481506,
"learning_rate": 3.4906312870331973e-08,
"loss": 0.7525,
"step": 600
},
{
"epoch": 2.9033816425120773,
"grad_norm": 0.3958161771297455,
"learning_rate": 3.166448464108629e-08,
"loss": 0.7512,
"step": 601
},
{
"epoch": 2.9082125603864735,
"grad_norm": 0.4121488630771637,
"learning_rate": 2.8580142108778354e-08,
"loss": 0.7784,
"step": 602
},
{
"epoch": 2.9130434782608696,
"grad_norm": 0.4234846234321594,
"learning_rate": 2.5653383040524228e-08,
"loss": 0.7282,
"step": 603
},
{
"epoch": 2.917874396135266,
"grad_norm": 0.4192802608013153,
"learning_rate": 2.2884300208378395e-08,
"loss": 0.7426,
"step": 604
},
{
"epoch": 2.922705314009662,
"grad_norm": 0.38649386167526245,
"learning_rate": 2.0272981386393332e-08,
"loss": 0.7535,
"step": 605
},
{
"epoch": 2.927536231884058,
"grad_norm": 0.39386647939682007,
"learning_rate": 1.781950934783505e-08,
"loss": 0.7877,
"step": 606
},
{
"epoch": 2.9323671497584543,
"grad_norm": 0.41560304164886475,
"learning_rate": 1.552396186256411e-08,
"loss": 0.738,
"step": 607
},
{
"epoch": 2.9371980676328504,
"grad_norm": 0.4333752393722534,
"learning_rate": 1.3386411694565894e-08,
"loss": 0.7643,
"step": 608
},
{
"epoch": 2.942028985507246,
"grad_norm": 0.3939555287361145,
"learning_rate": 1.1406926599646373e-08,
"loss": 0.7404,
"step": 609
},
{
"epoch": 2.9468599033816423,
"grad_norm": 0.4290173649787903,
"learning_rate": 9.585569323284915e-09,
"loss": 0.7406,
"step": 610
},
{
"epoch": 2.9516908212560384,
"grad_norm": 0.38762733340263367,
"learning_rate": 7.922397598642551e-09,
"loss": 0.7616,
"step": 611
},
{
"epoch": 2.9565217391304346,
"grad_norm": 0.44670021533966064,
"learning_rate": 6.417464144736208e-09,
"loss": 0.7457,
"step": 612
},
{
"epoch": 2.9613526570048307,
"grad_norm": 0.45436298847198486,
"learning_rate": 5.0708166647628345e-09,
"loss": 0.7153,
"step": 613
},
{
"epoch": 2.966183574879227,
"grad_norm": 0.44896742701530457,
"learning_rate": 3.88249784459227e-09,
"loss": 0.7393,
"step": 614
},
{
"epoch": 2.971014492753623,
"grad_norm": 0.4394199848175049,
"learning_rate": 2.8525453514099966e-09,
"loss": 0.7142,
"step": 615
},
{
"epoch": 2.975845410628019,
"grad_norm": 0.38278430700302124,
"learning_rate": 1.980991832524759e-09,
"loss": 0.7544,
"step": 616
},
{
"epoch": 2.9806763285024154,
"grad_norm": 0.403142511844635,
"learning_rate": 1.2678649143349485e-09,
"loss": 0.7158,
"step": 617
},
{
"epoch": 2.9855072463768115,
"grad_norm": 0.39844968914985657,
"learning_rate": 7.131872014509711e-10,
"loss": 0.7541,
"step": 618
},
{
"epoch": 2.9903381642512077,
"grad_norm": 0.4012044370174408,
"learning_rate": 3.1697627597970794e-10,
"loss": 0.7623,
"step": 619
},
{
"epoch": 2.995169082125604,
"grad_norm": 0.47099077701568604,
"learning_rate": 7.924469696718451e-11,
"loss": 0.7556,
"step": 620
},
{
"epoch": 3.0,
"grad_norm": 0.4217309057712555,
"learning_rate": 0.0,
"loss": 0.7747,
"step": 621
},
{
"epoch": 3.0,
"step": 621,
"total_flos": 536280491753472.0,
"train_loss": 0.8154453666889725,
"train_runtime": 29892.1358,
"train_samples_per_second": 1.994,
"train_steps_per_second": 0.021
}
],
"logging_steps": 1.0,
"max_steps": 621,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 536280491753472.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}