|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 621, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.004830917874396135, |
|
"grad_norm": 6.679134368896484, |
|
"learning_rate": 1.5873015873015874e-07, |
|
"loss": 1.2509, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00966183574879227, |
|
"grad_norm": 6.96745491027832, |
|
"learning_rate": 3.174603174603175e-07, |
|
"loss": 1.294, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.014492753623188406, |
|
"grad_norm": 6.5353102684021, |
|
"learning_rate": 4.7619047619047623e-07, |
|
"loss": 1.2325, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01932367149758454, |
|
"grad_norm": 6.902965068817139, |
|
"learning_rate": 6.34920634920635e-07, |
|
"loss": 1.2676, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.024154589371980676, |
|
"grad_norm": 6.7968244552612305, |
|
"learning_rate": 7.936507936507937e-07, |
|
"loss": 1.2375, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.028985507246376812, |
|
"grad_norm": 6.779526233673096, |
|
"learning_rate": 9.523809523809525e-07, |
|
"loss": 1.2592, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.033816425120772944, |
|
"grad_norm": 6.198831558227539, |
|
"learning_rate": 1.111111111111111e-06, |
|
"loss": 1.2144, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.03864734299516908, |
|
"grad_norm": 5.981875419616699, |
|
"learning_rate": 1.26984126984127e-06, |
|
"loss": 1.1995, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.043478260869565216, |
|
"grad_norm": 4.696550369262695, |
|
"learning_rate": 1.4285714285714286e-06, |
|
"loss": 1.1667, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.04830917874396135, |
|
"grad_norm": 4.801881790161133, |
|
"learning_rate": 1.5873015873015873e-06, |
|
"loss": 1.2129, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.05314009661835749, |
|
"grad_norm": 4.528290748596191, |
|
"learning_rate": 1.746031746031746e-06, |
|
"loss": 1.2057, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.057971014492753624, |
|
"grad_norm": 2.914287805557251, |
|
"learning_rate": 1.904761904761905e-06, |
|
"loss": 1.1337, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.06280193236714976, |
|
"grad_norm": 2.909637212753296, |
|
"learning_rate": 2.0634920634920634e-06, |
|
"loss": 1.1577, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.06763285024154589, |
|
"grad_norm": 2.6452553272247314, |
|
"learning_rate": 2.222222222222222e-06, |
|
"loss": 1.1333, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.07246376811594203, |
|
"grad_norm": 2.5321853160858154, |
|
"learning_rate": 2.380952380952381e-06, |
|
"loss": 1.1074, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.07729468599033816, |
|
"grad_norm": 3.3950793743133545, |
|
"learning_rate": 2.53968253968254e-06, |
|
"loss": 1.1057, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0821256038647343, |
|
"grad_norm": 3.9431092739105225, |
|
"learning_rate": 2.6984126984126986e-06, |
|
"loss": 1.1045, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.08695652173913043, |
|
"grad_norm": 3.998680830001831, |
|
"learning_rate": 2.8571428571428573e-06, |
|
"loss": 1.1112, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.09178743961352658, |
|
"grad_norm": 3.75687837600708, |
|
"learning_rate": 3.015873015873016e-06, |
|
"loss": 1.0995, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.0966183574879227, |
|
"grad_norm": 3.270338535308838, |
|
"learning_rate": 3.1746031746031746e-06, |
|
"loss": 1.0742, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.10144927536231885, |
|
"grad_norm": 2.4752681255340576, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 1.0497, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.10628019323671498, |
|
"grad_norm": 2.0890939235687256, |
|
"learning_rate": 3.492063492063492e-06, |
|
"loss": 1.0556, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.1111111111111111, |
|
"grad_norm": 1.8987175226211548, |
|
"learning_rate": 3.6507936507936507e-06, |
|
"loss": 1.0459, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.11594202898550725, |
|
"grad_norm": 1.8444652557373047, |
|
"learning_rate": 3.80952380952381e-06, |
|
"loss": 1.0059, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.12077294685990338, |
|
"grad_norm": 1.6351381540298462, |
|
"learning_rate": 3.968253968253968e-06, |
|
"loss": 0.9934, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.12560386473429952, |
|
"grad_norm": 1.5129081010818481, |
|
"learning_rate": 4.126984126984127e-06, |
|
"loss": 1.0023, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.13043478260869565, |
|
"grad_norm": 1.260694146156311, |
|
"learning_rate": 4.2857142857142855e-06, |
|
"loss": 1.0085, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.13526570048309178, |
|
"grad_norm": 1.1098384857177734, |
|
"learning_rate": 4.444444444444444e-06, |
|
"loss": 1.0117, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.14009661835748793, |
|
"grad_norm": 1.0046857595443726, |
|
"learning_rate": 4.603174603174604e-06, |
|
"loss": 1.022, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.14492753623188406, |
|
"grad_norm": 1.019243836402893, |
|
"learning_rate": 4.761904761904762e-06, |
|
"loss": 0.9697, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.1497584541062802, |
|
"grad_norm": 1.1034119129180908, |
|
"learning_rate": 4.920634920634921e-06, |
|
"loss": 0.9374, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.15458937198067632, |
|
"grad_norm": 1.0104053020477295, |
|
"learning_rate": 5.07936507936508e-06, |
|
"loss": 0.9662, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.15942028985507245, |
|
"grad_norm": 0.8873758316040039, |
|
"learning_rate": 5.2380952380952384e-06, |
|
"loss": 0.9419, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.1642512077294686, |
|
"grad_norm": 0.892684817314148, |
|
"learning_rate": 5.396825396825397e-06, |
|
"loss": 0.9036, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.16908212560386474, |
|
"grad_norm": 0.867564857006073, |
|
"learning_rate": 5.555555555555557e-06, |
|
"loss": 0.9402, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.17391304347826086, |
|
"grad_norm": 0.9065471887588501, |
|
"learning_rate": 5.7142857142857145e-06, |
|
"loss": 0.9335, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.178743961352657, |
|
"grad_norm": 0.8308513760566711, |
|
"learning_rate": 5.873015873015874e-06, |
|
"loss": 0.9084, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.18357487922705315, |
|
"grad_norm": 0.8211532831192017, |
|
"learning_rate": 6.031746031746032e-06, |
|
"loss": 0.9343, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.18840579710144928, |
|
"grad_norm": 0.9129012227058411, |
|
"learning_rate": 6.1904761904761914e-06, |
|
"loss": 0.9141, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.1932367149758454, |
|
"grad_norm": 0.9120469093322754, |
|
"learning_rate": 6.349206349206349e-06, |
|
"loss": 0.9409, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.19806763285024154, |
|
"grad_norm": 0.8367810845375061, |
|
"learning_rate": 6.507936507936509e-06, |
|
"loss": 0.9085, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.2028985507246377, |
|
"grad_norm": 0.7510740756988525, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 0.9286, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.20772946859903382, |
|
"grad_norm": 0.6897188425064087, |
|
"learning_rate": 6.825396825396826e-06, |
|
"loss": 0.9239, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.21256038647342995, |
|
"grad_norm": 0.7415544986724854, |
|
"learning_rate": 6.984126984126984e-06, |
|
"loss": 0.9135, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.21739130434782608, |
|
"grad_norm": 0.8625911474227905, |
|
"learning_rate": 7.1428571428571436e-06, |
|
"loss": 0.9253, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.2222222222222222, |
|
"grad_norm": 0.745945930480957, |
|
"learning_rate": 7.301587301587301e-06, |
|
"loss": 0.9366, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.22705314009661837, |
|
"grad_norm": 0.8231242895126343, |
|
"learning_rate": 7.460317460317461e-06, |
|
"loss": 0.9105, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.2318840579710145, |
|
"grad_norm": 0.6277362704277039, |
|
"learning_rate": 7.61904761904762e-06, |
|
"loss": 0.8868, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.23671497584541062, |
|
"grad_norm": 0.6948869228363037, |
|
"learning_rate": 7.77777777777778e-06, |
|
"loss": 0.887, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.24154589371980675, |
|
"grad_norm": 0.5923482775688171, |
|
"learning_rate": 7.936507936507936e-06, |
|
"loss": 0.906, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.2463768115942029, |
|
"grad_norm": 0.6197386980056763, |
|
"learning_rate": 8.095238095238097e-06, |
|
"loss": 0.9321, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.25120772946859904, |
|
"grad_norm": 0.8314595222473145, |
|
"learning_rate": 8.253968253968254e-06, |
|
"loss": 0.9195, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.2560386473429952, |
|
"grad_norm": 0.7024510502815247, |
|
"learning_rate": 8.412698412698414e-06, |
|
"loss": 0.932, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.2608695652173913, |
|
"grad_norm": 0.8203377723693848, |
|
"learning_rate": 8.571428571428571e-06, |
|
"loss": 0.9115, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.26570048309178745, |
|
"grad_norm": 0.7045034170150757, |
|
"learning_rate": 8.730158730158731e-06, |
|
"loss": 0.917, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.27053140096618356, |
|
"grad_norm": 0.6258231401443481, |
|
"learning_rate": 8.888888888888888e-06, |
|
"loss": 0.8916, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.2753623188405797, |
|
"grad_norm": 0.7873698472976685, |
|
"learning_rate": 9.047619047619049e-06, |
|
"loss": 0.8831, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.28019323671497587, |
|
"grad_norm": 0.9555951952934265, |
|
"learning_rate": 9.206349206349207e-06, |
|
"loss": 0.901, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.28502415458937197, |
|
"grad_norm": 0.6864539384841919, |
|
"learning_rate": 9.365079365079366e-06, |
|
"loss": 0.9124, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.2898550724637681, |
|
"grad_norm": 0.9573421478271484, |
|
"learning_rate": 9.523809523809525e-06, |
|
"loss": 0.8774, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.2946859903381642, |
|
"grad_norm": 0.775773286819458, |
|
"learning_rate": 9.682539682539683e-06, |
|
"loss": 0.9277, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.2995169082125604, |
|
"grad_norm": 0.639119029045105, |
|
"learning_rate": 9.841269841269842e-06, |
|
"loss": 0.8504, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.30434782608695654, |
|
"grad_norm": 0.912753164768219, |
|
"learning_rate": 1e-05, |
|
"loss": 0.9119, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.30917874396135264, |
|
"grad_norm": 0.695167064666748, |
|
"learning_rate": 9.999920755303033e-06, |
|
"loss": 0.8696, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.3140096618357488, |
|
"grad_norm": 0.8863600492477417, |
|
"learning_rate": 9.999683023724021e-06, |
|
"loss": 0.8948, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.3188405797101449, |
|
"grad_norm": 0.7954296469688416, |
|
"learning_rate": 9.99928681279855e-06, |
|
"loss": 0.9085, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.32367149758454106, |
|
"grad_norm": 0.7403387427330017, |
|
"learning_rate": 9.998732135085665e-06, |
|
"loss": 0.9368, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.3285024154589372, |
|
"grad_norm": 0.7288274168968201, |
|
"learning_rate": 9.998019008167476e-06, |
|
"loss": 0.8966, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.3333333333333333, |
|
"grad_norm": 0.836815595626831, |
|
"learning_rate": 9.99714745464859e-06, |
|
"loss": 0.8806, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.33816425120772947, |
|
"grad_norm": 0.8034535646438599, |
|
"learning_rate": 9.99611750215541e-06, |
|
"loss": 0.9018, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.34299516908212563, |
|
"grad_norm": 0.7000673413276672, |
|
"learning_rate": 9.994929183335237e-06, |
|
"loss": 0.8668, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.34782608695652173, |
|
"grad_norm": 0.7168377637863159, |
|
"learning_rate": 9.993582535855265e-06, |
|
"loss": 0.8809, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.3526570048309179, |
|
"grad_norm": 0.8486096858978271, |
|
"learning_rate": 9.992077602401358e-06, |
|
"loss": 0.8878, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.357487922705314, |
|
"grad_norm": 0.6581878662109375, |
|
"learning_rate": 9.990414430676716e-06, |
|
"loss": 0.8555, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.36231884057971014, |
|
"grad_norm": 0.8234553933143616, |
|
"learning_rate": 9.988593073400354e-06, |
|
"loss": 0.8852, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.3671497584541063, |
|
"grad_norm": 0.7650076746940613, |
|
"learning_rate": 9.986613588305435e-06, |
|
"loss": 0.8572, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.3719806763285024, |
|
"grad_norm": 0.6619766354560852, |
|
"learning_rate": 9.984476038137437e-06, |
|
"loss": 0.9015, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.37681159420289856, |
|
"grad_norm": 0.9842355847358704, |
|
"learning_rate": 9.982180490652165e-06, |
|
"loss": 0.8912, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.38164251207729466, |
|
"grad_norm": 0.8628110289573669, |
|
"learning_rate": 9.979727018613607e-06, |
|
"loss": 0.9033, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.3864734299516908, |
|
"grad_norm": 0.8167057633399963, |
|
"learning_rate": 9.977115699791622e-06, |
|
"loss": 0.8583, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.391304347826087, |
|
"grad_norm": 0.8511431813240051, |
|
"learning_rate": 9.974346616959476e-06, |
|
"loss": 0.8625, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.3961352657004831, |
|
"grad_norm": 0.8892733454704285, |
|
"learning_rate": 9.971419857891223e-06, |
|
"loss": 0.8967, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.40096618357487923, |
|
"grad_norm": 0.6108592748641968, |
|
"learning_rate": 9.968335515358916e-06, |
|
"loss": 0.8727, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.4057971014492754, |
|
"grad_norm": 0.9057552218437195, |
|
"learning_rate": 9.965093687129669e-06, |
|
"loss": 0.8489, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.4106280193236715, |
|
"grad_norm": 0.7670313715934753, |
|
"learning_rate": 9.961694475962562e-06, |
|
"loss": 0.8862, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.41545893719806765, |
|
"grad_norm": 0.7009760141372681, |
|
"learning_rate": 9.95813798960538e-06, |
|
"loss": 0.8481, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.42028985507246375, |
|
"grad_norm": 0.8547497391700745, |
|
"learning_rate": 9.954424340791195e-06, |
|
"loss": 0.8574, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.4251207729468599, |
|
"grad_norm": 0.6579576730728149, |
|
"learning_rate": 9.950553647234798e-06, |
|
"loss": 0.8876, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.42995169082125606, |
|
"grad_norm": 0.7313647270202637, |
|
"learning_rate": 9.94652603162896e-06, |
|
"loss": 0.8833, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.43478260869565216, |
|
"grad_norm": 0.8301870226860046, |
|
"learning_rate": 9.942341621640558e-06, |
|
"loss": 0.842, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.4396135265700483, |
|
"grad_norm": 0.634363055229187, |
|
"learning_rate": 9.938000549906509e-06, |
|
"loss": 0.8724, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.4444444444444444, |
|
"grad_norm": 0.6919889450073242, |
|
"learning_rate": 9.93350295402958e-06, |
|
"loss": 0.8459, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.4492753623188406, |
|
"grad_norm": 0.7910676598548889, |
|
"learning_rate": 9.92884897657402e-06, |
|
"loss": 0.8793, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.45410628019323673, |
|
"grad_norm": 0.7085094451904297, |
|
"learning_rate": 9.924038765061042e-06, |
|
"loss": 0.8618, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.45893719806763283, |
|
"grad_norm": 0.8088275194168091, |
|
"learning_rate": 9.919072471964146e-06, |
|
"loss": 0.8547, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.463768115942029, |
|
"grad_norm": 0.6322200298309326, |
|
"learning_rate": 9.913950254704291e-06, |
|
"loss": 0.8545, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.46859903381642515, |
|
"grad_norm": 0.7928032279014587, |
|
"learning_rate": 9.908672275644898e-06, |
|
"loss": 0.8642, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.47342995169082125, |
|
"grad_norm": 0.7358760237693787, |
|
"learning_rate": 9.903238702086707e-06, |
|
"loss": 0.8677, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.4782608695652174, |
|
"grad_norm": 0.7627845406532288, |
|
"learning_rate": 9.897649706262474e-06, |
|
"loss": 0.8784, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.4830917874396135, |
|
"grad_norm": 0.7139146327972412, |
|
"learning_rate": 9.89190546533151e-06, |
|
"loss": 0.8929, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.48792270531400966, |
|
"grad_norm": 0.7086196541786194, |
|
"learning_rate": 9.88600616137407e-06, |
|
"loss": 0.8687, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.4927536231884058, |
|
"grad_norm": 0.7507730722427368, |
|
"learning_rate": 9.879951981385577e-06, |
|
"loss": 0.8278, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.4975845410628019, |
|
"grad_norm": 0.7131170630455017, |
|
"learning_rate": 9.873743117270691e-06, |
|
"loss": 0.8751, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.5024154589371981, |
|
"grad_norm": 0.9851748943328857, |
|
"learning_rate": 9.867379765837237e-06, |
|
"loss": 0.8885, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.5072463768115942, |
|
"grad_norm": 0.7180512547492981, |
|
"learning_rate": 9.860862128789954e-06, |
|
"loss": 0.8686, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.5120772946859904, |
|
"grad_norm": 0.8217784762382507, |
|
"learning_rate": 9.854190412724114e-06, |
|
"loss": 0.8665, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.5169082125603864, |
|
"grad_norm": 0.863905668258667, |
|
"learning_rate": 9.847364829118963e-06, |
|
"loss": 0.86, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.5217391304347826, |
|
"grad_norm": 0.9537051916122437, |
|
"learning_rate": 9.840385594331022e-06, |
|
"loss": 0.8251, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.5265700483091788, |
|
"grad_norm": 0.6802311539649963, |
|
"learning_rate": 9.833252929587231e-06, |
|
"loss": 0.8688, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.5314009661835749, |
|
"grad_norm": 0.7750963568687439, |
|
"learning_rate": 9.825967060977933e-06, |
|
"loss": 0.8599, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.5362318840579711, |
|
"grad_norm": 0.7233647108078003, |
|
"learning_rate": 9.818528219449705e-06, |
|
"loss": 0.8659, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.5410628019323671, |
|
"grad_norm": 0.6166982054710388, |
|
"learning_rate": 9.810936640798046e-06, |
|
"loss": 0.853, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.5458937198067633, |
|
"grad_norm": 0.6997689008712769, |
|
"learning_rate": 9.803192565659898e-06, |
|
"loss": 0.8404, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.5507246376811594, |
|
"grad_norm": 0.6236562728881836, |
|
"learning_rate": 9.795296239506011e-06, |
|
"loss": 0.8722, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.5555555555555556, |
|
"grad_norm": 0.7246277332305908, |
|
"learning_rate": 9.78724791263318e-06, |
|
"loss": 0.8656, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.5603864734299517, |
|
"grad_norm": 0.6707713603973389, |
|
"learning_rate": 9.779047840156288e-06, |
|
"loss": 0.8821, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.5652173913043478, |
|
"grad_norm": 0.596466064453125, |
|
"learning_rate": 9.770696282000245e-06, |
|
"loss": 0.8461, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.5700483091787439, |
|
"grad_norm": 0.6369594931602478, |
|
"learning_rate": 9.762193502891726e-06, |
|
"loss": 0.8723, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.5748792270531401, |
|
"grad_norm": 0.6298103928565979, |
|
"learning_rate": 9.753539772350792e-06, |
|
"loss": 0.8751, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.5797101449275363, |
|
"grad_norm": 0.7989038825035095, |
|
"learning_rate": 9.744735364682347e-06, |
|
"loss": 0.8829, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.5845410628019324, |
|
"grad_norm": 0.6647093296051025, |
|
"learning_rate": 9.735780558967434e-06, |
|
"loss": 0.8725, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.5893719806763285, |
|
"grad_norm": 0.6525388360023499, |
|
"learning_rate": 9.726675639054403e-06, |
|
"loss": 0.8398, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.5942028985507246, |
|
"grad_norm": 0.6918199062347412, |
|
"learning_rate": 9.717420893549902e-06, |
|
"loss": 0.8523, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.5990338164251208, |
|
"grad_norm": 0.6548693776130676, |
|
"learning_rate": 9.70801661580973e-06, |
|
"loss": 0.8584, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.6038647342995169, |
|
"grad_norm": 0.7120753526687622, |
|
"learning_rate": 9.698463103929542e-06, |
|
"loss": 0.8821, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.6086956521739131, |
|
"grad_norm": 0.6639739871025085, |
|
"learning_rate": 9.688760660735403e-06, |
|
"loss": 0.8791, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.6135265700483091, |
|
"grad_norm": 0.5813497304916382, |
|
"learning_rate": 9.67890959377418e-06, |
|
"loss": 0.8738, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.6183574879227053, |
|
"grad_norm": 0.7182924747467041, |
|
"learning_rate": 9.668910215303797e-06, |
|
"loss": 0.8889, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.6231884057971014, |
|
"grad_norm": 0.6352912783622742, |
|
"learning_rate": 9.658762842283343e-06, |
|
"loss": 0.8625, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.6280193236714976, |
|
"grad_norm": 0.7149222493171692, |
|
"learning_rate": 9.648467796363019e-06, |
|
"loss": 0.8458, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.6328502415458938, |
|
"grad_norm": 0.591304361820221, |
|
"learning_rate": 9.638025403873939e-06, |
|
"loss": 0.8241, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.6376811594202898, |
|
"grad_norm": 0.6942040920257568, |
|
"learning_rate": 9.627435995817799e-06, |
|
"loss": 0.8459, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.642512077294686, |
|
"grad_norm": 0.6825373768806458, |
|
"learning_rate": 9.616699907856368e-06, |
|
"loss": 0.881, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.6473429951690821, |
|
"grad_norm": 0.7252668738365173, |
|
"learning_rate": 9.605817480300863e-06, |
|
"loss": 0.822, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.6521739130434783, |
|
"grad_norm": 0.702181875705719, |
|
"learning_rate": 9.594789058101154e-06, |
|
"loss": 0.8841, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.6570048309178744, |
|
"grad_norm": 0.6287996768951416, |
|
"learning_rate": 9.58361499083483e-06, |
|
"loss": 0.8266, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.6618357487922706, |
|
"grad_norm": 0.7371240258216858, |
|
"learning_rate": 9.57229563269612e-06, |
|
"loss": 0.8691, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.6666666666666666, |
|
"grad_norm": 0.8119018077850342, |
|
"learning_rate": 9.560831342484668e-06, |
|
"loss": 0.8842, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.6714975845410628, |
|
"grad_norm": 0.6123917102813721, |
|
"learning_rate": 9.549222483594154e-06, |
|
"loss": 0.82, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.6763285024154589, |
|
"grad_norm": 0.7341845631599426, |
|
"learning_rate": 9.53746942400078e-06, |
|
"loss": 0.8919, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.6811594202898551, |
|
"grad_norm": 0.6273525357246399, |
|
"learning_rate": 9.525572536251608e-06, |
|
"loss": 0.8453, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.6859903381642513, |
|
"grad_norm": 0.6209537386894226, |
|
"learning_rate": 9.513532197452737e-06, |
|
"loss": 0.833, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.6908212560386473, |
|
"grad_norm": 0.7676011919975281, |
|
"learning_rate": 9.501348789257373e-06, |
|
"loss": 0.8634, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.6956521739130435, |
|
"grad_norm": 0.5898526906967163, |
|
"learning_rate": 9.48902269785371e-06, |
|
"loss": 0.8844, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.7004830917874396, |
|
"grad_norm": 0.7219299077987671, |
|
"learning_rate": 9.476554313952697e-06, |
|
"loss": 0.8543, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.7053140096618358, |
|
"grad_norm": 0.6188173890113831, |
|
"learning_rate": 9.46394403277566e-06, |
|
"loss": 0.8325, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.7101449275362319, |
|
"grad_norm": 0.6188879609107971, |
|
"learning_rate": 9.451192254041759e-06, |
|
"loss": 0.8591, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.714975845410628, |
|
"grad_norm": 0.648301899433136, |
|
"learning_rate": 9.438299381955333e-06, |
|
"loss": 0.8646, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.7198067632850241, |
|
"grad_norm": 0.6337832808494568, |
|
"learning_rate": 9.425265825193077e-06, |
|
"loss": 0.8394, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.7246376811594203, |
|
"grad_norm": 0.5935664772987366, |
|
"learning_rate": 9.412091996891097e-06, |
|
"loss": 0.8865, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.7294685990338164, |
|
"grad_norm": 0.759688675403595, |
|
"learning_rate": 9.398778314631801e-06, |
|
"loss": 0.8644, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.7342995169082126, |
|
"grad_norm": 0.65163654088974, |
|
"learning_rate": 9.385325200430679e-06, |
|
"loss": 0.8421, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.7391304347826086, |
|
"grad_norm": 0.7104676365852356, |
|
"learning_rate": 9.371733080722911e-06, |
|
"loss": 0.8562, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.7439613526570048, |
|
"grad_norm": 0.6826310753822327, |
|
"learning_rate": 9.358002386349862e-06, |
|
"loss": 0.8867, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.748792270531401, |
|
"grad_norm": 0.6869804263114929, |
|
"learning_rate": 9.34413355254542e-06, |
|
"loss": 0.8403, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.7536231884057971, |
|
"grad_norm": 0.6028906106948853, |
|
"learning_rate": 9.330127018922195e-06, |
|
"loss": 0.8489, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.7584541062801933, |
|
"grad_norm": 0.8058885335922241, |
|
"learning_rate": 9.31598322945759e-06, |
|
"loss": 0.8623, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.7632850241545893, |
|
"grad_norm": 0.5606011748313904, |
|
"learning_rate": 9.301702632479734e-06, |
|
"loss": 0.8434, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.7681159420289855, |
|
"grad_norm": 0.7772120237350464, |
|
"learning_rate": 9.287285680653254e-06, |
|
"loss": 0.8924, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.7729468599033816, |
|
"grad_norm": 0.6535057425498962, |
|
"learning_rate": 9.272732830964948e-06, |
|
"loss": 0.8694, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.7777777777777778, |
|
"grad_norm": 0.6197230219841003, |
|
"learning_rate": 9.258044544709276e-06, |
|
"loss": 0.8541, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.782608695652174, |
|
"grad_norm": 0.6519168019294739, |
|
"learning_rate": 9.243221287473755e-06, |
|
"loss": 0.8679, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.7874396135265701, |
|
"grad_norm": 0.8101065158843994, |
|
"learning_rate": 9.228263529124199e-06, |
|
"loss": 0.8318, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.7922705314009661, |
|
"grad_norm": 0.6714143753051758, |
|
"learning_rate": 9.21317174378982e-06, |
|
"loss": 0.8526, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.7971014492753623, |
|
"grad_norm": 0.8356069922447205, |
|
"learning_rate": 9.197946409848196e-06, |
|
"loss": 0.8707, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.8019323671497585, |
|
"grad_norm": 0.6714097857475281, |
|
"learning_rate": 9.182588009910119e-06, |
|
"loss": 0.8601, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.8067632850241546, |
|
"grad_norm": 0.6646215319633484, |
|
"learning_rate": 9.167097030804289e-06, |
|
"loss": 0.8398, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.8115942028985508, |
|
"grad_norm": 0.6809929609298706, |
|
"learning_rate": 9.151473963561884e-06, |
|
"loss": 0.7931, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.8164251207729468, |
|
"grad_norm": 0.7758853435516357, |
|
"learning_rate": 9.135719303400995e-06, |
|
"loss": 0.8542, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.821256038647343, |
|
"grad_norm": 0.743437647819519, |
|
"learning_rate": 9.119833549710927e-06, |
|
"loss": 0.8251, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.8260869565217391, |
|
"grad_norm": 0.6809033155441284, |
|
"learning_rate": 9.103817206036383e-06, |
|
"loss": 0.8496, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.8309178743961353, |
|
"grad_norm": 0.6170421838760376, |
|
"learning_rate": 9.087670780061477e-06, |
|
"loss": 0.8558, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.8357487922705314, |
|
"grad_norm": 0.5970268845558167, |
|
"learning_rate": 9.071394783593664e-06, |
|
"loss": 0.8295, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.8405797101449275, |
|
"grad_norm": 0.6291173100471497, |
|
"learning_rate": 9.054989732547507e-06, |
|
"loss": 0.859, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.8454106280193237, |
|
"grad_norm": 0.6694471836090088, |
|
"learning_rate": 9.038456146928325e-06, |
|
"loss": 0.8654, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.8502415458937198, |
|
"grad_norm": 0.6172890663146973, |
|
"learning_rate": 9.021794550815713e-06, |
|
"loss": 0.8752, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.855072463768116, |
|
"grad_norm": 0.58245849609375, |
|
"learning_rate": 9.005005472346923e-06, |
|
"loss": 0.8326, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.8599033816425121, |
|
"grad_norm": 0.5857954621315002, |
|
"learning_rate": 8.988089443700131e-06, |
|
"loss": 0.8185, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.8647342995169082, |
|
"grad_norm": 0.5742454528808594, |
|
"learning_rate": 8.971047001077561e-06, |
|
"loss": 0.8957, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.8695652173913043, |
|
"grad_norm": 0.696999728679657, |
|
"learning_rate": 8.953878684688492e-06, |
|
"loss": 0.8769, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.8743961352657005, |
|
"grad_norm": 0.6432103514671326, |
|
"learning_rate": 8.936585038732143e-06, |
|
"loss": 0.863, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.8792270531400966, |
|
"grad_norm": 0.7360765933990479, |
|
"learning_rate": 8.919166611380397e-06, |
|
"loss": 0.8449, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.8840579710144928, |
|
"grad_norm": 0.6782159805297852, |
|
"learning_rate": 8.90162395476046e-06, |
|
"loss": 0.8473, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"grad_norm": 0.6303816437721252, |
|
"learning_rate": 8.883957624937333e-06, |
|
"loss": 0.839, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.893719806763285, |
|
"grad_norm": 0.6260365843772888, |
|
"learning_rate": 8.866168181896198e-06, |
|
"loss": 0.876, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.8985507246376812, |
|
"grad_norm": 0.5933541059494019, |
|
"learning_rate": 8.848256189524661e-06, |
|
"loss": 0.8496, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.9033816425120773, |
|
"grad_norm": 0.5709136724472046, |
|
"learning_rate": 8.83022221559489e-06, |
|
"loss": 0.8522, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.9082125603864735, |
|
"grad_norm": 0.5638396143913269, |
|
"learning_rate": 8.812066831745602e-06, |
|
"loss": 0.8649, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.9130434782608695, |
|
"grad_norm": 0.6326948404312134, |
|
"learning_rate": 8.793790613463956e-06, |
|
"loss": 0.8464, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.9178743961352657, |
|
"grad_norm": 0.6702078580856323, |
|
"learning_rate": 8.775394140067299e-06, |
|
"loss": 0.8051, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.9227053140096618, |
|
"grad_norm": 0.6150868535041809, |
|
"learning_rate": 8.756877994684818e-06, |
|
"loss": 0.8543, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.927536231884058, |
|
"grad_norm": 0.6802798509597778, |
|
"learning_rate": 8.738242764239046e-06, |
|
"loss": 0.8202, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.9323671497584541, |
|
"grad_norm": 0.6473823189735413, |
|
"learning_rate": 8.719489039427256e-06, |
|
"loss": 0.8058, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.9371980676328503, |
|
"grad_norm": 0.6301891207695007, |
|
"learning_rate": 8.700617414702746e-06, |
|
"loss": 0.8518, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.9420289855072463, |
|
"grad_norm": 0.5725292563438416, |
|
"learning_rate": 8.681628488255986e-06, |
|
"loss": 0.8444, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.9468599033816425, |
|
"grad_norm": 0.6120401620864868, |
|
"learning_rate": 8.66252286199567e-06, |
|
"loss": 0.8505, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.9516908212560387, |
|
"grad_norm": 0.6610070466995239, |
|
"learning_rate": 8.643301141529619e-06, |
|
"loss": 0.8238, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.9565217391304348, |
|
"grad_norm": 0.5605831146240234, |
|
"learning_rate": 8.6239639361456e-06, |
|
"loss": 0.8455, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.961352657004831, |
|
"grad_norm": 0.5961610674858093, |
|
"learning_rate": 8.604511858792006e-06, |
|
"loss": 0.8275, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.966183574879227, |
|
"grad_norm": 0.5695908069610596, |
|
"learning_rate": 8.584945526058426e-06, |
|
"loss": 0.8218, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.9710144927536232, |
|
"grad_norm": 0.6564455032348633, |
|
"learning_rate": 8.565265558156101e-06, |
|
"loss": 0.8291, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.9758454106280193, |
|
"grad_norm": 0.5902118682861328, |
|
"learning_rate": 8.545472578898276e-06, |
|
"loss": 0.87, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.9806763285024155, |
|
"grad_norm": 0.5982496738433838, |
|
"learning_rate": 8.525567215680397e-06, |
|
"loss": 0.8411, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.9855072463768116, |
|
"grad_norm": 0.5730376243591309, |
|
"learning_rate": 8.505550099460264e-06, |
|
"loss": 0.806, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.9903381642512077, |
|
"grad_norm": 0.6329107284545898, |
|
"learning_rate": 8.485421864737997e-06, |
|
"loss": 0.8181, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.9951690821256038, |
|
"grad_norm": 0.5955251455307007, |
|
"learning_rate": 8.465183149535939e-06, |
|
"loss": 0.8617, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.7374234199523926, |
|
"learning_rate": 8.444834595378434e-06, |
|
"loss": 0.8778, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 1.0048309178743962, |
|
"grad_norm": 0.5692369937896729, |
|
"learning_rate": 8.424376847271483e-06, |
|
"loss": 0.7686, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 1.0096618357487923, |
|
"grad_norm": 0.5868251919746399, |
|
"learning_rate": 8.403810553682307e-06, |
|
"loss": 0.7638, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 1.0144927536231885, |
|
"grad_norm": 0.6814491152763367, |
|
"learning_rate": 8.383136366518788e-06, |
|
"loss": 0.7903, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.0193236714975846, |
|
"grad_norm": 0.5972342491149902, |
|
"learning_rate": 8.362354941108803e-06, |
|
"loss": 0.7882, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 1.0241545893719808, |
|
"grad_norm": 0.709455668926239, |
|
"learning_rate": 8.341466936179457e-06, |
|
"loss": 0.7901, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 1.0289855072463767, |
|
"grad_norm": 0.6567758321762085, |
|
"learning_rate": 8.320473013836197e-06, |
|
"loss": 0.8157, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 1.0338164251207729, |
|
"grad_norm": 0.8612772226333618, |
|
"learning_rate": 8.299373839541829e-06, |
|
"loss": 0.7975, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 1.038647342995169, |
|
"grad_norm": 0.7711191177368164, |
|
"learning_rate": 8.278170082095422e-06, |
|
"loss": 0.7865, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 1.0434782608695652, |
|
"grad_norm": 0.6870383024215698, |
|
"learning_rate": 8.256862413611113e-06, |
|
"loss": 0.8089, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 1.0483091787439613, |
|
"grad_norm": 0.6059009432792664, |
|
"learning_rate": 8.23545150949679e-06, |
|
"loss": 0.7751, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 1.0531400966183575, |
|
"grad_norm": 0.7190768122673035, |
|
"learning_rate": 8.213938048432697e-06, |
|
"loss": 0.7912, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 1.0579710144927537, |
|
"grad_norm": 0.7017462253570557, |
|
"learning_rate": 8.192322712349917e-06, |
|
"loss": 0.779, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 1.0628019323671498, |
|
"grad_norm": 0.5871282815933228, |
|
"learning_rate": 8.170606186408756e-06, |
|
"loss": 0.8335, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.067632850241546, |
|
"grad_norm": 0.8193174004554749, |
|
"learning_rate": 8.148789158977012e-06, |
|
"loss": 0.8206, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 1.0724637681159421, |
|
"grad_norm": 0.6280666589736938, |
|
"learning_rate": 8.126872321608185e-06, |
|
"loss": 0.7918, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 1.077294685990338, |
|
"grad_norm": 0.593765139579773, |
|
"learning_rate": 8.104856369019525e-06, |
|
"loss": 0.8393, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 1.0821256038647342, |
|
"grad_norm": 0.6945616602897644, |
|
"learning_rate": 8.08274199907003e-06, |
|
"loss": 0.8082, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.0869565217391304, |
|
"grad_norm": 0.6425365209579468, |
|
"learning_rate": 8.060529912738316e-06, |
|
"loss": 0.7877, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.0917874396135265, |
|
"grad_norm": 0.602546215057373, |
|
"learning_rate": 8.038220814100403e-06, |
|
"loss": 0.7981, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 1.0966183574879227, |
|
"grad_norm": 0.5716975331306458, |
|
"learning_rate": 8.0158154103074e-06, |
|
"loss": 0.7971, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 1.1014492753623188, |
|
"grad_norm": 0.5490554571151733, |
|
"learning_rate": 7.993314411563075e-06, |
|
"loss": 0.7832, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.106280193236715, |
|
"grad_norm": 0.6210470795631409, |
|
"learning_rate": 7.970718531101365e-06, |
|
"loss": 0.8262, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 1.1111111111111112, |
|
"grad_norm": 0.5391402840614319, |
|
"learning_rate": 7.948028485163744e-06, |
|
"loss": 0.7775, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.1159420289855073, |
|
"grad_norm": 0.5497602224349976, |
|
"learning_rate": 7.925244992976538e-06, |
|
"loss": 0.7948, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 1.1207729468599035, |
|
"grad_norm": 0.6025332808494568, |
|
"learning_rate": 7.902368776728125e-06, |
|
"loss": 0.8187, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.1256038647342996, |
|
"grad_norm": 0.569249153137207, |
|
"learning_rate": 7.879400561546033e-06, |
|
"loss": 0.793, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 1.1304347826086956, |
|
"grad_norm": 0.6411007642745972, |
|
"learning_rate": 7.856341075473963e-06, |
|
"loss": 0.7905, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 1.1352657004830917, |
|
"grad_norm": 0.5739008188247681, |
|
"learning_rate": 7.833191049448706e-06, |
|
"loss": 0.7941, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.1400966183574879, |
|
"grad_norm": 0.641865074634552, |
|
"learning_rate": 7.809951217276986e-06, |
|
"loss": 0.8057, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 1.144927536231884, |
|
"grad_norm": 0.6288030743598938, |
|
"learning_rate": 7.786622315612182e-06, |
|
"loss": 0.7624, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 1.1497584541062802, |
|
"grad_norm": 0.6016022562980652, |
|
"learning_rate": 7.763205083930995e-06, |
|
"loss": 0.8071, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 1.1545893719806763, |
|
"grad_norm": 0.517977237701416, |
|
"learning_rate": 7.739700264509993e-06, |
|
"loss": 0.7639, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 1.1594202898550725, |
|
"grad_norm": 0.5750455856323242, |
|
"learning_rate": 7.716108602402094e-06, |
|
"loss": 0.7915, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.1642512077294687, |
|
"grad_norm": 0.5806918144226074, |
|
"learning_rate": 7.692430845412946e-06, |
|
"loss": 0.7944, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 1.1690821256038648, |
|
"grad_norm": 0.5185478925704956, |
|
"learning_rate": 7.668667744077215e-06, |
|
"loss": 0.8127, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 1.1739130434782608, |
|
"grad_norm": 0.5086768269538879, |
|
"learning_rate": 7.644820051634813e-06, |
|
"loss": 0.7772, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 1.178743961352657, |
|
"grad_norm": 0.6286478042602539, |
|
"learning_rate": 7.6208885240069995e-06, |
|
"loss": 0.7506, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.183574879227053, |
|
"grad_norm": 0.6119205355644226, |
|
"learning_rate": 7.596873919772438e-06, |
|
"loss": 0.7967, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 1.1884057971014492, |
|
"grad_norm": 0.5589543581008911, |
|
"learning_rate": 7.572777000143145e-06, |
|
"loss": 0.8363, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 1.1932367149758454, |
|
"grad_norm": 0.6248103380203247, |
|
"learning_rate": 7.548598528940354e-06, |
|
"loss": 0.8014, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 1.1980676328502415, |
|
"grad_norm": 0.5842779874801636, |
|
"learning_rate": 7.524339272570317e-06, |
|
"loss": 0.7897, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.2028985507246377, |
|
"grad_norm": 0.6430383324623108, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.7779, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 1.2077294685990339, |
|
"grad_norm": 0.5114961862564087, |
|
"learning_rate": 7.475581482732717e-06, |
|
"loss": 0.7671, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.21256038647343, |
|
"grad_norm": 0.5602841973304749, |
|
"learning_rate": 7.451084494783668e-06, |
|
"loss": 0.791, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 1.2173913043478262, |
|
"grad_norm": 0.6211389303207397, |
|
"learning_rate": 7.4265098126554065e-06, |
|
"loss": 0.8377, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 1.2222222222222223, |
|
"grad_norm": 0.6181166768074036, |
|
"learning_rate": 7.401858215313228e-06, |
|
"loss": 0.7738, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 1.2270531400966185, |
|
"grad_norm": 0.5928196907043457, |
|
"learning_rate": 7.3771304841604764e-06, |
|
"loss": 0.776, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 1.2318840579710144, |
|
"grad_norm": 0.5380995273590088, |
|
"learning_rate": 7.352327403013779e-06, |
|
"loss": 0.7791, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 1.2367149758454106, |
|
"grad_norm": 0.6451078057289124, |
|
"learning_rate": 7.327449758078194e-06, |
|
"loss": 0.816, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 1.2415458937198067, |
|
"grad_norm": 0.5372149348258972, |
|
"learning_rate": 7.302498337922293e-06, |
|
"loss": 0.8001, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 1.2463768115942029, |
|
"grad_norm": 0.5098853707313538, |
|
"learning_rate": 7.27747393345317e-06, |
|
"loss": 0.7566, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 1.251207729468599, |
|
"grad_norm": 0.6665354371070862, |
|
"learning_rate": 7.2523773378913655e-06, |
|
"loss": 0.7506, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 1.2560386473429952, |
|
"grad_norm": 0.654194712638855, |
|
"learning_rate": 7.2272093467457226e-06, |
|
"loss": 0.8192, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.2608695652173914, |
|
"grad_norm": 0.6488723754882812, |
|
"learning_rate": 7.201970757788172e-06, |
|
"loss": 0.8196, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 1.2657004830917875, |
|
"grad_norm": 0.6150570511817932, |
|
"learning_rate": 7.17666237102845e-06, |
|
"loss": 0.8084, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 1.2705314009661834, |
|
"grad_norm": 0.5813282132148743, |
|
"learning_rate": 7.151284988688731e-06, |
|
"loss": 0.7909, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 1.2753623188405796, |
|
"grad_norm": 0.6258924603462219, |
|
"learning_rate": 7.125839415178204e-06, |
|
"loss": 0.8144, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 1.2801932367149758, |
|
"grad_norm": 0.5735957622528076, |
|
"learning_rate": 7.100326457067576e-06, |
|
"loss": 0.7568, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 1.285024154589372, |
|
"grad_norm": 0.6329940557479858, |
|
"learning_rate": 7.074746923063497e-06, |
|
"loss": 0.7718, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 1.289855072463768, |
|
"grad_norm": 0.5378438830375671, |
|
"learning_rate": 7.049101623982938e-06, |
|
"loss": 0.8011, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 1.2946859903381642, |
|
"grad_norm": 0.6654214859008789, |
|
"learning_rate": 7.02339137272748e-06, |
|
"loss": 0.8156, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 1.2995169082125604, |
|
"grad_norm": 0.7466998100280762, |
|
"learning_rate": 6.9976169842575526e-06, |
|
"loss": 0.7883, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 1.3043478260869565, |
|
"grad_norm": 0.527633547782898, |
|
"learning_rate": 6.971779275566593e-06, |
|
"loss": 0.7967, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.3091787439613527, |
|
"grad_norm": 0.7060922384262085, |
|
"learning_rate": 6.945879065655164e-06, |
|
"loss": 0.8033, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 1.3140096618357489, |
|
"grad_norm": 0.6210879683494568, |
|
"learning_rate": 6.919917175504978e-06, |
|
"loss": 0.7613, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 1.318840579710145, |
|
"grad_norm": 0.5505712032318115, |
|
"learning_rate": 6.893894428052881e-06, |
|
"loss": 0.7572, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 1.3236714975845412, |
|
"grad_norm": 0.5760694742202759, |
|
"learning_rate": 6.867811648164769e-06, |
|
"loss": 0.794, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 1.3285024154589373, |
|
"grad_norm": 0.5254794955253601, |
|
"learning_rate": 6.841669662609437e-06, |
|
"loss": 0.7707, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 1.3333333333333333, |
|
"grad_norm": 0.5995458960533142, |
|
"learning_rate": 6.815469300032374e-06, |
|
"loss": 0.7778, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 1.3381642512077294, |
|
"grad_norm": 0.6530429124832153, |
|
"learning_rate": 6.789211390929497e-06, |
|
"loss": 0.7817, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 1.3429951690821256, |
|
"grad_norm": 0.6288939118385315, |
|
"learning_rate": 6.762896767620827e-06, |
|
"loss": 0.8001, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 1.3478260869565217, |
|
"grad_norm": 0.5950453281402588, |
|
"learning_rate": 6.736526264224101e-06, |
|
"loss": 0.8142, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 1.3526570048309179, |
|
"grad_norm": 0.6037265062332153, |
|
"learning_rate": 6.710100716628345e-06, |
|
"loss": 0.7762, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.357487922705314, |
|
"grad_norm": 0.784972071647644, |
|
"learning_rate": 6.6836209624673575e-06, |
|
"loss": 0.7972, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 1.3623188405797102, |
|
"grad_norm": 0.6714766621589661, |
|
"learning_rate": 6.657087841093179e-06, |
|
"loss": 0.7919, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 1.3671497584541064, |
|
"grad_norm": 0.6096466183662415, |
|
"learning_rate": 6.6305021935494755e-06, |
|
"loss": 0.79, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 1.3719806763285023, |
|
"grad_norm": 0.5656675696372986, |
|
"learning_rate": 6.603864862544879e-06, |
|
"loss": 0.7857, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 1.3768115942028984, |
|
"grad_norm": 0.7064409255981445, |
|
"learning_rate": 6.5771766924262795e-06, |
|
"loss": 0.8041, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 1.3816425120772946, |
|
"grad_norm": 0.6364306211471558, |
|
"learning_rate": 6.5504385291520554e-06, |
|
"loss": 0.8144, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 1.3864734299516908, |
|
"grad_norm": 0.5433293581008911, |
|
"learning_rate": 6.523651220265269e-06, |
|
"loss": 0.8049, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 1.391304347826087, |
|
"grad_norm": 0.6152275204658508, |
|
"learning_rate": 6.496815614866792e-06, |
|
"loss": 0.7886, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 1.396135265700483, |
|
"grad_norm": 0.5549913644790649, |
|
"learning_rate": 6.469932563588386e-06, |
|
"loss": 0.7809, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 1.4009661835748792, |
|
"grad_norm": 0.48793748021125793, |
|
"learning_rate": 6.443002918565754e-06, |
|
"loss": 0.7887, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.4057971014492754, |
|
"grad_norm": 0.6952255964279175, |
|
"learning_rate": 6.41602753341152e-06, |
|
"loss": 0.7703, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 1.4106280193236715, |
|
"grad_norm": 0.5199931859970093, |
|
"learning_rate": 6.389007263188176e-06, |
|
"loss": 0.8129, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 1.4154589371980677, |
|
"grad_norm": 0.5502813458442688, |
|
"learning_rate": 6.361942964380967e-06, |
|
"loss": 0.814, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 1.4202898550724639, |
|
"grad_norm": 0.6408051252365112, |
|
"learning_rate": 6.334835494870759e-06, |
|
"loss": 0.8181, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 1.42512077294686, |
|
"grad_norm": 0.6277318596839905, |
|
"learning_rate": 6.307685713906835e-06, |
|
"loss": 0.8188, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 1.4299516908212562, |
|
"grad_norm": 0.513602614402771, |
|
"learning_rate": 6.2804944820796596e-06, |
|
"loss": 0.8033, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 1.434782608695652, |
|
"grad_norm": 0.6088364124298096, |
|
"learning_rate": 6.2532626612936035e-06, |
|
"loss": 0.7887, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 1.4396135265700483, |
|
"grad_norm": 0.582086980342865, |
|
"learning_rate": 6.225991114739622e-06, |
|
"loss": 0.8111, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 1.4444444444444444, |
|
"grad_norm": 0.5153341889381409, |
|
"learning_rate": 6.1986807068678926e-06, |
|
"loss": 0.8097, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 1.4492753623188406, |
|
"grad_norm": 0.5599061846733093, |
|
"learning_rate": 6.171332303360411e-06, |
|
"loss": 0.7631, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.4541062801932367, |
|
"grad_norm": 0.5631583333015442, |
|
"learning_rate": 6.143946771103561e-06, |
|
"loss": 0.7744, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 1.458937198067633, |
|
"grad_norm": 0.6205547451972961, |
|
"learning_rate": 6.11652497816062e-06, |
|
"loss": 0.8067, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 1.463768115942029, |
|
"grad_norm": 0.5770376920700073, |
|
"learning_rate": 6.089067793744258e-06, |
|
"loss": 0.8072, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 1.4685990338164252, |
|
"grad_norm": 0.5772625803947449, |
|
"learning_rate": 6.061576088188981e-06, |
|
"loss": 0.7799, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 1.4734299516908211, |
|
"grad_norm": 0.6021287441253662, |
|
"learning_rate": 6.034050732923538e-06, |
|
"loss": 0.8046, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 1.4782608695652173, |
|
"grad_norm": 0.5101819634437561, |
|
"learning_rate": 6.006492600443301e-06, |
|
"loss": 0.7904, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 1.4830917874396135, |
|
"grad_norm": 0.5455768704414368, |
|
"learning_rate": 5.978902564282616e-06, |
|
"loss": 0.8112, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 1.4879227053140096, |
|
"grad_norm": 0.5585688352584839, |
|
"learning_rate": 5.951281498987106e-06, |
|
"loss": 0.778, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 1.4927536231884058, |
|
"grad_norm": 0.5855573415756226, |
|
"learning_rate": 5.923630280085948e-06, |
|
"loss": 0.7856, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 1.497584541062802, |
|
"grad_norm": 0.6162655353546143, |
|
"learning_rate": 5.895949784064126e-06, |
|
"loss": 0.7649, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.502415458937198, |
|
"grad_norm": 0.5619444847106934, |
|
"learning_rate": 5.8682408883346535e-06, |
|
"loss": 0.7734, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 1.5072463768115942, |
|
"grad_norm": 0.6392289400100708, |
|
"learning_rate": 5.840504471210742e-06, |
|
"loss": 0.8006, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 1.5120772946859904, |
|
"grad_norm": 0.5445345044136047, |
|
"learning_rate": 5.8127414118779825e-06, |
|
"loss": 0.807, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 1.5169082125603865, |
|
"grad_norm": 0.5203850865364075, |
|
"learning_rate": 5.7849525903664636e-06, |
|
"loss": 0.8264, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 1.5217391304347827, |
|
"grad_norm": 0.5838399529457092, |
|
"learning_rate": 5.757138887522884e-06, |
|
"loss": 0.8093, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 1.5265700483091789, |
|
"grad_norm": 0.4624084234237671, |
|
"learning_rate": 5.729301184982622e-06, |
|
"loss": 0.8046, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 1.531400966183575, |
|
"grad_norm": 0.5060942769050598, |
|
"learning_rate": 5.701440365141799e-06, |
|
"loss": 0.805, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 1.5362318840579712, |
|
"grad_norm": 0.5092195272445679, |
|
"learning_rate": 5.673557311129306e-06, |
|
"loss": 0.778, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 1.541062801932367, |
|
"grad_norm": 0.6425656080245972, |
|
"learning_rate": 5.645652906778808e-06, |
|
"loss": 0.7599, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 1.5458937198067633, |
|
"grad_norm": 0.6229380965232849, |
|
"learning_rate": 5.617728036600734e-06, |
|
"loss": 0.8282, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.5507246376811594, |
|
"grad_norm": 0.6080556511878967, |
|
"learning_rate": 5.5897835857542315e-06, |
|
"loss": 0.7854, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 1.5555555555555556, |
|
"grad_norm": 0.5633357167243958, |
|
"learning_rate": 5.561820440019117e-06, |
|
"loss": 0.7987, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 1.5603864734299517, |
|
"grad_norm": 0.62427818775177, |
|
"learning_rate": 5.533839485767795e-06, |
|
"loss": 0.8051, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 1.5652173913043477, |
|
"grad_norm": 0.5080029964447021, |
|
"learning_rate": 5.505841609937162e-06, |
|
"loss": 0.8136, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 1.5700483091787438, |
|
"grad_norm": 0.5406042337417603, |
|
"learning_rate": 5.477827700000492e-06, |
|
"loss": 0.8322, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.57487922705314, |
|
"grad_norm": 0.6476158499717712, |
|
"learning_rate": 5.449798643939305e-06, |
|
"loss": 0.8038, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 1.5797101449275361, |
|
"grad_norm": 0.6692044734954834, |
|
"learning_rate": 5.421755330215223e-06, |
|
"loss": 0.8115, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 1.5845410628019323, |
|
"grad_norm": 0.4540793299674988, |
|
"learning_rate": 5.39369864774181e-06, |
|
"loss": 0.7707, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 1.5893719806763285, |
|
"grad_norm": 0.683095395565033, |
|
"learning_rate": 5.365629485856381e-06, |
|
"loss": 0.8036, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 1.5942028985507246, |
|
"grad_norm": 0.6176734566688538, |
|
"learning_rate": 5.337548734291827e-06, |
|
"loss": 0.784, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.5990338164251208, |
|
"grad_norm": 0.5553250908851624, |
|
"learning_rate": 5.30945728314841e-06, |
|
"loss": 0.766, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 1.603864734299517, |
|
"grad_norm": 0.5356951951980591, |
|
"learning_rate": 5.281356022865542e-06, |
|
"loss": 0.7747, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 1.608695652173913, |
|
"grad_norm": 0.5851414203643799, |
|
"learning_rate": 5.253245844193564e-06, |
|
"loss": 0.7913, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 1.6135265700483092, |
|
"grad_norm": 0.4888221025466919, |
|
"learning_rate": 5.225127638165514e-06, |
|
"loss": 0.8018, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 1.6183574879227054, |
|
"grad_norm": 0.526310384273529, |
|
"learning_rate": 5.197002296068878e-06, |
|
"loss": 0.7714, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 1.6231884057971016, |
|
"grad_norm": 0.5102785229682922, |
|
"learning_rate": 5.168870709417342e-06, |
|
"loss": 0.7642, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.6280193236714977, |
|
"grad_norm": 0.602107048034668, |
|
"learning_rate": 5.140733769922525e-06, |
|
"loss": 0.755, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 1.6328502415458939, |
|
"grad_norm": 0.6300762295722961, |
|
"learning_rate": 5.112592369465731e-06, |
|
"loss": 0.8213, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 1.6376811594202898, |
|
"grad_norm": 0.5401408076286316, |
|
"learning_rate": 5.084447400069656e-06, |
|
"loss": 0.8322, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 1.642512077294686, |
|
"grad_norm": 0.6203901767730713, |
|
"learning_rate": 5.0562997538701295e-06, |
|
"loss": 0.7944, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.6473429951690821, |
|
"grad_norm": 0.6917441487312317, |
|
"learning_rate": 5.0281503230878304e-06, |
|
"loss": 0.7778, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 1.6521739130434783, |
|
"grad_norm": 0.4909766614437103, |
|
"learning_rate": 5e-06, |
|
"loss": 0.7992, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 1.6570048309178744, |
|
"grad_norm": 0.6045981049537659, |
|
"learning_rate": 4.971849676912172e-06, |
|
"loss": 0.8009, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 1.6618357487922706, |
|
"grad_norm": 0.7382987141609192, |
|
"learning_rate": 4.943700246129871e-06, |
|
"loss": 0.7973, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 1.6666666666666665, |
|
"grad_norm": 0.6006978154182434, |
|
"learning_rate": 4.915552599930345e-06, |
|
"loss": 0.7998, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.6714975845410627, |
|
"grad_norm": 0.6009412407875061, |
|
"learning_rate": 4.887407630534271e-06, |
|
"loss": 0.8081, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 1.6763285024154588, |
|
"grad_norm": 0.592892050743103, |
|
"learning_rate": 4.859266230077474e-06, |
|
"loss": 0.7681, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 1.681159420289855, |
|
"grad_norm": 0.6504932641983032, |
|
"learning_rate": 4.83112929058266e-06, |
|
"loss": 0.7812, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 1.6859903381642511, |
|
"grad_norm": 0.7033050656318665, |
|
"learning_rate": 4.802997703931124e-06, |
|
"loss": 0.8025, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 1.6908212560386473, |
|
"grad_norm": 0.6602729558944702, |
|
"learning_rate": 4.7748723618344865e-06, |
|
"loss": 0.7681, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.6956521739130435, |
|
"grad_norm": 0.5453357696533203, |
|
"learning_rate": 4.746754155806437e-06, |
|
"loss": 0.7494, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 1.7004830917874396, |
|
"grad_norm": 0.5276068449020386, |
|
"learning_rate": 4.71864397713446e-06, |
|
"loss": 0.7774, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 1.7053140096618358, |
|
"grad_norm": 0.6807137131690979, |
|
"learning_rate": 4.6905427168515914e-06, |
|
"loss": 0.7866, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 1.710144927536232, |
|
"grad_norm": 0.7194874882698059, |
|
"learning_rate": 4.662451265708174e-06, |
|
"loss": 0.7877, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 1.714975845410628, |
|
"grad_norm": 0.5802371501922607, |
|
"learning_rate": 4.63437051414362e-06, |
|
"loss": 0.7838, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 1.7198067632850242, |
|
"grad_norm": 0.5806293487548828, |
|
"learning_rate": 4.606301352258192e-06, |
|
"loss": 0.7698, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 1.7246376811594204, |
|
"grad_norm": 0.6043258309364319, |
|
"learning_rate": 4.5782446697847775e-06, |
|
"loss": 0.7921, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 1.7294685990338166, |
|
"grad_norm": 0.6771955490112305, |
|
"learning_rate": 4.550201356060695e-06, |
|
"loss": 0.7832, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 1.7342995169082127, |
|
"grad_norm": 0.5330072045326233, |
|
"learning_rate": 4.52217229999951e-06, |
|
"loss": 0.7826, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 1.7391304347826086, |
|
"grad_norm": 0.5096263885498047, |
|
"learning_rate": 4.49415839006284e-06, |
|
"loss": 0.7976, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.7439613526570048, |
|
"grad_norm": 0.6540196537971497, |
|
"learning_rate": 4.466160514232206e-06, |
|
"loss": 0.8119, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 1.748792270531401, |
|
"grad_norm": 0.6180545687675476, |
|
"learning_rate": 4.438179559980885e-06, |
|
"loss": 0.8178, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 1.7536231884057971, |
|
"grad_norm": 0.4621651768684387, |
|
"learning_rate": 4.410216414245771e-06, |
|
"loss": 0.7464, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 1.7584541062801933, |
|
"grad_norm": 0.543982744216919, |
|
"learning_rate": 4.382271963399268e-06, |
|
"loss": 0.8122, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 1.7632850241545892, |
|
"grad_norm": 0.5957825779914856, |
|
"learning_rate": 4.354347093221194e-06, |
|
"loss": 0.7915, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.7681159420289854, |
|
"grad_norm": 0.5544920563697815, |
|
"learning_rate": 4.326442688870697e-06, |
|
"loss": 0.7982, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 1.7729468599033815, |
|
"grad_norm": 0.4633856415748596, |
|
"learning_rate": 4.298559634858202e-06, |
|
"loss": 0.8165, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 1.7777777777777777, |
|
"grad_norm": 0.538975179195404, |
|
"learning_rate": 4.270698815017379e-06, |
|
"loss": 0.7815, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 1.7826086956521738, |
|
"grad_norm": 0.5998566746711731, |
|
"learning_rate": 4.2428611124771184e-06, |
|
"loss": 0.8045, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 1.78743961352657, |
|
"grad_norm": 0.5452315211296082, |
|
"learning_rate": 4.2150474096335356e-06, |
|
"loss": 0.7713, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.7922705314009661, |
|
"grad_norm": 0.47987136244773865, |
|
"learning_rate": 4.187258588122019e-06, |
|
"loss": 0.772, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 1.7971014492753623, |
|
"grad_norm": 0.5281115174293518, |
|
"learning_rate": 4.15949552878926e-06, |
|
"loss": 0.7981, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 1.8019323671497585, |
|
"grad_norm": 0.49062639474868774, |
|
"learning_rate": 4.131759111665349e-06, |
|
"loss": 0.763, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 1.8067632850241546, |
|
"grad_norm": 0.512249231338501, |
|
"learning_rate": 4.104050215935875e-06, |
|
"loss": 0.7778, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 1.8115942028985508, |
|
"grad_norm": 0.5328919887542725, |
|
"learning_rate": 4.076369719914055e-06, |
|
"loss": 0.7979, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.816425120772947, |
|
"grad_norm": 0.5143810510635376, |
|
"learning_rate": 4.048718501012895e-06, |
|
"loss": 0.7616, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 1.821256038647343, |
|
"grad_norm": 0.5369464159011841, |
|
"learning_rate": 4.021097435717386e-06, |
|
"loss": 0.7626, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 1.8260869565217392, |
|
"grad_norm": 0.5135899186134338, |
|
"learning_rate": 3.993507399556699e-06, |
|
"loss": 0.7478, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 1.8309178743961354, |
|
"grad_norm": 0.5900195240974426, |
|
"learning_rate": 3.965949267076465e-06, |
|
"loss": 0.8072, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 1.8357487922705316, |
|
"grad_norm": 0.6226658821105957, |
|
"learning_rate": 3.938423911811021e-06, |
|
"loss": 0.7713, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.8405797101449275, |
|
"grad_norm": 0.5240309238433838, |
|
"learning_rate": 3.910932206255742e-06, |
|
"loss": 0.7827, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 1.8454106280193237, |
|
"grad_norm": 0.5698277354240417, |
|
"learning_rate": 3.883475021839382e-06, |
|
"loss": 0.7783, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 1.8502415458937198, |
|
"grad_norm": 0.6924400925636292, |
|
"learning_rate": 3.856053228896442e-06, |
|
"loss": 0.7967, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 1.855072463768116, |
|
"grad_norm": 0.5029117465019226, |
|
"learning_rate": 3.8286676966395895e-06, |
|
"loss": 0.7897, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 1.8599033816425121, |
|
"grad_norm": 0.5228816866874695, |
|
"learning_rate": 3.8013192931321095e-06, |
|
"loss": 0.811, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.864734299516908, |
|
"grad_norm": 0.554703414440155, |
|
"learning_rate": 3.77400888526038e-06, |
|
"loss": 0.8261, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 1.8695652173913042, |
|
"grad_norm": 0.5827821493148804, |
|
"learning_rate": 3.7467373387063973e-06, |
|
"loss": 0.7898, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 1.8743961352657004, |
|
"grad_norm": 0.5374961495399475, |
|
"learning_rate": 3.719505517920342e-06, |
|
"loss": 0.8126, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 1.8792270531400965, |
|
"grad_norm": 0.5197350382804871, |
|
"learning_rate": 3.692314286093167e-06, |
|
"loss": 0.7839, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 1.8840579710144927, |
|
"grad_norm": 0.6360486149787903, |
|
"learning_rate": 3.6651645051292415e-06, |
|
"loss": 0.7879, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.8888888888888888, |
|
"grad_norm": 0.4924606680870056, |
|
"learning_rate": 3.6380570356190346e-06, |
|
"loss": 0.7753, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 1.893719806763285, |
|
"grad_norm": 0.5175696015357971, |
|
"learning_rate": 3.610992736811827e-06, |
|
"loss": 0.8057, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 1.8985507246376812, |
|
"grad_norm": 0.5930569171905518, |
|
"learning_rate": 3.58397246658848e-06, |
|
"loss": 0.7971, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 1.9033816425120773, |
|
"grad_norm": 0.4917478561401367, |
|
"learning_rate": 3.556997081434248e-06, |
|
"loss": 0.7796, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 1.9082125603864735, |
|
"grad_norm": 0.5090054869651794, |
|
"learning_rate": 3.5300674364116173e-06, |
|
"loss": 0.785, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.9130434782608696, |
|
"grad_norm": 0.5114875435829163, |
|
"learning_rate": 3.5031843851332105e-06, |
|
"loss": 0.805, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.9178743961352658, |
|
"grad_norm": 0.5346699953079224, |
|
"learning_rate": 3.476348779734732e-06, |
|
"loss": 0.8004, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 1.922705314009662, |
|
"grad_norm": 0.6212336421012878, |
|
"learning_rate": 3.449561470847947e-06, |
|
"loss": 0.8092, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 1.927536231884058, |
|
"grad_norm": 0.5195980668067932, |
|
"learning_rate": 3.4228233075737225e-06, |
|
"loss": 0.7694, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 1.9323671497584543, |
|
"grad_norm": 0.5516073703765869, |
|
"learning_rate": 3.3961351374551234e-06, |
|
"loss": 0.8211, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.9371980676328504, |
|
"grad_norm": 0.5031560659408569, |
|
"learning_rate": 3.3694978064505258e-06, |
|
"loss": 0.808, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 1.9420289855072463, |
|
"grad_norm": 0.6618154644966125, |
|
"learning_rate": 3.3429121589068213e-06, |
|
"loss": 0.814, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 1.9468599033816425, |
|
"grad_norm": 0.5210357904434204, |
|
"learning_rate": 3.316379037532644e-06, |
|
"loss": 0.8077, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 1.9516908212560387, |
|
"grad_norm": 0.5135445594787598, |
|
"learning_rate": 3.289899283371657e-06, |
|
"loss": 0.7805, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 1.9565217391304348, |
|
"grad_norm": 0.5435628294944763, |
|
"learning_rate": 3.2634737357758994e-06, |
|
"loss": 0.7839, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.961352657004831, |
|
"grad_norm": 0.5716672539710999, |
|
"learning_rate": 3.2371032323791757e-06, |
|
"loss": 0.7948, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 1.966183574879227, |
|
"grad_norm": 0.5514191389083862, |
|
"learning_rate": 3.2107886090705035e-06, |
|
"loss": 0.7613, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 1.971014492753623, |
|
"grad_norm": 0.5380987524986267, |
|
"learning_rate": 3.1845306999676274e-06, |
|
"loss": 0.7799, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 1.9758454106280192, |
|
"grad_norm": 0.5445787906646729, |
|
"learning_rate": 3.158330337390565e-06, |
|
"loss": 0.7906, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 1.9806763285024154, |
|
"grad_norm": 0.5519409775733948, |
|
"learning_rate": 3.132188351835232e-06, |
|
"loss": 0.7932, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.9855072463768115, |
|
"grad_norm": 0.6107297539710999, |
|
"learning_rate": 3.10610557194712e-06, |
|
"loss": 0.8045, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 1.9903381642512077, |
|
"grad_norm": 0.5743387937545776, |
|
"learning_rate": 3.080082824495024e-06, |
|
"loss": 0.8321, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 1.9951690821256038, |
|
"grad_norm": 0.5688542723655701, |
|
"learning_rate": 3.0541209343448373e-06, |
|
"loss": 0.7812, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.5131503343582153, |
|
"learning_rate": 3.0282207244334084e-06, |
|
"loss": 0.772, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 2.004830917874396, |
|
"grad_norm": 0.5330387949943542, |
|
"learning_rate": 3.0023830157424504e-06, |
|
"loss": 0.7047, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 2.0096618357487923, |
|
"grad_norm": 0.6632731556892395, |
|
"learning_rate": 2.97660862727252e-06, |
|
"loss": 0.7167, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 2.0144927536231885, |
|
"grad_norm": 0.529600203037262, |
|
"learning_rate": 2.950898376017064e-06, |
|
"loss": 0.7295, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 2.0193236714975846, |
|
"grad_norm": 0.6404242515563965, |
|
"learning_rate": 2.9252530769365053e-06, |
|
"loss": 0.7641, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 2.024154589371981, |
|
"grad_norm": 0.5049023628234863, |
|
"learning_rate": 2.8996735429324256e-06, |
|
"loss": 0.7486, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 2.028985507246377, |
|
"grad_norm": 0.5269747376441956, |
|
"learning_rate": 2.874160584821798e-06, |
|
"loss": 0.784, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.033816425120773, |
|
"grad_norm": 0.5848202109336853, |
|
"learning_rate": 2.848715011311271e-06, |
|
"loss": 0.7328, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 2.0386473429951693, |
|
"grad_norm": 0.6017028093338013, |
|
"learning_rate": 2.823337628971551e-06, |
|
"loss": 0.81, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 2.0434782608695654, |
|
"grad_norm": 0.5388191938400269, |
|
"learning_rate": 2.7980292422118282e-06, |
|
"loss": 0.7456, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 2.0483091787439616, |
|
"grad_norm": 0.4810311496257782, |
|
"learning_rate": 2.7727906532542783e-06, |
|
"loss": 0.7718, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 2.0531400966183573, |
|
"grad_norm": 0.519432008266449, |
|
"learning_rate": 2.7476226621086354e-06, |
|
"loss": 0.7562, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 2.0579710144927534, |
|
"grad_norm": 0.4953659474849701, |
|
"learning_rate": 2.72252606654683e-06, |
|
"loss": 0.7303, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 2.0628019323671496, |
|
"grad_norm": 0.5011581182479858, |
|
"learning_rate": 2.697501662077707e-06, |
|
"loss": 0.7589, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 2.0676328502415457, |
|
"grad_norm": 0.585135817527771, |
|
"learning_rate": 2.6725502419218084e-06, |
|
"loss": 0.7573, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 2.072463768115942, |
|
"grad_norm": 0.5075230002403259, |
|
"learning_rate": 2.6476725969862227e-06, |
|
"loss": 0.7592, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 2.077294685990338, |
|
"grad_norm": 0.45230966806411743, |
|
"learning_rate": 2.622869515839524e-06, |
|
"loss": 0.7663, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.082125603864734, |
|
"grad_norm": 0.4765617549419403, |
|
"learning_rate": 2.5981417846867753e-06, |
|
"loss": 0.726, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 2.0869565217391304, |
|
"grad_norm": 0.5006648898124695, |
|
"learning_rate": 2.573490187344596e-06, |
|
"loss": 0.7582, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 2.0917874396135265, |
|
"grad_norm": 0.48479005694389343, |
|
"learning_rate": 2.548915505216333e-06, |
|
"loss": 0.7271, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 2.0966183574879227, |
|
"grad_norm": 0.4033167064189911, |
|
"learning_rate": 2.524418517267283e-06, |
|
"loss": 0.7936, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 2.101449275362319, |
|
"grad_norm": 0.452640563249588, |
|
"learning_rate": 2.5000000000000015e-06, |
|
"loss": 0.7356, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 2.106280193236715, |
|
"grad_norm": 0.4914138913154602, |
|
"learning_rate": 2.4756607274296844e-06, |
|
"loss": 0.7526, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 2.111111111111111, |
|
"grad_norm": 0.4990180432796478, |
|
"learning_rate": 2.4514014710596467e-06, |
|
"loss": 0.7147, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 2.1159420289855073, |
|
"grad_norm": 0.46856606006622314, |
|
"learning_rate": 2.4272229998568576e-06, |
|
"loss": 0.7559, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 2.1207729468599035, |
|
"grad_norm": 0.48758986592292786, |
|
"learning_rate": 2.4031260802275623e-06, |
|
"loss": 0.7652, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 2.1256038647342996, |
|
"grad_norm": 0.4734245538711548, |
|
"learning_rate": 2.3791114759930013e-06, |
|
"loss": 0.7335, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.130434782608696, |
|
"grad_norm": 0.45640215277671814, |
|
"learning_rate": 2.3551799483651894e-06, |
|
"loss": 0.7535, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 2.135265700483092, |
|
"grad_norm": 0.4912705719470978, |
|
"learning_rate": 2.331332255922784e-06, |
|
"loss": 0.7503, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 2.140096618357488, |
|
"grad_norm": 0.46668151021003723, |
|
"learning_rate": 2.307569154587056e-06, |
|
"loss": 0.7335, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 2.1449275362318843, |
|
"grad_norm": 0.5260794162750244, |
|
"learning_rate": 2.283891397597908e-06, |
|
"loss": 0.7048, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 2.14975845410628, |
|
"grad_norm": 0.46279335021972656, |
|
"learning_rate": 2.2602997354900075e-06, |
|
"loss": 0.7288, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 2.154589371980676, |
|
"grad_norm": 0.464158296585083, |
|
"learning_rate": 2.236794916069007e-06, |
|
"loss": 0.7249, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 2.1594202898550723, |
|
"grad_norm": 0.46745479106903076, |
|
"learning_rate": 2.2133776843878185e-06, |
|
"loss": 0.7094, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 2.1642512077294684, |
|
"grad_norm": 0.4427666664123535, |
|
"learning_rate": 2.190048782723015e-06, |
|
"loss": 0.7349, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 2.1690821256038646, |
|
"grad_norm": 0.44838133454322815, |
|
"learning_rate": 2.166808950551296e-06, |
|
"loss": 0.7258, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 2.1739130434782608, |
|
"grad_norm": 0.49869880080223083, |
|
"learning_rate": 2.1436589245260375e-06, |
|
"loss": 0.7153, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.178743961352657, |
|
"grad_norm": 0.46117615699768066, |
|
"learning_rate": 2.120599438453968e-06, |
|
"loss": 0.7267, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 2.183574879227053, |
|
"grad_norm": 0.46450984477996826, |
|
"learning_rate": 2.0976312232718763e-06, |
|
"loss": 0.7158, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 2.1884057971014492, |
|
"grad_norm": 0.45363008975982666, |
|
"learning_rate": 2.074755007023461e-06, |
|
"loss": 0.7507, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 2.1932367149758454, |
|
"grad_norm": 0.48838719725608826, |
|
"learning_rate": 2.0519715148362585e-06, |
|
"loss": 0.7438, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 2.1980676328502415, |
|
"grad_norm": 0.5061821341514587, |
|
"learning_rate": 2.0292814688986375e-06, |
|
"loss": 0.7559, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 2.2028985507246377, |
|
"grad_norm": 0.4787469208240509, |
|
"learning_rate": 2.0066855884369246e-06, |
|
"loss": 0.7813, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 2.207729468599034, |
|
"grad_norm": 0.5006588697433472, |
|
"learning_rate": 1.9841845896926022e-06, |
|
"loss": 0.7688, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 2.21256038647343, |
|
"grad_norm": 0.4781847596168518, |
|
"learning_rate": 1.961779185899597e-06, |
|
"loss": 0.7436, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 2.217391304347826, |
|
"grad_norm": 0.4518663287162781, |
|
"learning_rate": 1.9394700872616856e-06, |
|
"loss": 0.736, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 2.2222222222222223, |
|
"grad_norm": 0.4903218746185303, |
|
"learning_rate": 1.9172580009299735e-06, |
|
"loss": 0.7169, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.2270531400966185, |
|
"grad_norm": 0.49644899368286133, |
|
"learning_rate": 1.8951436309804766e-06, |
|
"loss": 0.7392, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 2.2318840579710146, |
|
"grad_norm": 0.4554813802242279, |
|
"learning_rate": 1.8731276783918162e-06, |
|
"loss": 0.7329, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 2.236714975845411, |
|
"grad_norm": 0.42116937041282654, |
|
"learning_rate": 1.8512108410229878e-06, |
|
"loss": 0.7632, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 2.241545893719807, |
|
"grad_norm": 0.47068312764167786, |
|
"learning_rate": 1.8293938135912475e-06, |
|
"loss": 0.7631, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 2.246376811594203, |
|
"grad_norm": 0.4414040446281433, |
|
"learning_rate": 1.8076772876500831e-06, |
|
"loss": 0.7638, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 2.2512077294685993, |
|
"grad_norm": 0.44085627794265747, |
|
"learning_rate": 1.7860619515673034e-06, |
|
"loss": 0.7597, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 2.2560386473429954, |
|
"grad_norm": 0.41587820649147034, |
|
"learning_rate": 1.7645484905032129e-06, |
|
"loss": 0.723, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 2.260869565217391, |
|
"grad_norm": 0.4661046266555786, |
|
"learning_rate": 1.74313758638889e-06, |
|
"loss": 0.7493, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 2.2657004830917873, |
|
"grad_norm": 0.4823686182498932, |
|
"learning_rate": 1.7218299179045789e-06, |
|
"loss": 0.7505, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 2.2705314009661834, |
|
"grad_norm": 0.4889514744281769, |
|
"learning_rate": 1.7006261604581725e-06, |
|
"loss": 0.7372, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.2753623188405796, |
|
"grad_norm": 0.4301418960094452, |
|
"learning_rate": 1.6795269861638041e-06, |
|
"loss": 0.7674, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 2.2801932367149758, |
|
"grad_norm": 0.45111072063446045, |
|
"learning_rate": 1.6585330638205454e-06, |
|
"loss": 0.7547, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 2.285024154589372, |
|
"grad_norm": 0.46377280354499817, |
|
"learning_rate": 1.6376450588911985e-06, |
|
"loss": 0.7286, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 2.289855072463768, |
|
"grad_norm": 0.4372880458831787, |
|
"learning_rate": 1.6168636334812126e-06, |
|
"loss": 0.7425, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 2.2946859903381642, |
|
"grad_norm": 0.40830719470977783, |
|
"learning_rate": 1.5961894463176942e-06, |
|
"loss": 0.7568, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 2.2995169082125604, |
|
"grad_norm": 0.4759189784526825, |
|
"learning_rate": 1.5756231527285181e-06, |
|
"loss": 0.7517, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 2.3043478260869565, |
|
"grad_norm": 0.4460543990135193, |
|
"learning_rate": 1.555165404621567e-06, |
|
"loss": 0.7401, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 2.3091787439613527, |
|
"grad_norm": 0.43344846367836, |
|
"learning_rate": 1.5348168504640631e-06, |
|
"loss": 0.7509, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 2.314009661835749, |
|
"grad_norm": 0.4556834399700165, |
|
"learning_rate": 1.5145781352620054e-06, |
|
"loss": 0.7514, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 2.318840579710145, |
|
"grad_norm": 0.4389399290084839, |
|
"learning_rate": 1.4944499005397372e-06, |
|
"loss": 0.7228, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.323671497584541, |
|
"grad_norm": 0.4609167277812958, |
|
"learning_rate": 1.4744327843196043e-06, |
|
"loss": 0.7593, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 2.3285024154589373, |
|
"grad_norm": 0.494718462228775, |
|
"learning_rate": 1.4545274211017264e-06, |
|
"loss": 0.7569, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 2.3333333333333335, |
|
"grad_norm": 0.4857819974422455, |
|
"learning_rate": 1.434734441843899e-06, |
|
"loss": 0.7721, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 2.3381642512077296, |
|
"grad_norm": 0.4424082040786743, |
|
"learning_rate": 1.4150544739415755e-06, |
|
"loss": 0.7408, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 2.342995169082126, |
|
"grad_norm": 0.48782283067703247, |
|
"learning_rate": 1.3954881412079945e-06, |
|
"loss": 0.7648, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 2.3478260869565215, |
|
"grad_norm": 0.45818960666656494, |
|
"learning_rate": 1.3760360638544012e-06, |
|
"loss": 0.7492, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 2.3526570048309177, |
|
"grad_norm": 0.49244335293769836, |
|
"learning_rate": 1.3566988584703817e-06, |
|
"loss": 0.7357, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 2.357487922705314, |
|
"grad_norm": 0.4363764822483063, |
|
"learning_rate": 1.3374771380043306e-06, |
|
"loss": 0.7283, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 2.36231884057971, |
|
"grad_norm": 0.44438326358795166, |
|
"learning_rate": 1.3183715117440143e-06, |
|
"loss": 0.7493, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 2.367149758454106, |
|
"grad_norm": 0.43266430497169495, |
|
"learning_rate": 1.2993825852972559e-06, |
|
"loss": 0.7782, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.3719806763285023, |
|
"grad_norm": 0.4312337338924408, |
|
"learning_rate": 1.280510960572745e-06, |
|
"loss": 0.7685, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 2.3768115942028984, |
|
"grad_norm": 0.4180818498134613, |
|
"learning_rate": 1.2617572357609565e-06, |
|
"loss": 0.7557, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 2.3816425120772946, |
|
"grad_norm": 0.4240153431892395, |
|
"learning_rate": 1.2431220053151832e-06, |
|
"loss": 0.746, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 2.3864734299516908, |
|
"grad_norm": 0.47555652260780334, |
|
"learning_rate": 1.2246058599327021e-06, |
|
"loss": 0.7249, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 2.391304347826087, |
|
"grad_norm": 0.4910667836666107, |
|
"learning_rate": 1.2062093865360458e-06, |
|
"loss": 0.7812, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 2.396135265700483, |
|
"grad_norm": 0.44273853302001953, |
|
"learning_rate": 1.1879331682543972e-06, |
|
"loss": 0.7552, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 2.4009661835748792, |
|
"grad_norm": 0.4478420913219452, |
|
"learning_rate": 1.1697777844051105e-06, |
|
"loss": 0.7505, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 2.4057971014492754, |
|
"grad_norm": 0.4299091398715973, |
|
"learning_rate": 1.1517438104753386e-06, |
|
"loss": 0.7401, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 2.4106280193236715, |
|
"grad_norm": 0.4194658696651459, |
|
"learning_rate": 1.1338318181038037e-06, |
|
"loss": 0.7372, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 2.4154589371980677, |
|
"grad_norm": 0.44260597229003906, |
|
"learning_rate": 1.1160423750626693e-06, |
|
"loss": 0.7408, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.420289855072464, |
|
"grad_norm": 0.48269572854042053, |
|
"learning_rate": 1.0983760452395415e-06, |
|
"loss": 0.7874, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 2.42512077294686, |
|
"grad_norm": 0.4460936188697815, |
|
"learning_rate": 1.0808333886196038e-06, |
|
"loss": 0.7575, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 2.429951690821256, |
|
"grad_norm": 0.42938852310180664, |
|
"learning_rate": 1.063414961267859e-06, |
|
"loss": 0.7842, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 2.4347826086956523, |
|
"grad_norm": 0.4390149712562561, |
|
"learning_rate": 1.046121315311508e-06, |
|
"loss": 0.753, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 2.4396135265700485, |
|
"grad_norm": 0.42639121413230896, |
|
"learning_rate": 1.02895299892244e-06, |
|
"loss": 0.7587, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 2.4444444444444446, |
|
"grad_norm": 0.4300375282764435, |
|
"learning_rate": 1.01191055629987e-06, |
|
"loss": 0.7457, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 2.449275362318841, |
|
"grad_norm": 0.45614680647850037, |
|
"learning_rate": 9.949945276530782e-07, |
|
"loss": 0.7336, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 2.454106280193237, |
|
"grad_norm": 0.43812501430511475, |
|
"learning_rate": 9.782054491842879e-07, |
|
"loss": 0.7686, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 2.4589371980676327, |
|
"grad_norm": 0.4528813064098358, |
|
"learning_rate": 9.615438530716753e-07, |
|
"loss": 0.71, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 2.463768115942029, |
|
"grad_norm": 0.4434235394001007, |
|
"learning_rate": 9.450102674524952e-07, |
|
"loss": 0.6996, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.468599033816425, |
|
"grad_norm": 0.42420923709869385, |
|
"learning_rate": 9.286052164063369e-07, |
|
"loss": 0.7661, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 2.473429951690821, |
|
"grad_norm": 0.4244982600212097, |
|
"learning_rate": 9.123292199385247e-07, |
|
"loss": 0.7905, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 2.4782608695652173, |
|
"grad_norm": 0.45026880502700806, |
|
"learning_rate": 8.961827939636198e-07, |
|
"loss": 0.7306, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 2.4830917874396135, |
|
"grad_norm": 0.4042074978351593, |
|
"learning_rate": 8.801664502890722e-07, |
|
"loss": 0.757, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 2.4879227053140096, |
|
"grad_norm": 0.41660401225090027, |
|
"learning_rate": 8.64280696599008e-07, |
|
"loss": 0.7432, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 2.4927536231884058, |
|
"grad_norm": 0.42759013175964355, |
|
"learning_rate": 8.485260364381187e-07, |
|
"loss": 0.7641, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 2.497584541062802, |
|
"grad_norm": 0.40827131271362305, |
|
"learning_rate": 8.329029691957124e-07, |
|
"loss": 0.7504, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 2.502415458937198, |
|
"grad_norm": 0.41131073236465454, |
|
"learning_rate": 8.17411990089883e-07, |
|
"loss": 0.7909, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 2.5072463768115942, |
|
"grad_norm": 0.4378702640533447, |
|
"learning_rate": 8.02053590151805e-07, |
|
"loss": 0.753, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 2.5120772946859904, |
|
"grad_norm": 0.4602629542350769, |
|
"learning_rate": 7.868282562101819e-07, |
|
"loss": 0.7367, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.5169082125603865, |
|
"grad_norm": 0.4568740129470825, |
|
"learning_rate": 7.717364708758024e-07, |
|
"loss": 0.7604, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 2.5217391304347827, |
|
"grad_norm": 0.46792659163475037, |
|
"learning_rate": 7.567787125262449e-07, |
|
"loss": 0.7315, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 2.526570048309179, |
|
"grad_norm": 0.4383016526699066, |
|
"learning_rate": 7.41955455290726e-07, |
|
"loss": 0.7408, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 2.531400966183575, |
|
"grad_norm": 0.40000349283218384, |
|
"learning_rate": 7.27267169035053e-07, |
|
"loss": 0.732, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 2.536231884057971, |
|
"grad_norm": 0.3945160508155823, |
|
"learning_rate": 7.127143193467445e-07, |
|
"loss": 0.7751, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 2.541062801932367, |
|
"grad_norm": 0.4534020721912384, |
|
"learning_rate": 6.982973675202676e-07, |
|
"loss": 0.7625, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 2.545893719806763, |
|
"grad_norm": 0.3865754008293152, |
|
"learning_rate": 6.840167705424106e-07, |
|
"loss": 0.7565, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 2.550724637681159, |
|
"grad_norm": 0.441019207239151, |
|
"learning_rate": 6.698729810778065e-07, |
|
"loss": 0.7452, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 2.5555555555555554, |
|
"grad_norm": 0.42603588104248047, |
|
"learning_rate": 6.558664474545817e-07, |
|
"loss": 0.7297, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 2.5603864734299515, |
|
"grad_norm": 0.4300025999546051, |
|
"learning_rate": 6.419976136501377e-07, |
|
"loss": 0.7457, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.5652173913043477, |
|
"grad_norm": 0.48573917150497437, |
|
"learning_rate": 6.282669192770896e-07, |
|
"loss": 0.7646, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 2.570048309178744, |
|
"grad_norm": 0.4400321841239929, |
|
"learning_rate": 6.146747995693225e-07, |
|
"loss": 0.7085, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 2.57487922705314, |
|
"grad_norm": 0.4423598349094391, |
|
"learning_rate": 6.012216853682001e-07, |
|
"loss": 0.7478, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 2.579710144927536, |
|
"grad_norm": 0.46172380447387695, |
|
"learning_rate": 5.879080031089047e-07, |
|
"loss": 0.7339, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 2.5845410628019323, |
|
"grad_norm": 0.4211689233779907, |
|
"learning_rate": 5.747341748069229e-07, |
|
"loss": 0.774, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 2.5893719806763285, |
|
"grad_norm": 0.4327799379825592, |
|
"learning_rate": 5.617006180446688e-07, |
|
"loss": 0.7417, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 2.5942028985507246, |
|
"grad_norm": 0.4260788559913635, |
|
"learning_rate": 5.488077459582425e-07, |
|
"loss": 0.7593, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 2.5990338164251208, |
|
"grad_norm": 0.44230300188064575, |
|
"learning_rate": 5.360559672243421e-07, |
|
"loss": 0.7384, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 2.603864734299517, |
|
"grad_norm": 0.4378679394721985, |
|
"learning_rate": 5.234456860473042e-07, |
|
"loss": 0.7479, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 2.608695652173913, |
|
"grad_norm": 0.3903830647468567, |
|
"learning_rate": 5.109773021462921e-07, |
|
"loss": 0.7484, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.6135265700483092, |
|
"grad_norm": 0.4425307810306549, |
|
"learning_rate": 4.986512107426283e-07, |
|
"loss": 0.7575, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 2.6183574879227054, |
|
"grad_norm": 0.43397703766822815, |
|
"learning_rate": 4.864678025472635e-07, |
|
"loss": 0.7421, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 2.6231884057971016, |
|
"grad_norm": 0.4102035164833069, |
|
"learning_rate": 4.7442746374839363e-07, |
|
"loss": 0.7684, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 2.6280193236714977, |
|
"grad_norm": 0.4425765872001648, |
|
"learning_rate": 4.625305759992205e-07, |
|
"loss": 0.7295, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 2.632850241545894, |
|
"grad_norm": 0.38346314430236816, |
|
"learning_rate": 4.50777516405847e-07, |
|
"loss": 0.767, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 2.63768115942029, |
|
"grad_norm": 0.44841113686561584, |
|
"learning_rate": 4.3916865751533313e-07, |
|
"loss": 0.7601, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 2.642512077294686, |
|
"grad_norm": 0.42210566997528076, |
|
"learning_rate": 4.2770436730388166e-07, |
|
"loss": 0.7117, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 2.6473429951690823, |
|
"grad_norm": 0.40515097975730896, |
|
"learning_rate": 4.163850091651717e-07, |
|
"loss": 0.7199, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 2.6521739130434785, |
|
"grad_norm": 0.43395963311195374, |
|
"learning_rate": 4.05210941898847e-07, |
|
"loss": 0.7549, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 2.6570048309178746, |
|
"grad_norm": 0.41985607147216797, |
|
"learning_rate": 3.941825196991378e-07, |
|
"loss": 0.7058, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.661835748792271, |
|
"grad_norm": 0.44412142038345337, |
|
"learning_rate": 3.8330009214363197e-07, |
|
"loss": 0.7635, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 2.6666666666666665, |
|
"grad_norm": 0.44575539231300354, |
|
"learning_rate": 3.725640041822026e-07, |
|
"loss": 0.688, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 2.6714975845410627, |
|
"grad_norm": 0.42022770643234253, |
|
"learning_rate": 3.619745961260623e-07, |
|
"loss": 0.776, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 2.676328502415459, |
|
"grad_norm": 0.4149934947490692, |
|
"learning_rate": 3.5153220363698225e-07, |
|
"loss": 0.7615, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 2.681159420289855, |
|
"grad_norm": 0.4015776216983795, |
|
"learning_rate": 3.4123715771665786e-07, |
|
"loss": 0.774, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 2.685990338164251, |
|
"grad_norm": 0.4558965265750885, |
|
"learning_rate": 3.310897846962041e-07, |
|
"loss": 0.6754, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 2.6908212560386473, |
|
"grad_norm": 0.44629231095314026, |
|
"learning_rate": 3.2109040622582186e-07, |
|
"loss": 0.7403, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 2.6956521739130435, |
|
"grad_norm": 0.42704764008522034, |
|
"learning_rate": 3.112393392645985e-07, |
|
"loss": 0.7617, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 2.7004830917874396, |
|
"grad_norm": 0.4057011604309082, |
|
"learning_rate": 3.015368960704584e-07, |
|
"loss": 0.7681, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 2.7053140096618358, |
|
"grad_norm": 0.4393041431903839, |
|
"learning_rate": 2.919833841902714e-07, |
|
"loss": 0.735, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 2.710144927536232, |
|
"grad_norm": 0.44212228059768677, |
|
"learning_rate": 2.8257910645009935e-07, |
|
"loss": 0.7628, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 2.714975845410628, |
|
"grad_norm": 0.4455677270889282, |
|
"learning_rate": 2.733243609455971e-07, |
|
"loss": 0.7659, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 2.7198067632850242, |
|
"grad_norm": 0.4192866384983063, |
|
"learning_rate": 2.6421944103256657e-07, |
|
"loss": 0.7334, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 2.7246376811594204, |
|
"grad_norm": 0.40188413858413696, |
|
"learning_rate": 2.5526463531765467e-07, |
|
"loss": 0.7454, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 2.7294685990338166, |
|
"grad_norm": 0.46752142906188965, |
|
"learning_rate": 2.4646022764920843e-07, |
|
"loss": 0.7245, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 2.7342995169082127, |
|
"grad_norm": 0.43731552362442017, |
|
"learning_rate": 2.3780649710827552e-07, |
|
"loss": 0.7286, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 2.7391304347826084, |
|
"grad_norm": 0.4043439030647278, |
|
"learning_rate": 2.2930371799975593e-07, |
|
"loss": 0.749, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 2.7439613526570046, |
|
"grad_norm": 0.45707106590270996, |
|
"learning_rate": 2.20952159843712e-07, |
|
"loss": 0.7253, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 2.7487922705314007, |
|
"grad_norm": 0.42659294605255127, |
|
"learning_rate": 2.1275208736682262e-07, |
|
"loss": 0.74, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 2.753623188405797, |
|
"grad_norm": 0.4186418652534485, |
|
"learning_rate": 2.0470376049398944e-07, |
|
"loss": 0.7382, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 2.758454106280193, |
|
"grad_norm": 0.4962987005710602, |
|
"learning_rate": 1.9680743434010385e-07, |
|
"loss": 0.74, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 2.763285024154589, |
|
"grad_norm": 0.45507651567459106, |
|
"learning_rate": 1.8906335920195418e-07, |
|
"loss": 0.7158, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 2.7681159420289854, |
|
"grad_norm": 0.4043714702129364, |
|
"learning_rate": 1.814717805502958e-07, |
|
"loss": 0.7602, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 2.7729468599033815, |
|
"grad_norm": 0.4091343581676483, |
|
"learning_rate": 1.7403293902206851e-07, |
|
"loss": 0.7612, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 2.7777777777777777, |
|
"grad_norm": 0.46215108036994934, |
|
"learning_rate": 1.667470704127694e-07, |
|
"loss": 0.7294, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 2.782608695652174, |
|
"grad_norm": 0.4418387711048126, |
|
"learning_rate": 1.5961440566897913e-07, |
|
"loss": 0.7653, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 2.78743961352657, |
|
"grad_norm": 0.4204854667186737, |
|
"learning_rate": 1.5263517088103862e-07, |
|
"loss": 0.7491, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 2.792270531400966, |
|
"grad_norm": 0.42175278067588806, |
|
"learning_rate": 1.4580958727588746e-07, |
|
"loss": 0.7647, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 2.7971014492753623, |
|
"grad_norm": 0.39392223954200745, |
|
"learning_rate": 1.3913787121004717e-07, |
|
"loss": 0.7367, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 2.8019323671497585, |
|
"grad_norm": 0.44153186678886414, |
|
"learning_rate": 1.3262023416276414e-07, |
|
"loss": 0.7382, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 2.8067632850241546, |
|
"grad_norm": 0.42436960339546204, |
|
"learning_rate": 1.2625688272930925e-07, |
|
"loss": 0.7637, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 2.8115942028985508, |
|
"grad_norm": 0.41522523760795593, |
|
"learning_rate": 1.2004801861442373e-07, |
|
"loss": 0.7504, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 2.816425120772947, |
|
"grad_norm": 0.4469909071922302, |
|
"learning_rate": 1.1399383862592928e-07, |
|
"loss": 0.7511, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 2.821256038647343, |
|
"grad_norm": 0.46624404191970825, |
|
"learning_rate": 1.0809453466849029e-07, |
|
"loss": 0.7385, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 2.8260869565217392, |
|
"grad_norm": 0.41857796907424927, |
|
"learning_rate": 1.0235029373752758e-07, |
|
"loss": 0.7778, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 2.8309178743961354, |
|
"grad_norm": 0.42972105741500854, |
|
"learning_rate": 9.676129791329481e-08, |
|
"loss": 0.7589, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 2.8357487922705316, |
|
"grad_norm": 0.40334904193878174, |
|
"learning_rate": 9.132772435510362e-08, |
|
"loss": 0.7381, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 2.8405797101449277, |
|
"grad_norm": 0.4147365689277649, |
|
"learning_rate": 8.604974529571042e-08, |
|
"loss": 0.7363, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 2.845410628019324, |
|
"grad_norm": 0.3909109830856323, |
|
"learning_rate": 8.092752803585513e-08, |
|
"loss": 0.7541, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 2.85024154589372, |
|
"grad_norm": 0.4555034041404724, |
|
"learning_rate": 7.59612349389599e-08, |
|
"loss": 0.7431, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 2.855072463768116, |
|
"grad_norm": 0.4270079433917999, |
|
"learning_rate": 7.115102342598101e-08, |
|
"loss": 0.7417, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 2.8599033816425123, |
|
"grad_norm": 0.41564372181892395, |
|
"learning_rate": 6.649704597042061e-08, |
|
"loss": 0.78, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 2.864734299516908, |
|
"grad_norm": 0.4386310875415802, |
|
"learning_rate": 6.199945009349173e-08, |
|
"loss": 0.735, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 2.869565217391304, |
|
"grad_norm": 0.47602468729019165, |
|
"learning_rate": 5.7658378359443104e-08, |
|
"loss": 0.7267, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 2.8743961352657004, |
|
"grad_norm": 0.41371598839759827, |
|
"learning_rate": 5.3473968371040575e-08, |
|
"loss": 0.7403, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 2.8792270531400965, |
|
"grad_norm": 0.404881089925766, |
|
"learning_rate": 4.944635276520393e-08, |
|
"loss": 0.7199, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 2.8840579710144927, |
|
"grad_norm": 0.4284209609031677, |
|
"learning_rate": 4.55756592088058e-08, |
|
"loss": 0.7354, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 2.888888888888889, |
|
"grad_norm": 0.40044698119163513, |
|
"learning_rate": 4.186201039462046e-08, |
|
"loss": 0.7433, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 2.893719806763285, |
|
"grad_norm": 0.43060269951820374, |
|
"learning_rate": 3.8305524037438035e-08, |
|
"loss": 0.7518, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 2.898550724637681, |
|
"grad_norm": 0.42064300179481506, |
|
"learning_rate": 3.4906312870331973e-08, |
|
"loss": 0.7525, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 2.9033816425120773, |
|
"grad_norm": 0.3958161771297455, |
|
"learning_rate": 3.166448464108629e-08, |
|
"loss": 0.7512, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 2.9082125603864735, |
|
"grad_norm": 0.4121488630771637, |
|
"learning_rate": 2.8580142108778354e-08, |
|
"loss": 0.7784, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 2.9130434782608696, |
|
"grad_norm": 0.4234846234321594, |
|
"learning_rate": 2.5653383040524228e-08, |
|
"loss": 0.7282, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 2.917874396135266, |
|
"grad_norm": 0.4192802608013153, |
|
"learning_rate": 2.2884300208378395e-08, |
|
"loss": 0.7426, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 2.922705314009662, |
|
"grad_norm": 0.38649386167526245, |
|
"learning_rate": 2.0272981386393332e-08, |
|
"loss": 0.7535, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 2.927536231884058, |
|
"grad_norm": 0.39386647939682007, |
|
"learning_rate": 1.781950934783505e-08, |
|
"loss": 0.7877, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 2.9323671497584543, |
|
"grad_norm": 0.41560304164886475, |
|
"learning_rate": 1.552396186256411e-08, |
|
"loss": 0.738, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 2.9371980676328504, |
|
"grad_norm": 0.4333752393722534, |
|
"learning_rate": 1.3386411694565894e-08, |
|
"loss": 0.7643, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 2.942028985507246, |
|
"grad_norm": 0.3939555287361145, |
|
"learning_rate": 1.1406926599646373e-08, |
|
"loss": 0.7404, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 2.9468599033816423, |
|
"grad_norm": 0.4290173649787903, |
|
"learning_rate": 9.585569323284915e-09, |
|
"loss": 0.7406, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.9516908212560384, |
|
"grad_norm": 0.38762733340263367, |
|
"learning_rate": 7.922397598642551e-09, |
|
"loss": 0.7616, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 2.9565217391304346, |
|
"grad_norm": 0.44670021533966064, |
|
"learning_rate": 6.417464144736208e-09, |
|
"loss": 0.7457, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 2.9613526570048307, |
|
"grad_norm": 0.45436298847198486, |
|
"learning_rate": 5.0708166647628345e-09, |
|
"loss": 0.7153, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 2.966183574879227, |
|
"grad_norm": 0.44896742701530457, |
|
"learning_rate": 3.88249784459227e-09, |
|
"loss": 0.7393, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 2.971014492753623, |
|
"grad_norm": 0.4394199848175049, |
|
"learning_rate": 2.8525453514099966e-09, |
|
"loss": 0.7142, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 2.975845410628019, |
|
"grad_norm": 0.38278430700302124, |
|
"learning_rate": 1.980991832524759e-09, |
|
"loss": 0.7544, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 2.9806763285024154, |
|
"grad_norm": 0.403142511844635, |
|
"learning_rate": 1.2678649143349485e-09, |
|
"loss": 0.7158, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 2.9855072463768115, |
|
"grad_norm": 0.39844968914985657, |
|
"learning_rate": 7.131872014509711e-10, |
|
"loss": 0.7541, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 2.9903381642512077, |
|
"grad_norm": 0.4012044370174408, |
|
"learning_rate": 3.1697627597970794e-10, |
|
"loss": 0.7623, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 2.995169082125604, |
|
"grad_norm": 0.47099077701568604, |
|
"learning_rate": 7.924469696718451e-11, |
|
"loss": 0.7556, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 0.4217309057712555, |
|
"learning_rate": 0.0, |
|
"loss": 0.7747, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 621, |
|
"total_flos": 536280491753472.0, |
|
"train_loss": 0.8154453666889725, |
|
"train_runtime": 29892.1358, |
|
"train_samples_per_second": 1.994, |
|
"train_steps_per_second": 0.021 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 621, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 536280491753472.0, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|