|
{ |
|
"best_global_step": null, |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 11920, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0008389525677192026, |
|
"grad_norm": 290.0, |
|
"learning_rate": 3.6000000000000003e-06, |
|
"loss": 6.6219, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0016779051354384052, |
|
"grad_norm": 4.15625, |
|
"learning_rate": 7.600000000000001e-06, |
|
"loss": 1.5478, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0025168577031576076, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 1.16e-05, |
|
"loss": 0.9445, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0033558102708768104, |
|
"grad_norm": 1.3125, |
|
"learning_rate": 1.5600000000000003e-05, |
|
"loss": 0.693, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.004194762838596013, |
|
"grad_norm": 1.3671875, |
|
"learning_rate": 1.9600000000000002e-05, |
|
"loss": 0.6403, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.005033715406315215, |
|
"grad_norm": 1.2578125, |
|
"learning_rate": 2.36e-05, |
|
"loss": 0.6304, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.005872667974034418, |
|
"grad_norm": 1.515625, |
|
"learning_rate": 2.76e-05, |
|
"loss": 0.599, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.006711620541753621, |
|
"grad_norm": 1.28125, |
|
"learning_rate": 3.16e-05, |
|
"loss": 0.591, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.007550573109472823, |
|
"grad_norm": 1.1484375, |
|
"learning_rate": 3.5600000000000005e-05, |
|
"loss": 0.6058, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.008389525677192027, |
|
"grad_norm": 1.3046875, |
|
"learning_rate": 3.96e-05, |
|
"loss": 0.6065, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.009228478244911228, |
|
"grad_norm": 1.15625, |
|
"learning_rate": 4.360000000000001e-05, |
|
"loss": 0.6054, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.01006743081263043, |
|
"grad_norm": 1.2265625, |
|
"learning_rate": 4.7600000000000005e-05, |
|
"loss": 0.6284, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.010906383380349634, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 5.160000000000001e-05, |
|
"loss": 0.6296, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.011745335948068836, |
|
"grad_norm": 1.0390625, |
|
"learning_rate": 5.56e-05, |
|
"loss": 0.534, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.012584288515788038, |
|
"grad_norm": 1.03125, |
|
"learning_rate": 5.9600000000000005e-05, |
|
"loss": 0.5969, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.013423241083507242, |
|
"grad_norm": 0.97265625, |
|
"learning_rate": 6.360000000000001e-05, |
|
"loss": 0.6417, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.014262193651226444, |
|
"grad_norm": 0.796875, |
|
"learning_rate": 6.76e-05, |
|
"loss": 0.6363, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.015101146218945646, |
|
"grad_norm": 1.0078125, |
|
"learning_rate": 7.16e-05, |
|
"loss": 0.6482, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.015940098786664848, |
|
"grad_norm": 0.9140625, |
|
"learning_rate": 7.560000000000001e-05, |
|
"loss": 0.6197, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.016779051354384053, |
|
"grad_norm": 0.83984375, |
|
"learning_rate": 7.960000000000001e-05, |
|
"loss": 0.599, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.017618003922103255, |
|
"grad_norm": 0.85546875, |
|
"learning_rate": 7.993856655290104e-05, |
|
"loss": 0.6171, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.018456956489822457, |
|
"grad_norm": 0.81640625, |
|
"learning_rate": 7.98703071672355e-05, |
|
"loss": 0.621, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.01929590905754166, |
|
"grad_norm": 0.84765625, |
|
"learning_rate": 7.980204778156997e-05, |
|
"loss": 0.5833, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.02013486162526086, |
|
"grad_norm": 0.76953125, |
|
"learning_rate": 7.973378839590445e-05, |
|
"loss": 0.62, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.020973814192980063, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 7.966552901023891e-05, |
|
"loss": 0.6168, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.021812766760699268, |
|
"grad_norm": 0.68359375, |
|
"learning_rate": 7.959726962457338e-05, |
|
"loss": 0.596, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.02265171932841847, |
|
"grad_norm": 0.69140625, |
|
"learning_rate": 7.952901023890786e-05, |
|
"loss": 0.5832, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.023490671896137672, |
|
"grad_norm": 0.7109375, |
|
"learning_rate": 7.946075085324232e-05, |
|
"loss": 0.6071, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.024329624463856874, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 7.93924914675768e-05, |
|
"loss": 0.6342, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.025168577031576076, |
|
"grad_norm": 0.71875, |
|
"learning_rate": 7.932423208191127e-05, |
|
"loss": 0.6141, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.02600752959929528, |
|
"grad_norm": 2.921875, |
|
"learning_rate": 7.925597269624573e-05, |
|
"loss": 0.6573, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.026846482167014483, |
|
"grad_norm": 0.70703125, |
|
"learning_rate": 7.918771331058021e-05, |
|
"loss": 0.5759, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.027685434734733685, |
|
"grad_norm": 0.76953125, |
|
"learning_rate": 7.911945392491469e-05, |
|
"loss": 0.5954, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.028524387302452887, |
|
"grad_norm": 0.67578125, |
|
"learning_rate": 7.905119453924916e-05, |
|
"loss": 0.6044, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.02936333987017209, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 7.898293515358362e-05, |
|
"loss": 0.6609, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.03020229243789129, |
|
"grad_norm": 0.70703125, |
|
"learning_rate": 7.891467576791809e-05, |
|
"loss": 0.6294, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.031041245005610497, |
|
"grad_norm": 0.68359375, |
|
"learning_rate": 7.884641638225257e-05, |
|
"loss": 0.6168, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.031880197573329695, |
|
"grad_norm": 0.64453125, |
|
"learning_rate": 7.877815699658705e-05, |
|
"loss": 0.5928, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.0327191501410489, |
|
"grad_norm": 0.72265625, |
|
"learning_rate": 7.870989761092151e-05, |
|
"loss": 0.6392, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.033558102708768106, |
|
"grad_norm": 0.671875, |
|
"learning_rate": 7.864163822525598e-05, |
|
"loss": 0.595, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.03439705527648731, |
|
"grad_norm": 0.6328125, |
|
"learning_rate": 7.857337883959044e-05, |
|
"loss": 0.5803, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.03523600784420651, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 7.850511945392492e-05, |
|
"loss": 0.5735, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.03607496041192571, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 7.843686006825939e-05, |
|
"loss": 0.5848, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.036913912979644914, |
|
"grad_norm": 0.69921875, |
|
"learning_rate": 7.836860068259385e-05, |
|
"loss": 0.6076, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.037752865547364116, |
|
"grad_norm": 0.6640625, |
|
"learning_rate": 7.830034129692833e-05, |
|
"loss": 0.6163, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.03859181811508332, |
|
"grad_norm": 0.6796875, |
|
"learning_rate": 7.823208191126281e-05, |
|
"loss": 0.598, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.03943077068280252, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 7.816382252559728e-05, |
|
"loss": 0.5121, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.04026972325052172, |
|
"grad_norm": 0.7265625, |
|
"learning_rate": 7.809556313993174e-05, |
|
"loss": 0.5521, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.041108675818240924, |
|
"grad_norm": 0.67578125, |
|
"learning_rate": 7.802730375426621e-05, |
|
"loss": 0.62, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.041947628385960126, |
|
"grad_norm": 0.625, |
|
"learning_rate": 7.795904436860069e-05, |
|
"loss": 0.5939, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.042786580953679335, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 7.789078498293517e-05, |
|
"loss": 0.5646, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.043625533521398537, |
|
"grad_norm": 0.60546875, |
|
"learning_rate": 7.782252559726963e-05, |
|
"loss": 0.595, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.04446448608911774, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 7.77542662116041e-05, |
|
"loss": 0.5661, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.04530343865683694, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 7.768600682593857e-05, |
|
"loss": 0.5619, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.04614239122455614, |
|
"grad_norm": 0.6328125, |
|
"learning_rate": 7.761774744027304e-05, |
|
"loss": 0.5654, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.046981343792275344, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 7.754948805460752e-05, |
|
"loss": 0.5833, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.047820296359994546, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 7.748122866894199e-05, |
|
"loss": 0.5263, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.04865924892771375, |
|
"grad_norm": 0.578125, |
|
"learning_rate": 7.741296928327645e-05, |
|
"loss": 0.582, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.04949820149543295, |
|
"grad_norm": 0.65234375, |
|
"learning_rate": 7.734470989761092e-05, |
|
"loss": 0.5396, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.05033715406315215, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 7.72764505119454e-05, |
|
"loss": 0.5699, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.051176106630871354, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 7.720819112627987e-05, |
|
"loss": 0.5638, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.05201505919859056, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 7.713993174061433e-05, |
|
"loss": 0.5641, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.052854011766309765, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 7.707167235494881e-05, |
|
"loss": 0.5358, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.05369296433402897, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 7.700341296928329e-05, |
|
"loss": 0.565, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.05453191690174817, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 7.693515358361776e-05, |
|
"loss": 0.5771, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.05537086946946737, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 7.686689419795222e-05, |
|
"loss": 0.5628, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.05620982203718657, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 7.679863481228669e-05, |
|
"loss": 0.589, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.057048774604905775, |
|
"grad_norm": 0.63671875, |
|
"learning_rate": 7.673037542662117e-05, |
|
"loss": 0.5531, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.05788772717262498, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 7.666211604095565e-05, |
|
"loss": 0.5415, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.05872667974034418, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 7.659385665529011e-05, |
|
"loss": 0.5729, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.05956563230806338, |
|
"grad_norm": 0.62890625, |
|
"learning_rate": 7.652559726962458e-05, |
|
"loss": 0.5536, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.06040458487578258, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 7.645733788395904e-05, |
|
"loss": 0.5431, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.061243537443501785, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 7.638907849829352e-05, |
|
"loss": 0.5555, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.062082490011220993, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7.6320819112628e-05, |
|
"loss": 0.5444, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.0629214425789402, |
|
"grad_norm": 0.640625, |
|
"learning_rate": 7.625255972696247e-05, |
|
"loss": 0.5285, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.06376039514665939, |
|
"grad_norm": 0.60546875, |
|
"learning_rate": 7.618430034129693e-05, |
|
"loss": 0.5676, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.0645993477143786, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 7.611604095563141e-05, |
|
"loss": 0.5007, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.0654383002820978, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.604778156996588e-05, |
|
"loss": 0.5397, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.066277252849817, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 7.597952218430036e-05, |
|
"loss": 0.5538, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.06711620541753621, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 7.591126279863481e-05, |
|
"loss": 0.5183, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.06795515798525541, |
|
"grad_norm": 0.58984375, |
|
"learning_rate": 7.584300341296929e-05, |
|
"loss": 0.5385, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.06879411055297462, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 7.577474402730377e-05, |
|
"loss": 0.5702, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.06963306312069381, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 7.570648464163823e-05, |
|
"loss": 0.5602, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.07047201568841302, |
|
"grad_norm": 0.6328125, |
|
"learning_rate": 7.56382252559727e-05, |
|
"loss": 0.5444, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.07131096825613222, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 7.556996587030716e-05, |
|
"loss": 0.5568, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.07214992082385142, |
|
"grad_norm": 0.60546875, |
|
"learning_rate": 7.550170648464164e-05, |
|
"loss": 0.5868, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.07298887339157062, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 7.543344709897612e-05, |
|
"loss": 0.5531, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.07382782595928983, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 7.536518771331059e-05, |
|
"loss": 0.5235, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.07466677852700902, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 7.529692832764505e-05, |
|
"loss": 0.5234, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.07550573109472823, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 7.522866894197953e-05, |
|
"loss": 0.4985, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.07634468366244744, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 7.5160409556314e-05, |
|
"loss": 0.5425, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.07718363623016664, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7.509215017064848e-05, |
|
"loss": 0.5343, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.07802258879788584, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 7.502389078498294e-05, |
|
"loss": 0.5259, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.07886154136560504, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 7.495563139931741e-05, |
|
"loss": 0.5176, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.07970049393332425, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 7.488737201365189e-05, |
|
"loss": 0.5314, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.08053944650104344, |
|
"grad_norm": 0.60546875, |
|
"learning_rate": 7.481911262798635e-05, |
|
"loss": 0.5583, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.08137839906876265, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 7.475085324232083e-05, |
|
"loss": 0.5162, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.08221735163648185, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.46825938566553e-05, |
|
"loss": 0.5017, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.08305630420420106, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 7.461433447098976e-05, |
|
"loss": 0.5186, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.08389525677192025, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 7.454607508532424e-05, |
|
"loss": 0.5654, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.08473420933963946, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 7.447781569965871e-05, |
|
"loss": 0.5267, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.08557316190735867, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 7.440955631399318e-05, |
|
"loss": 0.5202, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.08641211447507786, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 7.434129692832765e-05, |
|
"loss": 0.5497, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.08725106704279707, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 7.427303754266212e-05, |
|
"loss": 0.517, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.08809001961051627, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 7.42047781569966e-05, |
|
"loss": 0.5921, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.08892897217823548, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 7.413651877133106e-05, |
|
"loss": 0.5061, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.08976792474595467, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 7.406825938566553e-05, |
|
"loss": 0.5261, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.09060687731367388, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 7.400000000000001e-05, |
|
"loss": 0.5014, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.09144582988139308, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 7.393174061433448e-05, |
|
"loss": 0.5191, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.09228478244911228, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 7.386348122866895e-05, |
|
"loss": 0.5717, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.09312373501683148, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 7.379522184300342e-05, |
|
"loss": 0.5369, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.09396268758455069, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 7.372696245733789e-05, |
|
"loss": 0.5232, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.0948016401522699, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 7.365870307167237e-05, |
|
"loss": 0.5449, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.09564059271998909, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 7.359044368600683e-05, |
|
"loss": 0.537, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.0964795452877083, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7.352218430034131e-05, |
|
"loss": 0.5266, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.0973184978554275, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 7.345392491467578e-05, |
|
"loss": 0.5448, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.0981574504231467, |
|
"grad_norm": 0.7109375, |
|
"learning_rate": 7.338566552901024e-05, |
|
"loss": 0.5589, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.0989964029908659, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 7.331740614334472e-05, |
|
"loss": 0.5174, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.09983535555858511, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.324914675767919e-05, |
|
"loss": 0.5537, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.1006743081263043, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 7.318088737201365e-05, |
|
"loss": 0.5181, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.10151326069402351, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7.311262798634813e-05, |
|
"loss": 0.5371, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.10235221326174271, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 7.30443686006826e-05, |
|
"loss": 0.502, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.10319116582946192, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 7.297610921501708e-05, |
|
"loss": 0.5532, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.10403011839718113, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 7.290784982935154e-05, |
|
"loss": 0.51, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.10486907096490032, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 7.283959044368601e-05, |
|
"loss": 0.5173, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.10570802353261953, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 7.277133105802049e-05, |
|
"loss": 0.5247, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.10654697610033873, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 7.270307167235495e-05, |
|
"loss": 0.5094, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.10738592866805793, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.263481228668943e-05, |
|
"loss": 0.4913, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.10822488123577713, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 7.25665529010239e-05, |
|
"loss": 0.517, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.10906383380349634, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.249829351535836e-05, |
|
"loss": 0.5303, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.10990278637121553, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 7.243003412969284e-05, |
|
"loss": 0.5111, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.11074173893893474, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 7.236177474402731e-05, |
|
"loss": 0.5216, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.11158069150665394, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 7.229351535836179e-05, |
|
"loss": 0.5426, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.11241964407437315, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 7.222525597269625e-05, |
|
"loss": 0.4916, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.11325859664209235, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 7.215699658703072e-05, |
|
"loss": 0.4829, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.11409754920981155, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7.20887372013652e-05, |
|
"loss": 0.5266, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.11493650177753076, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 7.202047781569966e-05, |
|
"loss": 0.5466, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.11577545434524995, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 7.195221843003413e-05, |
|
"loss": 0.5389, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.11661440691296916, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.188395904436861e-05, |
|
"loss": 0.5168, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.11745335948068836, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 7.181569965870307e-05, |
|
"loss": 0.5077, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.11829231204840757, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 7.174744027303755e-05, |
|
"loss": 0.4983, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.11913126461612676, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7.167918088737202e-05, |
|
"loss": 0.4972, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.11997021718384597, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 7.161092150170648e-05, |
|
"loss": 0.5226, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.12080916975156517, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 7.154266211604096e-05, |
|
"loss": 0.503, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.12164812231928437, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 7.147440273037543e-05, |
|
"loss": 0.5348, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.12248707488700357, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 7.140614334470991e-05, |
|
"loss": 0.4928, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.12332602745472278, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 7.133788395904437e-05, |
|
"loss": 0.5024, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.12416498002244199, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7.126962457337884e-05, |
|
"loss": 0.4768, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.12500393259016118, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 7.120136518771332e-05, |
|
"loss": 0.4918, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.1258428851578804, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 7.113310580204779e-05, |
|
"loss": 0.4975, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.1266818377255996, |
|
"grad_norm": 0.62109375, |
|
"learning_rate": 7.106484641638226e-05, |
|
"loss": 0.4952, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.12752079029331878, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 7.099658703071673e-05, |
|
"loss": 0.5037, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.128359742861038, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 7.09283276450512e-05, |
|
"loss": 0.5083, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.1291986954287572, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 7.086006825938567e-05, |
|
"loss": 0.4955, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.1300376479964764, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 7.079180887372014e-05, |
|
"loss": 0.5227, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.1308766005641956, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 7.07235494880546e-05, |
|
"loss": 0.5049, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.1317155531319148, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 7.065529010238909e-05, |
|
"loss": 0.5296, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.132554505699634, |
|
"grad_norm": 0.5, |
|
"learning_rate": 7.058703071672355e-05, |
|
"loss": 0.5114, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.13339345826735322, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 7.051877133105803e-05, |
|
"loss": 0.4872, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.13423241083507242, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 7.04505119453925e-05, |
|
"loss": 0.5246, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.1350713634027916, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 7.038225255972696e-05, |
|
"loss": 0.5132, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.13591031597051081, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 7.031399317406144e-05, |
|
"loss": 0.5115, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.13674926853823002, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7.02457337883959e-05, |
|
"loss": 0.5134, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.13758822110594923, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 7.017747440273039e-05, |
|
"loss": 0.5062, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.1384271736736684, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 7.010921501706485e-05, |
|
"loss": 0.4901, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.13926612624138762, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 7.004095563139932e-05, |
|
"loss": 0.5184, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.14010507880910683, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.99726962457338e-05, |
|
"loss": 0.5085, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.14094403137682604, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.990443686006826e-05, |
|
"loss": 0.4884, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.14178298394454525, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 6.983617747440274e-05, |
|
"loss": 0.5136, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.14262193651226443, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 6.976791808873721e-05, |
|
"loss": 0.4814, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.14346088907998364, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 6.969965870307167e-05, |
|
"loss": 0.5241, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.14429984164770285, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 6.963139931740615e-05, |
|
"loss": 0.4941, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.14513879421542206, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 6.956313993174062e-05, |
|
"loss": 0.4865, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.14597774678314124, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 6.949488054607508e-05, |
|
"loss": 0.4577, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.14681669935086045, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 6.942662116040956e-05, |
|
"loss": 0.4727, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.14765565191857966, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 6.935836177474403e-05, |
|
"loss": 0.4624, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.14849460448629886, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 6.929010238907851e-05, |
|
"loss": 0.477, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.14933355705401805, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 6.922184300341297e-05, |
|
"loss": 0.4987, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.15017250962173725, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 6.915358361774744e-05, |
|
"loss": 0.5018, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.15101146218945646, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 6.908532423208192e-05, |
|
"loss": 0.506, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.15185041475717567, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 6.901706484641638e-05, |
|
"loss": 0.494, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.15268936732489488, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 6.894880546075086e-05, |
|
"loss": 0.5037, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.15352831989261406, |
|
"grad_norm": 0.81640625, |
|
"learning_rate": 6.888054607508533e-05, |
|
"loss": 0.4933, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.15436727246033327, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 6.88122866894198e-05, |
|
"loss": 0.4891, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.15520622502805248, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 6.874402730375427e-05, |
|
"loss": 0.5019, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.1560451775957717, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 6.867576791808875e-05, |
|
"loss": 0.4845, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.15688413016349087, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 6.860750853242322e-05, |
|
"loss": 0.464, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.15772308273121008, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 6.853924914675768e-05, |
|
"loss": 0.4677, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.1585620352989293, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 6.847098976109215e-05, |
|
"loss": 0.462, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.1594009878666485, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 6.840273037542663e-05, |
|
"loss": 0.4755, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.16023994043436768, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 6.83344709897611e-05, |
|
"loss": 0.5176, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.1610788930020869, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 6.826621160409556e-05, |
|
"loss": 0.452, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.1619178455698061, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 6.819795221843004e-05, |
|
"loss": 0.5008, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.1627567981375253, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 6.81296928327645e-05, |
|
"loss": 0.4664, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.1635957507052445, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 6.806143344709898e-05, |
|
"loss": 0.4695, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.1644347032729637, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 6.799317406143345e-05, |
|
"loss": 0.4591, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.1652736558406829, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 6.792491467576792e-05, |
|
"loss": 0.4595, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.1661126084084021, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 6.78566552901024e-05, |
|
"loss": 0.4739, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.16695156097612132, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 6.778839590443687e-05, |
|
"loss": 0.4793, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.1677905135438405, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 6.772013651877134e-05, |
|
"loss": 0.4749, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.1686294661115597, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 6.76518771331058e-05, |
|
"loss": 0.5148, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.16946841867927892, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 6.758361774744027e-05, |
|
"loss": 0.5101, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.17030737124699813, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 6.751535836177475e-05, |
|
"loss": 0.4781, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.17114632381471734, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 6.744709897610923e-05, |
|
"loss": 0.4708, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.17198527638243652, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 6.73788395904437e-05, |
|
"loss": 0.4739, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.17282422895015573, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 6.731058020477816e-05, |
|
"loss": 0.458, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.17366318151787494, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 6.724232081911263e-05, |
|
"loss": 0.4762, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.17450213408559415, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 6.71740614334471e-05, |
|
"loss": 0.4539, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.17534108665331333, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 6.710580204778159e-05, |
|
"loss": 0.4771, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.17618003922103254, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 6.703754266211605e-05, |
|
"loss": 0.478, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.17701899178875175, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 6.696928327645052e-05, |
|
"loss": 0.4498, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.17785794435647095, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 6.690102389078498e-05, |
|
"loss": 0.4836, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.17869689692419014, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 6.683276450511946e-05, |
|
"loss": 0.4617, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.17953584949190934, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 6.676450511945393e-05, |
|
"loss": 0.5011, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.18037480205962855, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 6.669624573378839e-05, |
|
"loss": 0.4577, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.18121375462734776, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.662798634812287e-05, |
|
"loss": 0.4857, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.18205270719506697, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 6.655972696245735e-05, |
|
"loss": 0.4846, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.18289165976278615, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 6.649146757679182e-05, |
|
"loss": 0.4992, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.18373061233050536, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 6.642320819112628e-05, |
|
"loss": 0.452, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.18456956489822457, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 6.635494880546075e-05, |
|
"loss": 0.4598, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.18540851746594378, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 6.628668941979523e-05, |
|
"loss": 0.4405, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.18624747003366296, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 6.62184300341297e-05, |
|
"loss": 0.4606, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.18708642260138217, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 6.615017064846417e-05, |
|
"loss": 0.4765, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.18792537516910138, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 6.608191126279864e-05, |
|
"loss": 0.4773, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.1887643277368206, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 6.60136518771331e-05, |
|
"loss": 0.4657, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.1896032803045398, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 6.594539249146758e-05, |
|
"loss": 0.448, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.19044223287225898, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 6.587713310580206e-05, |
|
"loss": 0.4578, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.19128118543997819, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 6.580887372013653e-05, |
|
"loss": 0.455, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.1921201380076974, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 6.5740614334471e-05, |
|
"loss": 0.4657, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.1929590905754166, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 6.567235494880547e-05, |
|
"loss": 0.4754, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.19379804314313578, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 6.560409556313994e-05, |
|
"loss": 0.4405, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.194636995710855, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 6.55358361774744e-05, |
|
"loss": 0.4605, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.1954759482785742, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 6.546757679180887e-05, |
|
"loss": 0.5133, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.1963149008462934, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 6.539931740614335e-05, |
|
"loss": 0.4318, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.1971538534140126, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 6.533105802047783e-05, |
|
"loss": 0.4743, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.1979928059817318, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 6.52627986348123e-05, |
|
"loss": 0.4667, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.198831758549451, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 6.519453924914676e-05, |
|
"loss": 0.4878, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.19967071111717022, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 6.512627986348123e-05, |
|
"loss": 0.4279, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.20050966368488943, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 6.50580204778157e-05, |
|
"loss": 0.4737, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.2013486162526086, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 6.498976109215018e-05, |
|
"loss": 0.4355, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.20218756882032782, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 6.492150170648465e-05, |
|
"loss": 0.4458, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.20302652138804703, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 6.485324232081912e-05, |
|
"loss": 0.4723, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.20386547395576624, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 6.47849829351536e-05, |
|
"loss": 0.4331, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.20470442652348542, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 6.471672354948806e-05, |
|
"loss": 0.4613, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.20554337909120463, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 6.464846416382254e-05, |
|
"loss": 0.4682, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.20638233165892383, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 6.4580204778157e-05, |
|
"loss": 0.4343, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.20722128422664304, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 6.451194539249147e-05, |
|
"loss": 0.4582, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.20806023679436225, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 6.444368600682595e-05, |
|
"loss": 0.4496, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.20889918936208143, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 6.437542662116042e-05, |
|
"loss": 0.4507, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.20973814192980064, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 6.430716723549488e-05, |
|
"loss": 0.4319, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.21057709449751985, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 6.423890784982935e-05, |
|
"loss": 0.4843, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.21141604706523906, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 6.417064846416383e-05, |
|
"loss": 0.4452, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.21225499963295824, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 6.41023890784983e-05, |
|
"loss": 0.4459, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.21309395220067745, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 6.403412969283277e-05, |
|
"loss": 0.4298, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.21393290476839666, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 6.396587030716724e-05, |
|
"loss": 0.4787, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.21477185733611587, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 6.389761092150172e-05, |
|
"loss": 0.4694, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.21561080990383505, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 6.382935153583618e-05, |
|
"loss": 0.4112, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.21644976247155426, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 6.376109215017066e-05, |
|
"loss": 0.483, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.21728871503927347, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 6.369283276450513e-05, |
|
"loss": 0.4553, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.21812766760699268, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.362457337883959e-05, |
|
"loss": 0.4915, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.21896662017471188, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 6.355631399317407e-05, |
|
"loss": 0.4574, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.21980557274243107, |
|
"grad_norm": 0.5, |
|
"learning_rate": 6.348805460750854e-05, |
|
"loss": 0.4874, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.22064452531015027, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 6.341979522184302e-05, |
|
"loss": 0.4514, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.22148347787786948, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 6.335153583617748e-05, |
|
"loss": 0.4272, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.2223224304455887, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 6.328327645051195e-05, |
|
"loss": 0.451, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.22316138301330787, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 6.321501706484643e-05, |
|
"loss": 0.4915, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.22400033558102708, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 6.314675767918089e-05, |
|
"loss": 0.4305, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.2248392881487463, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 6.307849829351536e-05, |
|
"loss": 0.4472, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.2256782407164655, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 6.301023890784984e-05, |
|
"loss": 0.4676, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.2265171932841847, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 6.29419795221843e-05, |
|
"loss": 0.4966, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.2273561458519039, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 6.287372013651878e-05, |
|
"loss": 0.4278, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.2281950984196231, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 6.280546075085325e-05, |
|
"loss": 0.4478, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.2290340509873423, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 6.273720136518771e-05, |
|
"loss": 0.4217, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.22987300355506152, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 6.266894197952219e-05, |
|
"loss": 0.4293, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.2307119561227807, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 6.260068259385666e-05, |
|
"loss": 0.46, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.2315509086904999, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 6.253242320819114e-05, |
|
"loss": 0.4623, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.23238986125821912, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 6.24641638225256e-05, |
|
"loss": 0.4705, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.23322881382593832, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 6.239590443686007e-05, |
|
"loss": 0.4416, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.2340677663936575, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 6.232764505119455e-05, |
|
"loss": 0.4106, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.23490671896137671, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 6.225938566552901e-05, |
|
"loss": 0.4642, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.23574567152909592, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 6.21911262798635e-05, |
|
"loss": 0.4476, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.23658462409681513, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 6.212286689419796e-05, |
|
"loss": 0.455, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.23742357666453434, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 6.205460750853242e-05, |
|
"loss": 0.4555, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.23826252923225352, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 6.19863481228669e-05, |
|
"loss": 0.495, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.23910148179997273, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 6.191808873720137e-05, |
|
"loss": 0.4076, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.23994043436769194, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 6.184982935153584e-05, |
|
"loss": 0.4398, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.24077938693541115, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 6.178156996587031e-05, |
|
"loss": 0.4383, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.24161833950313033, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 6.171331058020478e-05, |
|
"loss": 0.4432, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.24245729207084954, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 6.164505119453926e-05, |
|
"loss": 0.4463, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.24329624463856875, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 6.157679180887373e-05, |
|
"loss": 0.4596, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.24413519720628796, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 6.150853242320819e-05, |
|
"loss": 0.5046, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.24497414977400714, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 6.144027303754267e-05, |
|
"loss": 0.4507, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.24581310234172635, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 6.137201365187714e-05, |
|
"loss": 0.4575, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.24665205490944556, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 6.130375426621161e-05, |
|
"loss": 0.4668, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.24749100747716477, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 6.123549488054608e-05, |
|
"loss": 0.4656, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.24832996004488397, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 6.116723549488055e-05, |
|
"loss": 0.4224, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.24916891261260316, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 6.109897610921503e-05, |
|
"loss": 0.4412, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.25000786518032236, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 6.103071672354949e-05, |
|
"loss": 0.4512, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.2508468177480416, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 6.0962457337883964e-05, |
|
"loss": 0.4302, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.2516857703157608, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 6.0894197952218436e-05, |
|
"loss": 0.4406, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.25252472288348, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 6.08259385665529e-05, |
|
"loss": 0.4241, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.2533636754511992, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 6.075767918088738e-05, |
|
"loss": 0.4216, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.25420262801891835, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 6.068941979522185e-05, |
|
"loss": 0.4415, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.25504158058663756, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 6.062116040955632e-05, |
|
"loss": 0.4302, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.25588053315435677, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 6.055290102389079e-05, |
|
"loss": 0.4468, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.256719485722076, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 6.048464163822526e-05, |
|
"loss": 0.4397, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.2575584382897952, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 6.0416382252559736e-05, |
|
"loss": 0.4542, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.2583973908575144, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 6.0348122866894195e-05, |
|
"loss": 0.4199, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.2592363434252336, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 6.0279863481228675e-05, |
|
"loss": 0.4216, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.2600752959929528, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 6.021160409556315e-05, |
|
"loss": 0.4309, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.260914248560672, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 6.014334470989761e-05, |
|
"loss": 0.4567, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.2617532011283912, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 6.0075085324232085e-05, |
|
"loss": 0.4273, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.2625921536961104, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 6.0006825938566564e-05, |
|
"loss": 0.4112, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.2634311062638296, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 5.993856655290103e-05, |
|
"loss": 0.4498, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.2642700588315488, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 5.98703071672355e-05, |
|
"loss": 0.427, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.265109011399268, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 5.980204778156997e-05, |
|
"loss": 0.4167, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.2659479639669872, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 5.973378839590444e-05, |
|
"loss": 0.3787, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.26678691653470643, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 5.966552901023891e-05, |
|
"loss": 0.4317, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.26762586910242564, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 5.959726962457338e-05, |
|
"loss": 0.4098, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.26846482167014485, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 5.952901023890786e-05, |
|
"loss": 0.4357, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.269303774237864, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 5.9460750853242324e-05, |
|
"loss": 0.4277, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.2701427268055832, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 5.9392491467576796e-05, |
|
"loss": 0.4205, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.2709816793733024, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 5.932423208191127e-05, |
|
"loss": 0.4259, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.27182063194102163, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 5.9255972696245734e-05, |
|
"loss": 0.4296, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.27265958450874084, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 5.9187713310580214e-05, |
|
"loss": 0.4092, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.27349853707646005, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 5.9119453924914686e-05, |
|
"loss": 0.4199, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.27433748964417926, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 5.905119453924915e-05, |
|
"loss": 0.3933, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.27517644221189846, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 5.8982935153583624e-05, |
|
"loss": 0.4492, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.2760153947796177, |
|
"grad_norm": 0.6640625, |
|
"learning_rate": 5.891467576791809e-05, |
|
"loss": 0.4505, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.2768543473473368, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 5.884641638225256e-05, |
|
"loss": 0.4073, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.27769329991505604, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 5.877815699658704e-05, |
|
"loss": 0.4388, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.27853225248277524, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 5.870989761092151e-05, |
|
"loss": 0.4237, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.27937120505049445, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 5.864163822525598e-05, |
|
"loss": 0.4001, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.28021015761821366, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 5.8573378839590445e-05, |
|
"loss": 0.4324, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.28104911018593287, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 5.850511945392492e-05, |
|
"loss": 0.4404, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.2818880627536521, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 5.843686006825939e-05, |
|
"loss": 0.4181, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.2827270153213713, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 5.8368600682593856e-05, |
|
"loss": 0.4123, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.2835659678890905, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 5.8300341296928335e-05, |
|
"loss": 0.4224, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.28440492045680965, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 5.823208191126281e-05, |
|
"loss": 0.4059, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.28524387302452886, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 5.816382252559727e-05, |
|
"loss": 0.4295, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.28608282559224807, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 5.8095563139931746e-05, |
|
"loss": 0.4251, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.2869217781599673, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 5.802730375426621e-05, |
|
"loss": 0.4448, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.2877607307276865, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 5.795904436860069e-05, |
|
"loss": 0.4223, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.2885996832954057, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 5.789078498293516e-05, |
|
"loss": 0.4372, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.2894386358631249, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 5.782252559726963e-05, |
|
"loss": 0.3961, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.2902775884308441, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 5.77542662116041e-05, |
|
"loss": 0.4195, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.29111654099856327, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 5.768600682593857e-05, |
|
"loss": 0.4226, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.2919554935662825, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 5.761774744027304e-05, |
|
"loss": 0.4318, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.2927944461340017, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 5.754948805460752e-05, |
|
"loss": 0.4511, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.2936333987017209, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 5.7481228668941984e-05, |
|
"loss": 0.4321, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.2944723512694401, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 5.741296928327646e-05, |
|
"loss": 0.4016, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.2953113038371593, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 5.734470989761092e-05, |
|
"loss": 0.3936, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.2961502564048785, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 5.7276450511945395e-05, |
|
"loss": 0.4152, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.29698920897259773, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 5.720819112627987e-05, |
|
"loss": 0.4021, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.29782816154031694, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 5.713993174061433e-05, |
|
"loss": 0.4409, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.2986671141080361, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 5.707167235494881e-05, |
|
"loss": 0.397, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.2995060666757553, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 5.7003412969283285e-05, |
|
"loss": 0.4422, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.3003450192434745, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 5.693515358361775e-05, |
|
"loss": 0.4104, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.3011839718111937, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 5.686689419795222e-05, |
|
"loss": 0.3982, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.3020229243789129, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 5.679863481228669e-05, |
|
"loss": 0.4193, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.30286187694663214, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 5.673037542662117e-05, |
|
"loss": 0.4249, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.30370082951435134, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 5.666211604095564e-05, |
|
"loss": 0.4162, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.30453978208207055, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 5.6593856655290106e-05, |
|
"loss": 0.3931, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.30537873464978976, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 5.652559726962458e-05, |
|
"loss": 0.3981, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.3062176872175089, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 5.6457337883959044e-05, |
|
"loss": 0.4286, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.3070566397852281, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 5.6389078498293516e-05, |
|
"loss": 0.431, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.30789559235294733, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 5.6320819112627996e-05, |
|
"loss": 0.402, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.30873454492066654, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 5.625255972696246e-05, |
|
"loss": 0.4212, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.30957349748838575, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 5.6184300341296934e-05, |
|
"loss": 0.4355, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.31041245005610496, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 5.6116040955631406e-05, |
|
"loss": 0.4358, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.31125140262382417, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 5.604778156996587e-05, |
|
"loss": 0.4846, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.3120903551915434, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 5.597952218430035e-05, |
|
"loss": 0.3781, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.3129293077592626, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 5.591126279863481e-05, |
|
"loss": 0.4198, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.31376826032698174, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 5.584300341296929e-05, |
|
"loss": 0.4387, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.31460721289470095, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 5.577474402730376e-05, |
|
"loss": 0.4313, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.31544616546242016, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 5.570648464163823e-05, |
|
"loss": 0.4184, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.31628511803013937, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 5.56382252559727e-05, |
|
"loss": 0.4007, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.3171240705978586, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 5.5569965870307165e-05, |
|
"loss": 0.421, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.3179630231655778, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 5.5501706484641645e-05, |
|
"loss": 0.4016, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.318801975733297, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 5.543344709897612e-05, |
|
"loss": 0.3737, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.3196409283010162, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 5.536518771331058e-05, |
|
"loss": 0.422, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.32047988086873536, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 5.5296928327645055e-05, |
|
"loss": 0.42, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.32131883343645457, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 5.522866894197953e-05, |
|
"loss": 0.4118, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.3221577860041738, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 5.5160409556313993e-05, |
|
"loss": 0.3828, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.322996738571893, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 5.509215017064847e-05, |
|
"loss": 0.4322, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.3238356911396122, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 5.502389078498294e-05, |
|
"loss": 0.3888, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.3246746437073314, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 5.495563139931741e-05, |
|
"loss": 0.3914, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.3255135962750506, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 5.488737201365188e-05, |
|
"loss": 0.3864, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.3263525488427698, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 5.481911262798635e-05, |
|
"loss": 0.3853, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.327191501410489, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 5.475085324232083e-05, |
|
"loss": 0.4229, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.3280304539782082, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 5.4682593856655294e-05, |
|
"loss": 0.432, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.3288694065459274, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 5.4614334470989766e-05, |
|
"loss": 0.3766, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.3297083591136466, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 5.454607508532424e-05, |
|
"loss": 0.3977, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.3305473116813658, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 5.4477815699658704e-05, |
|
"loss": 0.3929, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.331386264249085, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 5.440955631399318e-05, |
|
"loss": 0.4273, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.3322252168168042, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 5.4341296928327656e-05, |
|
"loss": 0.3961, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.33306416938452343, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 5.427303754266212e-05, |
|
"loss": 0.3848, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.33390312195224264, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 5.4204778156996594e-05, |
|
"loss": 0.3883, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.33474207451996185, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 5.413651877133106e-05, |
|
"loss": 0.3954, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.335581027087681, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 5.406825938566553e-05, |
|
"loss": 0.4048, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.3364199796554002, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 5.4000000000000005e-05, |
|
"loss": 0.3781, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.3372589322231194, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 5.393174061433447e-05, |
|
"loss": 0.3827, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.33809788479083863, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 5.386348122866895e-05, |
|
"loss": 0.402, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.33893683735855784, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 5.3795221843003415e-05, |
|
"loss": 0.403, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.33977578992627705, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 5.372696245733789e-05, |
|
"loss": 0.4247, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.34061474249399626, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 5.365870307167236e-05, |
|
"loss": 0.3697, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.34145369506171547, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 5.3590443686006826e-05, |
|
"loss": 0.4241, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.3422926476294347, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 5.3522184300341305e-05, |
|
"loss": 0.4229, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.34313160019715383, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 5.345392491467578e-05, |
|
"loss": 0.3971, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.34397055276487304, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 5.338566552901024e-05, |
|
"loss": 0.412, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.34480950533259225, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 5.3317406143344716e-05, |
|
"loss": 0.3692, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.34564845790031146, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 5.324914675767918e-05, |
|
"loss": 0.4232, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.34648741046803067, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 5.3180887372013654e-05, |
|
"loss": 0.4087, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.3473263630357499, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 5.311262798634813e-05, |
|
"loss": 0.4235, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.3481653156034691, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 5.30443686006826e-05, |
|
"loss": 0.4036, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.3490042681711883, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 5.297610921501707e-05, |
|
"loss": 0.3665, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.3498432207389075, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 5.290784982935154e-05, |
|
"loss": 0.3914, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.35068217330662665, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 5.283959044368601e-05, |
|
"loss": 0.3714, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.35152112587434586, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 5.277133105802048e-05, |
|
"loss": 0.4048, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.35236007844206507, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 5.270307167235495e-05, |
|
"loss": 0.3795, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.3531990310097843, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 5.263481228668943e-05, |
|
"loss": 0.4002, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.3540379835775035, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 5.25665529010239e-05, |
|
"loss": 0.4126, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.3548769361452227, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 5.2498293515358365e-05, |
|
"loss": 0.3944, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.3557158887129419, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 5.243003412969284e-05, |
|
"loss": 0.4179, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.3565548412806611, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 5.23617747440273e-05, |
|
"loss": 0.3878, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.35739379384838027, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 5.229351535836178e-05, |
|
"loss": 0.378, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.3582327464160995, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 5.2225255972696255e-05, |
|
"loss": 0.3705, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.3590716989838187, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 5.215699658703072e-05, |
|
"loss": 0.4008, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.3599106515515379, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 5.208873720136519e-05, |
|
"loss": 0.3957, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.3607496041192571, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 5.202047781569966e-05, |
|
"loss": 0.4047, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.3615885566869763, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 5.195221843003413e-05, |
|
"loss": 0.4013, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.3624275092546955, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 5.188395904436861e-05, |
|
"loss": 0.4057, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.36326646182241473, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 5.1815699658703076e-05, |
|
"loss": 0.3759, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.36410541439013394, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 5.174744027303755e-05, |
|
"loss": 0.4095, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.3649443669578531, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 5.1679180887372014e-05, |
|
"loss": 0.3885, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.3657833195255723, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 5.1610921501706486e-05, |
|
"loss": 0.4258, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.3666222720932915, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 5.1542662116040966e-05, |
|
"loss": 0.396, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.3674612246610107, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 5.1474402730375425e-05, |
|
"loss": 0.3749, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.36830017722872993, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 5.1406143344709904e-05, |
|
"loss": 0.3873, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.36913912979644914, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 5.1337883959044376e-05, |
|
"loss": 0.3725, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.36997808236416835, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 5.126962457337884e-05, |
|
"loss": 0.3794, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.37081703493188756, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 5.1201365187713314e-05, |
|
"loss": 0.3908, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.37165598749960677, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 5.113310580204778e-05, |
|
"loss": 0.3812, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.3724949400673259, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 5.106484641638226e-05, |
|
"loss": 0.4344, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.37333389263504513, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 5.099658703071673e-05, |
|
"loss": 0.3884, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.37417284520276434, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 5.09283276450512e-05, |
|
"loss": 0.381, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.37501179777048355, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 5.086006825938567e-05, |
|
"loss": 0.3803, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.37585075033820275, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 5.0791808873720135e-05, |
|
"loss": 0.4266, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.37668970290592196, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 5.072354948805461e-05, |
|
"loss": 0.377, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.3775286554736412, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 5.065529010238909e-05, |
|
"loss": 0.3635, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.3783676080413604, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 5.058703071672355e-05, |
|
"loss": 0.4067, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.3792065606090796, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 5.0518771331058025e-05, |
|
"loss": 0.3815, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.38004551317679874, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 5.04505119453925e-05, |
|
"loss": 0.356, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.38088446574451795, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 5.0382252559726963e-05, |
|
"loss": 0.394, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.38172341831223716, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 5.031399317406144e-05, |
|
"loss": 0.3757, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.38256237087995637, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 5.024573378839591e-05, |
|
"loss": 0.4115, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.3834013234476756, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 5.017747440273038e-05, |
|
"loss": 0.3735, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.3842402760153948, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 5.010921501706485e-05, |
|
"loss": 0.3696, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.385079228583114, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 5.004095563139932e-05, |
|
"loss": 0.4015, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.3859181811508332, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 4.997269624573379e-05, |
|
"loss": 0.388, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.3867571337185524, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 4.990443686006826e-05, |
|
"loss": 0.3666, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.38759608628627157, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 4.9836177474402736e-05, |
|
"loss": 0.3942, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.3884350388539908, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 4.976791808873721e-05, |
|
"loss": 0.3514, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.38927399142171, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 4.9699658703071674e-05, |
|
"loss": 0.3885, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.3901129439894292, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 4.963139931740615e-05, |
|
"loss": 0.3653, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.3909518965571484, |
|
"grad_norm": 0.98046875, |
|
"learning_rate": 4.956313993174062e-05, |
|
"loss": 0.4304, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.3917908491248676, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 4.9494880546075085e-05, |
|
"loss": 0.4087, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.3926298016925868, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 4.9426621160409564e-05, |
|
"loss": 0.3567, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.39346875426030603, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 4.935836177474403e-05, |
|
"loss": 0.3968, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.3943077068280252, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 4.92901023890785e-05, |
|
"loss": 0.4048, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.3951466593957444, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 4.9221843003412975e-05, |
|
"loss": 0.4115, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.3959856119634636, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 4.915358361774744e-05, |
|
"loss": 0.3706, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.3968245645311828, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 4.908532423208192e-05, |
|
"loss": 0.3831, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.397663517098902, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 4.9017064846416385e-05, |
|
"loss": 0.394, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.39850246966662123, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 4.894880546075086e-05, |
|
"loss": 0.3928, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.39934142223434044, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 4.888054607508533e-05, |
|
"loss": 0.3838, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.40018037480205965, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 4.8812286689419796e-05, |
|
"loss": 0.4014, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.40101932736977886, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 4.874402730375427e-05, |
|
"loss": 0.3832, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.401858279937498, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 4.867576791808875e-05, |
|
"loss": 0.3779, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.4026972325052172, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 4.860750853242321e-05, |
|
"loss": 0.3528, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.4035361850729364, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 4.8539249146757686e-05, |
|
"loss": 0.3457, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.40437513764065564, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 4.847098976109215e-05, |
|
"loss": 0.3936, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.40521409020837484, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 4.8402730375426624e-05, |
|
"loss": 0.3731, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.40605304277609405, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 4.8334470989761096e-05, |
|
"loss": 0.3713, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.40689199534381326, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 4.826621160409556e-05, |
|
"loss": 0.3551, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.40773094791153247, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 4.819795221843004e-05, |
|
"loss": 0.3949, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.4085699004792517, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 4.812969283276451e-05, |
|
"loss": 0.4025, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.40940885304697083, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 4.806143344709898e-05, |
|
"loss": 0.4235, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.41024780561469004, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 4.799317406143345e-05, |
|
"loss": 0.3775, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.41108675818240925, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 4.792491467576792e-05, |
|
"loss": 0.3746, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.41192571075012846, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 4.78566552901024e-05, |
|
"loss": 0.3956, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.41276466331784767, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 4.778839590443687e-05, |
|
"loss": 0.3876, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.4136036158855669, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 4.7720136518771335e-05, |
|
"loss": 0.387, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.4144425684532861, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 4.765187713310581e-05, |
|
"loss": 0.4316, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.4152815210210053, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 4.758361774744027e-05, |
|
"loss": 0.3687, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.4161204735887245, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 4.7515358361774745e-05, |
|
"loss": 0.3934, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.41695942615644366, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 4.7447098976109225e-05, |
|
"loss": 0.3847, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.41779837872416287, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 4.737883959044369e-05, |
|
"loss": 0.3557, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.4186373312918821, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 4.731058020477816e-05, |
|
"loss": 0.4218, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.4194762838596013, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 4.724232081911263e-05, |
|
"loss": 0.4181, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.4203152364273205, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 4.71740614334471e-05, |
|
"loss": 0.3795, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.4211541889950397, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 4.710580204778158e-05, |
|
"loss": 0.4028, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.4219931415627589, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 4.7037542662116046e-05, |
|
"loss": 0.3825, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.4228320941304781, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 4.696928327645052e-05, |
|
"loss": 0.3888, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.4236710466981973, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 4.6901023890784984e-05, |
|
"loss": 0.4167, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.4245099992659165, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 4.6832764505119456e-05, |
|
"loss": 0.3944, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.4253489518336357, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 4.676450511945393e-05, |
|
"loss": 0.3643, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.4261879044013549, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 4.6696245733788395e-05, |
|
"loss": 0.3612, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.4270268569690741, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 4.6627986348122874e-05, |
|
"loss": 0.3632, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.4278658095367933, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 4.6559726962457346e-05, |
|
"loss": 0.3677, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.4287047621045125, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 4.649146757679181e-05, |
|
"loss": 0.3721, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.42954371467223174, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 4.6423208191126284e-05, |
|
"loss": 0.3836, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.43038266723995094, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 4.635494880546075e-05, |
|
"loss": 0.3635, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.4312216198076701, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 4.628668941979522e-05, |
|
"loss": 0.4082, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.4320605723753893, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 4.62184300341297e-05, |
|
"loss": 0.3845, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.4328995249431085, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 4.615017064846417e-05, |
|
"loss": 0.356, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.4337384775108277, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 4.608191126279864e-05, |
|
"loss": 0.3948, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.43457743007854693, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 4.6013651877133106e-05, |
|
"loss": 0.3754, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.43541638264626614, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 4.594539249146758e-05, |
|
"loss": 0.3837, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.43625533521398535, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 4.587713310580206e-05, |
|
"loss": 0.3595, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.43709428778170456, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 4.580887372013652e-05, |
|
"loss": 0.3998, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.43793324034942377, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 4.5740614334470995e-05, |
|
"loss": 0.365, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.4387721929171429, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 4.567235494880547e-05, |
|
"loss": 0.3782, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.43961114548486213, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 4.5604095563139933e-05, |
|
"loss": 0.362, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.44045009805258134, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 4.5535836177474406e-05, |
|
"loss": 0.3825, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.44128905062030055, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 4.546757679180887e-05, |
|
"loss": 0.363, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.44212800318801976, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 4.539931740614335e-05, |
|
"loss": 0.3932, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.44296695575573897, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 4.533105802047782e-05, |
|
"loss": 0.3776, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.4438059083234582, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 4.526279863481229e-05, |
|
"loss": 0.3596, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.4446448608911774, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 4.519453924914676e-05, |
|
"loss": 0.3707, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.4454838134588966, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 4.512627986348123e-05, |
|
"loss": 0.3773, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.44632276602661575, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 4.50580204778157e-05, |
|
"loss": 0.3686, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.44716171859433496, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 4.498976109215018e-05, |
|
"loss": 0.3864, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.44800067116205416, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 4.4921501706484644e-05, |
|
"loss": 0.3432, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.4488396237297734, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 4.485324232081912e-05, |
|
"loss": 0.3807, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.4496785762974926, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 4.478498293515359e-05, |
|
"loss": 0.4029, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.4505175288652118, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 4.4716723549488055e-05, |
|
"loss": 0.3901, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.451356481432931, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 4.4648464163822534e-05, |
|
"loss": 0.3755, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.4521954340006502, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 4.4580204778157e-05, |
|
"loss": 0.3339, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.4530343865683694, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 4.451194539249147e-05, |
|
"loss": 0.3822, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.45387333913608857, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 4.4443686006825945e-05, |
|
"loss": 0.3514, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.4547122917038078, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 4.437542662116041e-05, |
|
"loss": 0.3611, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.455551244271527, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 4.430716723549488e-05, |
|
"loss": 0.3644, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.4563901968392462, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 4.423890784982935e-05, |
|
"loss": 0.3892, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.4572291494069654, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 4.417064846416383e-05, |
|
"loss": 0.3801, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.4580681019746846, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 4.41023890784983e-05, |
|
"loss": 0.3483, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.4589070545424038, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 4.4034129692832766e-05, |
|
"loss": 0.3817, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.45974600711012303, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 4.396587030716724e-05, |
|
"loss": 0.3588, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.4605849596778422, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 4.389761092150172e-05, |
|
"loss": 0.3589, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.4614239122455614, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 4.3829351535836177e-05, |
|
"loss": 0.393, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.4622628648132806, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 4.3761092150170656e-05, |
|
"loss": 0.3636, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.4631018173809998, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 4.369283276450512e-05, |
|
"loss": 0.3836, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.463940769948719, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 4.3624573378839594e-05, |
|
"loss": 0.3607, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.46477972251643823, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 4.3556313993174066e-05, |
|
"loss": 0.3756, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.46561867508415744, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 4.348805460750853e-05, |
|
"loss": 0.3677, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.46645762765187665, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 4.341979522184301e-05, |
|
"loss": 0.3466, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.46729658021959586, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 4.335153583617748e-05, |
|
"loss": 0.3478, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.468135532787315, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 4.328327645051195e-05, |
|
"loss": 0.3491, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.4689744853550342, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 4.321501706484642e-05, |
|
"loss": 0.3583, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.46981343792275343, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 4.314675767918089e-05, |
|
"loss": 0.4017, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.47065239049047264, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 4.307849829351536e-05, |
|
"loss": 0.355, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.47149134305819185, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 4.301023890784984e-05, |
|
"loss": 0.3713, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.47233029562591106, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 4.2941979522184305e-05, |
|
"loss": 0.3893, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.47316924819363027, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 4.287372013651878e-05, |
|
"loss": 0.3784, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.4740082007613495, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 4.280546075085324e-05, |
|
"loss": 0.3578, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.4748471533290687, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 4.2737201365187716e-05, |
|
"loss": 0.3649, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.47568610589678784, |
|
"grad_norm": 0.375, |
|
"learning_rate": 4.2668941979522195e-05, |
|
"loss": 0.3454, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.47652505846450705, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 4.260068259385666e-05, |
|
"loss": 0.3565, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.47736401103222625, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 4.253242320819113e-05, |
|
"loss": 0.3402, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.47820296359994546, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 4.24641638225256e-05, |
|
"loss": 0.359, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.47904191616766467, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 4.239590443686007e-05, |
|
"loss": 0.3324, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.4798808687353839, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 4.2327645051194543e-05, |
|
"loss": 0.376, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.4807198213031031, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 4.225938566552901e-05, |
|
"loss": 0.3775, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.4815587738708223, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 4.219112627986349e-05, |
|
"loss": 0.3478, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.4823977264385415, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 4.2122866894197954e-05, |
|
"loss": 0.3594, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.48323667900626066, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 4.2054607508532426e-05, |
|
"loss": 0.3592, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.48407563157397987, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 4.19863481228669e-05, |
|
"loss": 0.355, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.4849145841416991, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 4.1918088737201365e-05, |
|
"loss": 0.3213, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.4857535367094183, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 4.184982935153584e-05, |
|
"loss": 0.3554, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.4865924892771375, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 4.1781569965870316e-05, |
|
"loss": 0.4197, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.4874314418448567, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 4.171331058020478e-05, |
|
"loss": 0.3765, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.4882703944125759, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 4.1645051194539254e-05, |
|
"loss": 0.3412, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.4891093469802951, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 4.157679180887372e-05, |
|
"loss": 0.3748, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.4899482995480143, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 4.150853242320819e-05, |
|
"loss": 0.3282, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.4907872521157335, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 4.144027303754267e-05, |
|
"loss": 0.3495, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.4916262046834527, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 4.137201365187714e-05, |
|
"loss": 0.3633, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.4924651572511719, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 4.130375426621161e-05, |
|
"loss": 0.3537, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.4933041098188911, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 4.1235494880546076e-05, |
|
"loss": 0.3498, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.4941430623866103, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 4.116723549488055e-05, |
|
"loss": 0.3607, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.49498201495432953, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 4.109897610921502e-05, |
|
"loss": 0.3749, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.49582096752204874, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 4.1030716723549486e-05, |
|
"loss": 0.3455, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.49665992008976795, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 4.0962457337883965e-05, |
|
"loss": 0.3392, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.4974988726574871, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 4.089419795221844e-05, |
|
"loss": 0.3759, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.4983378252252063, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 4.0825938566552904e-05, |
|
"loss": 0.3544, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.4991767777929255, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 4.0757679180887376e-05, |
|
"loss": 0.3617, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.5000157303606447, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 4.068941979522184e-05, |
|
"loss": 0.3591, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.500854682928364, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 4.0621160409556314e-05, |
|
"loss": 0.3589, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.5016936354960831, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 4.055290102389079e-05, |
|
"loss": 0.3423, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.5025325880638023, |
|
"grad_norm": 0.375, |
|
"learning_rate": 4.048464163822526e-05, |
|
"loss": 0.3565, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.5033715406315216, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 4.041638225255973e-05, |
|
"loss": 0.3573, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.5042104931992407, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 4.03481228668942e-05, |
|
"loss": 0.3922, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 0.50504944576696, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 4.027986348122867e-05, |
|
"loss": 0.378, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.5058883983346791, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 4.021160409556315e-05, |
|
"loss": 0.3494, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 0.5067273509023984, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 4.0143344709897615e-05, |
|
"loss": 0.3469, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.5075663034701176, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 4.007508532423209e-05, |
|
"loss": 0.337, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.5084052560378367, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 4.000682593856656e-05, |
|
"loss": 0.3258, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.509244208605556, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 3.9938566552901025e-05, |
|
"loss": 0.3762, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 0.5100831611732751, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 3.98703071672355e-05, |
|
"loss": 0.3797, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.5109221137409944, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 3.980204778156997e-05, |
|
"loss": 0.3369, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 0.5117610663087135, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 3.973378839590444e-05, |
|
"loss": 0.3597, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.5126000188764328, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 3.966552901023891e-05, |
|
"loss": 0.3399, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 0.513438971444152, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 3.959726962457338e-05, |
|
"loss": 0.3617, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.5142779240118712, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 3.952901023890785e-05, |
|
"loss": 0.3236, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 0.5151168765795904, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 3.9460750853242325e-05, |
|
"loss": 0.3636, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.5159558291473095, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 3.939249146757679e-05, |
|
"loss": 0.3515, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.5167947817150288, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 3.9324232081911264e-05, |
|
"loss": 0.374, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.517633734282748, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 3.9255972696245736e-05, |
|
"loss": 0.335, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 0.5184726868504672, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 3.918771331058021e-05, |
|
"loss": 0.368, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.5193116394181864, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 3.911945392491468e-05, |
|
"loss": 0.3317, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 0.5201505919859056, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 3.905119453924915e-05, |
|
"loss": 0.3271, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.5209895445536248, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 3.8982935153583626e-05, |
|
"loss": 0.3437, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 0.521828497121344, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 3.891467576791809e-05, |
|
"loss": 0.3447, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.5226674496890632, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 3.8846416382252564e-05, |
|
"loss": 0.3611, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 0.5235064022567824, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 3.877815699658703e-05, |
|
"loss": 0.376, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.5243453548245016, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 3.87098976109215e-05, |
|
"loss": 0.3582, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.5251843073922208, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 3.8641638225255975e-05, |
|
"loss": 0.3765, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.52602325995994, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 3.857337883959045e-05, |
|
"loss": 0.3421, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 0.5268622125276592, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 3.850511945392492e-05, |
|
"loss": 0.3694, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.5277011650953785, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 3.8436860068259385e-05, |
|
"loss": 0.3709, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 0.5285401176630976, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 3.8368600682593864e-05, |
|
"loss": 0.3469, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.5293790702308169, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 3.830034129692833e-05, |
|
"loss": 0.3826, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 0.530218022798536, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 3.82320819112628e-05, |
|
"loss": 0.3406, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.5310569753662552, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 3.8163822525597275e-05, |
|
"loss": 0.3424, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 0.5318959279339744, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 3.809556313993174e-05, |
|
"loss": 0.3457, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.5327348805016936, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 3.802730375426621e-05, |
|
"loss": 0.3666, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.5335738330694129, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 3.7959044368600686e-05, |
|
"loss": 0.3582, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.534412785637132, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 3.789078498293516e-05, |
|
"loss": 0.3327, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 0.5352517382048513, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 3.7822525597269624e-05, |
|
"loss": 0.3599, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.5360906907725704, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 3.77542662116041e-05, |
|
"loss": 0.3429, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 0.5369296433402897, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 3.768600682593857e-05, |
|
"loss": 0.3605, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.5377685959080089, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 3.761774744027304e-05, |
|
"loss": 0.3699, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 0.538607548475728, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 3.7549488054607514e-05, |
|
"loss": 0.3136, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.5394465010434473, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 3.7481228668941986e-05, |
|
"loss": 0.3854, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 0.5402854536111664, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 3.741296928327645e-05, |
|
"loss": 0.3788, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.5411244061788857, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 3.7344709897610924e-05, |
|
"loss": 0.3655, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.5419633587466048, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 3.7276450511945397e-05, |
|
"loss": 0.3541, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.5428023113143241, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 3.720819112627986e-05, |
|
"loss": 0.3621, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 0.5436412638820433, |
|
"grad_norm": 0.375, |
|
"learning_rate": 3.713993174061434e-05, |
|
"loss": 0.3473, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.5444802164497625, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 3.707167235494881e-05, |
|
"loss": 0.3243, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 0.5453191690174817, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 3.700341296928328e-05, |
|
"loss": 0.3351, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.5461581215852008, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 3.693515358361775e-05, |
|
"loss": 0.3426, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 0.5469970741529201, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 3.6866894197952224e-05, |
|
"loss": 0.3556, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.5478360267206392, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 3.679863481228669e-05, |
|
"loss": 0.3233, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 0.5486749792883585, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 3.673037542662116e-05, |
|
"loss": 0.3248, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.5495139318560777, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 3.6662116040955635e-05, |
|
"loss": 0.3561, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.5503528844237969, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 3.659385665529011e-05, |
|
"loss": 0.3514, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.5511918369915161, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 3.652559726962458e-05, |
|
"loss": 0.3569, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 0.5520307895592353, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 3.6457337883959046e-05, |
|
"loss": 0.3809, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.5528697421269545, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 3.638907849829352e-05, |
|
"loss": 0.3447, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 0.5537086946946737, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 3.632081911262799e-05, |
|
"loss": 0.3543, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.5545476472623929, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 3.625255972696246e-05, |
|
"loss": 0.3384, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 0.5553865998301121, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 3.618430034129693e-05, |
|
"loss": 0.354, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.5562255523978313, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 3.61160409556314e-05, |
|
"loss": 0.3303, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 0.5570645049655505, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 3.6047781569965874e-05, |
|
"loss": 0.356, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.5579034575332698, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 3.5979522184300346e-05, |
|
"loss": 0.3523, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.5587424101009889, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 3.591126279863482e-05, |
|
"loss": 0.3382, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.5595813626687082, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 3.5843003412969284e-05, |
|
"loss": 0.3311, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 0.5604203152364273, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 3.577474402730376e-05, |
|
"loss": 0.3281, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.5612592678041465, |
|
"grad_norm": 0.375, |
|
"learning_rate": 3.570648464163823e-05, |
|
"loss": 0.3499, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 0.5620982203718657, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 3.56382252559727e-05, |
|
"loss": 0.3492, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.5629371729395849, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 3.556996587030717e-05, |
|
"loss": 0.3639, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 0.5637761255073042, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 3.550170648464164e-05, |
|
"loss": 0.3119, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.5646150780750233, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 3.543344709897611e-05, |
|
"loss": 0.3589, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 0.5654540306427426, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 3.5365187713310585e-05, |
|
"loss": 0.3621, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.5662929832104617, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 3.529692832764506e-05, |
|
"loss": 0.3264, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.567131935778181, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 3.522866894197952e-05, |
|
"loss": 0.3505, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.5679708883459001, |
|
"grad_norm": 0.375, |
|
"learning_rate": 3.5160409556313995e-05, |
|
"loss": 0.3302, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 0.5688098409136193, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 3.509215017064847e-05, |
|
"loss": 0.3317, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.5696487934813386, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 3.502389078498294e-05, |
|
"loss": 0.3364, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 0.5704877460490577, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 3.4955631399317406e-05, |
|
"loss": 0.3216, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.571326698616777, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 3.488737201365188e-05, |
|
"loss": 0.3494, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 0.5721656511844961, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 3.481911262798635e-05, |
|
"loss": 0.3391, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.5730046037522154, |
|
"grad_norm": 0.375, |
|
"learning_rate": 3.475085324232082e-05, |
|
"loss": 0.3256, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 0.5738435563199346, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 3.4682593856655296e-05, |
|
"loss": 0.3261, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.5746825088876537, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 3.461433447098976e-05, |
|
"loss": 0.3597, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.575521461455373, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 3.4546075085324234e-05, |
|
"loss": 0.3247, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.5763604140230921, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 3.4477815699658706e-05, |
|
"loss": 0.3373, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 0.5771993665908114, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 3.440955631399318e-05, |
|
"loss": 0.3392, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.5780383191585305, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 3.4341296928327644e-05, |
|
"loss": 0.3333, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 0.5788772717262498, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 3.427303754266212e-05, |
|
"loss": 0.3332, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.579716224293969, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 3.420477815699659e-05, |
|
"loss": 0.3262, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 0.5805551768616882, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 3.413651877133106e-05, |
|
"loss": 0.3367, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.5813941294294074, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 3.4068259385665534e-05, |
|
"loss": 0.3586, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 0.5822330819971265, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 3.4e-05, |
|
"loss": 0.3445, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.5830720345648458, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 3.393174061433447e-05, |
|
"loss": 0.3563, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.583910987132565, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 3.3863481228668945e-05, |
|
"loss": 0.3265, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.5847499397002842, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 3.379522184300342e-05, |
|
"loss": 0.3438, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 0.5855888922680034, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 3.372696245733789e-05, |
|
"loss": 0.3318, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.5864278448357226, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 3.3658703071672355e-05, |
|
"loss": 0.3124, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 0.5872667974034418, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 3.359044368600683e-05, |
|
"loss": 0.3352, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.588105749971161, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 3.35221843003413e-05, |
|
"loss": 0.3296, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 0.5889447025388802, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 3.345392491467577e-05, |
|
"loss": 0.3408, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.5897836551065994, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 3.338566552901024e-05, |
|
"loss": 0.381, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 0.5906226076743186, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 3.331740614334471e-05, |
|
"loss": 0.3506, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.5914615602420378, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 3.324914675767918e-05, |
|
"loss": 0.3229, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.592300512809757, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 3.3180887372013656e-05, |
|
"loss": 0.3506, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 0.5931394653774762, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 3.311262798634813e-05, |
|
"loss": 0.3208, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 0.5939784179451955, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 3.3044368600682594e-05, |
|
"loss": 0.3423, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 0.5948173705129146, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 3.2976109215017066e-05, |
|
"loss": 0.3651, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 0.5956563230806339, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 3.290784982935154e-05, |
|
"loss": 0.3303, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.596495275648353, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 3.283959044368601e-05, |
|
"loss": 0.3473, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 0.5973342282160722, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 3.277133105802048e-05, |
|
"loss": 0.3125, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 0.5981731807837914, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 3.2703071672354956e-05, |
|
"loss": 0.3107, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 0.5990121333515106, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 3.263481228668942e-05, |
|
"loss": 0.3162, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 0.5998510859192299, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 3.2566552901023894e-05, |
|
"loss": 0.3519, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.600690038486949, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 3.2498293515358367e-05, |
|
"loss": 0.3237, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 0.6015289910546683, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 3.243003412969283e-05, |
|
"loss": 0.3399, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 0.6023679436223874, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 3.2361774744027305e-05, |
|
"loss": 0.3439, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 0.6032068961901067, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 3.229351535836178e-05, |
|
"loss": 0.3279, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 0.6040458487578259, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 3.222525597269625e-05, |
|
"loss": 0.343, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.604884801325545, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 3.2156996587030715e-05, |
|
"loss": 0.3484, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 0.6057237538932643, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 3.2088737201365195e-05, |
|
"loss": 0.333, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 0.6065627064609834, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 3.202047781569966e-05, |
|
"loss": 0.3551, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 0.6074016590287027, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 3.195221843003413e-05, |
|
"loss": 0.3482, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 0.6082406115964218, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 3.1883959044368605e-05, |
|
"loss": 0.3644, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.6090795641641411, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 3.181569965870308e-05, |
|
"loss": 0.3326, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 0.6099185167318603, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 3.174744027303754e-05, |
|
"loss": 0.3247, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 0.6107574692995795, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 3.1679180887372016e-05, |
|
"loss": 0.3696, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 0.6115964218672987, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 3.161092150170649e-05, |
|
"loss": 0.3406, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 0.6124353744350178, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 3.1542662116040954e-05, |
|
"loss": 0.3316, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.6132743270027371, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 3.147440273037543e-05, |
|
"loss": 0.317, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 0.6141132795704562, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 3.14061433447099e-05, |
|
"loss": 0.3533, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 0.6149522321381755, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 3.133788395904437e-05, |
|
"loss": 0.3467, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 0.6157911847058947, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 3.1269624573378844e-05, |
|
"loss": 0.3187, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 0.6166301372736139, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 3.1201365187713316e-05, |
|
"loss": 0.3329, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.6174690898413331, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 3.113310580204778e-05, |
|
"loss": 0.344, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 0.6183080424090523, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 3.1064846416382254e-05, |
|
"loss": 0.3199, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 0.6191469949767715, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 3.099658703071673e-05, |
|
"loss": 0.3271, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 0.6199859475444907, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 3.092832764505119e-05, |
|
"loss": 0.3497, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 0.6208249001122099, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 3.086006825938567e-05, |
|
"loss": 0.3234, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.6216638526799291, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 3.079180887372014e-05, |
|
"loss": 0.3571, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 0.6225028052476483, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 3.072354948805461e-05, |
|
"loss": 0.3376, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 0.6233417578153675, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 3.065529010238908e-05, |
|
"loss": 0.3077, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 0.6241807103830868, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 3.0587030716723555e-05, |
|
"loss": 0.336, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.6250196629508059, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 3.051877133105802e-05, |
|
"loss": 0.3266, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.6258586155185252, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 3.0450511945392493e-05, |
|
"loss": 0.3642, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 0.6266975680862443, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 3.0382252559726965e-05, |
|
"loss": 0.3294, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 0.6275365206539635, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 3.0313993174061438e-05, |
|
"loss": 0.3365, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 0.6283754732216827, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 3.0245733788395907e-05, |
|
"loss": 0.3473, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 0.6292144257894019, |
|
"grad_norm": 0.375, |
|
"learning_rate": 3.017747440273038e-05, |
|
"loss": 0.3268, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.6300533783571212, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 3.0109215017064848e-05, |
|
"loss": 0.3146, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 0.6308923309248403, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 3.0040955631399317e-05, |
|
"loss": 0.3401, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 0.6317312834925596, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 2.9972696245733793e-05, |
|
"loss": 0.3098, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 0.6325702360602787, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 2.9904436860068262e-05, |
|
"loss": 0.306, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 0.633409188627998, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 2.983617747440273e-05, |
|
"loss": 0.3441, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.6342481411957172, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 2.9767918088737204e-05, |
|
"loss": 0.3614, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 0.6350870937634363, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 2.9699658703071676e-05, |
|
"loss": 0.3234, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 0.6359260463311556, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 2.9631399317406145e-05, |
|
"loss": 0.2922, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 0.6367649988988747, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 2.9563139931740618e-05, |
|
"loss": 0.3327, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 0.637603951466594, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 2.9494880546075087e-05, |
|
"loss": 0.3221, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.6384429040343131, |
|
"grad_norm": 0.375, |
|
"learning_rate": 2.942662116040956e-05, |
|
"loss": 0.3255, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 0.6392818566020324, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 2.935836177474403e-05, |
|
"loss": 0.2936, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 0.6401208091697516, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 2.92901023890785e-05, |
|
"loss": 0.3064, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 0.6409597617374707, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 2.922184300341297e-05, |
|
"loss": 0.3213, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 0.64179871430519, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 2.9153583617747442e-05, |
|
"loss": 0.3455, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.6426376668729091, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 2.9085324232081915e-05, |
|
"loss": 0.3068, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 0.6434766194406284, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 2.9017064846416384e-05, |
|
"loss": 0.3188, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 0.6443155720083475, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 2.8948805460750856e-05, |
|
"loss": 0.2951, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 0.6451545245760668, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 2.8880546075085325e-05, |
|
"loss": 0.3078, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 0.645993477143786, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 2.88122866894198e-05, |
|
"loss": 0.3267, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.6468324297115052, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 2.874402730375427e-05, |
|
"loss": 0.3431, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 0.6476713822792244, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 2.867576791808874e-05, |
|
"loss": 0.3016, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 0.6485103348469435, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 2.8607508532423208e-05, |
|
"loss": 0.3218, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 0.6493492874146628, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 2.853924914675768e-05, |
|
"loss": 0.3162, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 0.650188239982382, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 2.8470989761092153e-05, |
|
"loss": 0.2939, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.6510271925501012, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 2.8402730375426622e-05, |
|
"loss": 0.3235, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 0.6518661451178204, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 2.8334470989761095e-05, |
|
"loss": 0.3244, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 0.6527050976855396, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 2.8266211604095564e-05, |
|
"loss": 0.3491, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 0.6535440502532588, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 2.819795221843004e-05, |
|
"loss": 0.3061, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 0.654383002820978, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 2.812969283276451e-05, |
|
"loss": 0.3574, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.6552219553886972, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 2.8061433447098978e-05, |
|
"loss": 0.3341, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 0.6560609079564164, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 2.7993174061433447e-05, |
|
"loss": 0.333, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 0.6568998605241356, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 2.7924914675767923e-05, |
|
"loss": 0.3102, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 0.6577388130918548, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 2.7856655290102392e-05, |
|
"loss": 0.3315, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 0.658577765659574, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 2.778839590443686e-05, |
|
"loss": 0.3534, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.6594167182272932, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 2.7720136518771333e-05, |
|
"loss": 0.3407, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 0.6602556707950125, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 2.7651877133105802e-05, |
|
"loss": 0.2997, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 0.6610946233627316, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 2.7583617747440278e-05, |
|
"loss": 0.3279, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 0.6619335759304509, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 2.7515358361774747e-05, |
|
"loss": 0.3335, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 0.66277252849817, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 2.7447098976109216e-05, |
|
"loss": 0.3275, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.6636114810658892, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 2.7378839590443685e-05, |
|
"loss": 0.3384, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 0.6644504336336085, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 2.731058020477816e-05, |
|
"loss": 0.3416, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 0.6652893862013276, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 2.724232081911263e-05, |
|
"loss": 0.3141, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 0.6661283387690469, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 2.71740614334471e-05, |
|
"loss": 0.2982, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 0.666967291336766, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 2.7105802047781572e-05, |
|
"loss": 0.2999, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.6678062439044853, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 2.7037542662116044e-05, |
|
"loss": 0.3125, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 0.6686451964722044, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 2.6969283276450517e-05, |
|
"loss": 0.3205, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 0.6694841490399237, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 2.6901023890784986e-05, |
|
"loss": 0.3469, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 0.6703231016076429, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 2.6832764505119455e-05, |
|
"loss": 0.3269, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 0.671162054175362, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 2.6764505119453924e-05, |
|
"loss": 0.3178, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.6720010067430813, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 2.66962457337884e-05, |
|
"loss": 0.3161, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 0.6728399593108004, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 2.662798634812287e-05, |
|
"loss": 0.3221, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 0.6736789118785197, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 2.6559726962457338e-05, |
|
"loss": 0.3135, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 0.6745178644462388, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 2.649146757679181e-05, |
|
"loss": 0.3181, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 0.6753568170139581, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 2.6423208191126283e-05, |
|
"loss": 0.3273, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.6761957695816773, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 2.6354948805460755e-05, |
|
"loss": 0.3148, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 0.6770347221493965, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 2.6286689419795224e-05, |
|
"loss": 0.3343, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 0.6778736747171157, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 2.6218430034129693e-05, |
|
"loss": 0.3092, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 0.6787126272848348, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 2.6150170648464162e-05, |
|
"loss": 0.3048, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 0.6795515798525541, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 2.6081911262798638e-05, |
|
"loss": 0.2975, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.6803905324202733, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 2.6013651877133107e-05, |
|
"loss": 0.3113, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 0.6812294849879925, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 2.594539249146758e-05, |
|
"loss": 0.3479, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 0.6820684375557117, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 2.587713310580205e-05, |
|
"loss": 0.3433, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 0.6829073901234309, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 2.580887372013652e-05, |
|
"loss": 0.3153, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 0.6837463426911501, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 2.5740614334470994e-05, |
|
"loss": 0.3238, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.6845852952588694, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 2.5672354948805463e-05, |
|
"loss": 0.3233, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 0.6854242478265885, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 2.5604095563139932e-05, |
|
"loss": 0.3207, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 0.6862632003943077, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 2.5535836177474408e-05, |
|
"loss": 0.3189, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 0.6871021529620269, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 2.5467576791808877e-05, |
|
"loss": 0.312, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 0.6879411055297461, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 2.5399317406143346e-05, |
|
"loss": 0.3189, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.6887800580974653, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 2.5331058020477818e-05, |
|
"loss": 0.3152, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 0.6896190106651845, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 2.5262798634812287e-05, |
|
"loss": 0.3393, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 0.6904579632329038, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 2.519453924914676e-05, |
|
"loss": 0.3043, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 0.6912969158006229, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 2.5126279863481232e-05, |
|
"loss": 0.3183, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 0.6921358683683422, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 2.50580204778157e-05, |
|
"loss": 0.312, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.6929748209360613, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 2.498976109215017e-05, |
|
"loss": 0.3176, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 0.6938137735037805, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 2.4921501706484646e-05, |
|
"loss": 0.3067, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 0.6946527260714997, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 2.4853242320819115e-05, |
|
"loss": 0.3491, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 0.6954916786392189, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 2.4784982935153584e-05, |
|
"loss": 0.3276, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 0.6963306312069382, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 2.4716723549488057e-05, |
|
"loss": 0.3044, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.6971695837746573, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 2.464846416382253e-05, |
|
"loss": 0.3063, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 0.6980085363423766, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 2.4580204778157e-05, |
|
"loss": 0.3146, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 0.6988474889100957, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 2.451194539249147e-05, |
|
"loss": 0.3358, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 0.699686441477815, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 2.444368600682594e-05, |
|
"loss": 0.3146, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 0.7005253940455342, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 2.437542662116041e-05, |
|
"loss": 0.3078, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.7013643466132533, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 2.4307167235494885e-05, |
|
"loss": 0.3059, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 0.7022032991809726, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 2.4238907849829354e-05, |
|
"loss": 0.3047, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 0.7030422517486917, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 2.4170648464163823e-05, |
|
"loss": 0.3344, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 0.703881204316411, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 2.4102389078498295e-05, |
|
"loss": 0.3081, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 0.7047201568841301, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 2.4034129692832768e-05, |
|
"loss": 0.2912, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.7055591094518494, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 2.3965870307167237e-05, |
|
"loss": 0.3128, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 0.7063980620195686, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 2.389761092150171e-05, |
|
"loss": 0.3087, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 0.7072370145872878, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 2.382935153583618e-05, |
|
"loss": 0.3116, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 0.708075967155007, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 2.3761092150170654e-05, |
|
"loss": 0.321, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 0.7089149197227261, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 2.3692832764505123e-05, |
|
"loss": 0.2924, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.7097538722904454, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 2.3624573378839592e-05, |
|
"loss": 0.3071, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 0.7105928248581646, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 2.355631399317406e-05, |
|
"loss": 0.3217, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 0.7114317774258838, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 2.3488054607508534e-05, |
|
"loss": 0.3306, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 0.712270729993603, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 2.3419795221843006e-05, |
|
"loss": 0.3178, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 0.7131096825613222, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 2.3351535836177475e-05, |
|
"loss": 0.3169, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.7139486351290414, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 2.3283276450511948e-05, |
|
"loss": 0.2919, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 0.7147875876967605, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 2.3215017064846417e-05, |
|
"loss": 0.3017, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 0.7156265402644798, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 2.3146757679180893e-05, |
|
"loss": 0.3285, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 0.716465492832199, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 2.3078498293515362e-05, |
|
"loss": 0.302, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 0.7173044453999182, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 2.301023890784983e-05, |
|
"loss": 0.3347, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.7181433979676374, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 2.29419795221843e-05, |
|
"loss": 0.3461, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 0.7189823505353566, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 2.2873720136518772e-05, |
|
"loss": 0.3167, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 0.7198213031030758, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 2.2805460750853245e-05, |
|
"loss": 0.3068, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 0.7206602556707951, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 2.2737201365187714e-05, |
|
"loss": 0.3143, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 0.7214992082385142, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 2.2668941979522186e-05, |
|
"loss": 0.3136, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.7223381608062334, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 2.2600682593856655e-05, |
|
"loss": 0.298, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 0.7231771133739526, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 2.253242320819113e-05, |
|
"loss": 0.3019, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 0.7240160659416718, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 2.24641638225256e-05, |
|
"loss": 0.3116, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 0.724855018509391, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 2.239590443686007e-05, |
|
"loss": 0.2842, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 0.7256939710771102, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 2.232764505119454e-05, |
|
"loss": 0.3347, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.7265329236448295, |
|
"grad_norm": 0.296875, |
|
"learning_rate": 2.2259385665529014e-05, |
|
"loss": 0.2836, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 0.7273718762125486, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 2.2191126279863483e-05, |
|
"loss": 0.3001, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 0.7282108287802679, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 2.2122866894197952e-05, |
|
"loss": 0.3101, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 0.729049781347987, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 2.2054607508532425e-05, |
|
"loss": 0.3106, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 0.7298887339157062, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 2.1986348122866894e-05, |
|
"loss": 0.3473, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.7307276864834255, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 2.191808873720137e-05, |
|
"loss": 0.3223, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 0.7315666390511446, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 2.184982935153584e-05, |
|
"loss": 0.3213, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 0.7324055916188639, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 2.1781569965870308e-05, |
|
"loss": 0.3069, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 0.733244544186583, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 2.1713310580204777e-05, |
|
"loss": 0.3324, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 0.7340834967543023, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 2.1645051194539253e-05, |
|
"loss": 0.3394, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.7349224493220214, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 2.1576791808873722e-05, |
|
"loss": 0.3097, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 0.7357614018897407, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 2.1508532423208194e-05, |
|
"loss": 0.3354, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 0.7366003544574599, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 2.1440273037542663e-05, |
|
"loss": 0.2888, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 0.737439307025179, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 2.1372013651877136e-05, |
|
"loss": 0.3024, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 0.7382782595928983, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 2.1303754266211608e-05, |
|
"loss": 0.3279, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.7391172121606174, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 2.1235494880546077e-05, |
|
"loss": 0.3129, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 0.7399561647283367, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 2.1167235494880546e-05, |
|
"loss": 0.291, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 0.7407951172960558, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 2.1098976109215015e-05, |
|
"loss": 0.285, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 0.7416340698637751, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 2.103071672354949e-05, |
|
"loss": 0.332, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 0.7424730224314943, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 2.096245733788396e-05, |
|
"loss": 0.3272, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.7433119749992135, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 2.0894197952218433e-05, |
|
"loss": 0.2977, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 0.7441509275669327, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 2.0825938566552902e-05, |
|
"loss": 0.3118, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 0.7449898801346518, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 2.0757679180887374e-05, |
|
"loss": 0.3112, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 0.7458288327023711, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 2.0689419795221847e-05, |
|
"loss": 0.3137, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 0.7466677852700903, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 2.0621160409556316e-05, |
|
"loss": 0.3162, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.7475067378378095, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 2.0552901023890785e-05, |
|
"loss": 0.3101, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 0.7483456904055287, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 2.0484641638225254e-05, |
|
"loss": 0.3113, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 0.7491846429732479, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 2.041638225255973e-05, |
|
"loss": 0.2909, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 0.7500235955409671, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 2.03481228668942e-05, |
|
"loss": 0.3233, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 0.7508625481086864, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 2.027986348122867e-05, |
|
"loss": 0.2976, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.7517015006764055, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 2.021160409556314e-05, |
|
"loss": 0.3359, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 0.7525404532441247, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 2.0143344709897613e-05, |
|
"loss": 0.308, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 0.7533794058118439, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 2.0075085324232085e-05, |
|
"loss": 0.3131, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 0.7542183583795631, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 2.0006825938566554e-05, |
|
"loss": 0.3192, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 0.7550573109472823, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 1.9938566552901027e-05, |
|
"loss": 0.3064, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.7558962635150015, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 1.9870307167235496e-05, |
|
"loss": 0.2888, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 0.7567352160827208, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 1.980204778156997e-05, |
|
"loss": 0.3259, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 0.7575741686504399, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 1.9733788395904437e-05, |
|
"loss": 0.3083, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 0.7584131212181592, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 1.966552901023891e-05, |
|
"loss": 0.3019, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 0.7592520737858783, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 1.9597269624573382e-05, |
|
"loss": 0.2958, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 0.7600910263535975, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 1.952901023890785e-05, |
|
"loss": 0.292, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 0.7609299789213168, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 1.9460750853242324e-05, |
|
"loss": 0.3166, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 0.7617689314890359, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 1.9392491467576793e-05, |
|
"loss": 0.3226, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 0.7626078840567552, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 1.9324232081911265e-05, |
|
"loss": 0.3234, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 0.7634468366244743, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 1.9255972696245734e-05, |
|
"loss": 0.3238, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.7642857891921936, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 1.9187713310580207e-05, |
|
"loss": 0.3389, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 0.7651247417599127, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 1.9119453924914676e-05, |
|
"loss": 0.3019, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 0.765963694327632, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 1.905119453924915e-05, |
|
"loss": 0.3517, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 0.7668026468953512, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 1.898293515358362e-05, |
|
"loss": 0.2821, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 0.7676415994630703, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 1.891467576791809e-05, |
|
"loss": 0.318, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 0.7684805520307896, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 1.8846416382252562e-05, |
|
"loss": 0.3093, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 0.7693195045985087, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 1.877815699658703e-05, |
|
"loss": 0.3299, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 0.770158457166228, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 1.8709897610921504e-05, |
|
"loss": 0.3146, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 0.7709974097339471, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 1.8641638225255973e-05, |
|
"loss": 0.2951, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 0.7718363623016664, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 1.8573378839590445e-05, |
|
"loss": 0.3226, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.7726753148693856, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 1.8505119453924914e-05, |
|
"loss": 0.3018, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 0.7735142674371048, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 1.8436860068259387e-05, |
|
"loss": 0.3199, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 0.774353220004824, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 1.836860068259386e-05, |
|
"loss": 0.3067, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 0.7751921725725431, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 1.830034129692833e-05, |
|
"loss": 0.2861, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 0.7760311251402624, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 1.82320819112628e-05, |
|
"loss": 0.3188, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 0.7768700777079816, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 1.8163822525597273e-05, |
|
"loss": 0.3236, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 0.7777090302757008, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 1.8095563139931742e-05, |
|
"loss": 0.3075, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 0.77854798284342, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 1.802730375426621e-05, |
|
"loss": 0.308, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 0.7793869354111392, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 1.7959044368600684e-05, |
|
"loss": 0.3105, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 0.7802258879788584, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 1.7890784982935153e-05, |
|
"loss": 0.291, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.7810648405465775, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 1.7822525597269625e-05, |
|
"loss": 0.2903, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 0.7819037931142968, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 1.7754266211604098e-05, |
|
"loss": 0.3183, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 0.782742745682016, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 1.768600682593857e-05, |
|
"loss": 0.3169, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 0.7835816982497352, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 1.761774744027304e-05, |
|
"loss": 0.2839, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 0.7844206508174544, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 1.7549488054607512e-05, |
|
"loss": 0.2799, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 0.7852596033851736, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 1.748122866894198e-05, |
|
"loss": 0.3072, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 0.7860985559528928, |
|
"grad_norm": 0.375, |
|
"learning_rate": 1.7412969283276453e-05, |
|
"loss": 0.322, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 0.7869375085206121, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 1.7344709897610922e-05, |
|
"loss": 0.3228, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 0.7877764610883312, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 1.727645051194539e-05, |
|
"loss": 0.3282, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 0.7886154136560504, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 1.7208191126279864e-05, |
|
"loss": 0.3361, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.7894543662237696, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 1.7139931740614336e-05, |
|
"loss": 0.3411, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 0.7902933187914888, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 1.707167235494881e-05, |
|
"loss": 0.2957, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 0.791132271359208, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 1.7003412969283278e-05, |
|
"loss": 0.2747, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 0.7919712239269272, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 1.693515358361775e-05, |
|
"loss": 0.301, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 0.7928101764946465, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 1.686689419795222e-05, |
|
"loss": 0.3174, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 0.7936491290623656, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 1.6798634812286692e-05, |
|
"loss": 0.3069, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 0.7944880816300849, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 1.673037542662116e-05, |
|
"loss": 0.2877, |
|
"step": 9470 |
|
}, |
|
{ |
|
"epoch": 0.795327034197804, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 1.6662116040955633e-05, |
|
"loss": 0.3435, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 0.7961659867655232, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 1.6593856655290102e-05, |
|
"loss": 0.2862, |
|
"step": 9490 |
|
}, |
|
{ |
|
"epoch": 0.7970049393332425, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 1.6525597269624575e-05, |
|
"loss": 0.3139, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.7978438919009616, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 1.6457337883959047e-05, |
|
"loss": 0.3095, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 0.7986828444686809, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 1.6389078498293516e-05, |
|
"loss": 0.3284, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 0.7995217970364, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 1.632081911262799e-05, |
|
"loss": 0.3056, |
|
"step": 9530 |
|
}, |
|
{ |
|
"epoch": 0.8003607496041193, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 1.6252559726962458e-05, |
|
"loss": 0.3237, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 0.8011997021718384, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 1.618430034129693e-05, |
|
"loss": 0.2981, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 0.8020386547395577, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 1.61160409556314e-05, |
|
"loss": 0.3133, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 0.8028776073072769, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 1.6047781569965872e-05, |
|
"loss": 0.3057, |
|
"step": 9570 |
|
}, |
|
{ |
|
"epoch": 0.803716559874996, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 1.597952218430034e-05, |
|
"loss": 0.3068, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 0.8045555124427153, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 1.5911262798634813e-05, |
|
"loss": 0.3063, |
|
"step": 9590 |
|
}, |
|
{ |
|
"epoch": 0.8053944650104344, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 1.5843003412969286e-05, |
|
"loss": 0.3034, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.8062334175781537, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 1.5774744027303755e-05, |
|
"loss": 0.3159, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 0.8070723701458729, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 1.5706484641638227e-05, |
|
"loss": 0.3064, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 0.8079113227135921, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 1.5638225255972696e-05, |
|
"loss": 0.307, |
|
"step": 9630 |
|
}, |
|
{ |
|
"epoch": 0.8087502752813113, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 1.556996587030717e-05, |
|
"loss": 0.2971, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 0.8095892278490305, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 1.5501706484641638e-05, |
|
"loss": 0.2962, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 0.8104281804167497, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 1.543344709897611e-05, |
|
"loss": 0.312, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 0.8112671329844688, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 1.536518771331058e-05, |
|
"loss": 0.3082, |
|
"step": 9670 |
|
}, |
|
{ |
|
"epoch": 0.8121060855521881, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 1.5296928327645052e-05, |
|
"loss": 0.312, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 0.8129450381199073, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 1.5228668941979523e-05, |
|
"loss": 0.3184, |
|
"step": 9690 |
|
}, |
|
{ |
|
"epoch": 0.8137839906876265, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 1.5160409556313995e-05, |
|
"loss": 0.2945, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.8146229432553457, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 1.5092150170648466e-05, |
|
"loss": 0.3275, |
|
"step": 9710 |
|
}, |
|
{ |
|
"epoch": 0.8154618958230649, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 1.5023890784982937e-05, |
|
"loss": 0.3295, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 0.8163008483907841, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 1.4955631399317407e-05, |
|
"loss": 0.2932, |
|
"step": 9730 |
|
}, |
|
{ |
|
"epoch": 0.8171398009585034, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 1.4887372013651877e-05, |
|
"loss": 0.3053, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 0.8179787535262225, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 1.4819112627986349e-05, |
|
"loss": 0.3193, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 0.8188177060939417, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 1.475085324232082e-05, |
|
"loss": 0.3288, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 0.8196566586616609, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 1.4682593856655292e-05, |
|
"loss": 0.2928, |
|
"step": 9770 |
|
}, |
|
{ |
|
"epoch": 0.8204956112293801, |
|
"grad_norm": 0.5, |
|
"learning_rate": 1.4614334470989761e-05, |
|
"loss": 0.3261, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 0.8213345637970993, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 1.4546075085324234e-05, |
|
"loss": 0.2812, |
|
"step": 9790 |
|
}, |
|
{ |
|
"epoch": 0.8221735163648185, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 1.4477815699658704e-05, |
|
"loss": 0.2891, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.8230124689325378, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 1.4409556313993175e-05, |
|
"loss": 0.3091, |
|
"step": 9810 |
|
}, |
|
{ |
|
"epoch": 0.8238514215002569, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 1.4341296928327646e-05, |
|
"loss": 0.3233, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 0.8246903740679762, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 1.4273037542662118e-05, |
|
"loss": 0.2931, |
|
"step": 9830 |
|
}, |
|
{ |
|
"epoch": 0.8255293266356953, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 1.4204778156996588e-05, |
|
"loss": 0.327, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 0.8263682792034145, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 1.4136518771331058e-05, |
|
"loss": 0.3128, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 0.8272072317711338, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 1.406825938566553e-05, |
|
"loss": 0.3158, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 0.8280461843388529, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 1.4e-05, |
|
"loss": 0.3069, |
|
"step": 9870 |
|
}, |
|
{ |
|
"epoch": 0.8288851369065722, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 1.3931740614334472e-05, |
|
"loss": 0.2917, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 0.8297240894742913, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 1.3863481228668943e-05, |
|
"loss": 0.3108, |
|
"step": 9890 |
|
}, |
|
{ |
|
"epoch": 0.8305630420420106, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 1.3795221843003414e-05, |
|
"loss": 0.3268, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.8314019946097297, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 1.3726962457337885e-05, |
|
"loss": 0.3301, |
|
"step": 9910 |
|
}, |
|
{ |
|
"epoch": 0.832240947177449, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 1.3658703071672357e-05, |
|
"loss": 0.2925, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 0.8330798997451682, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 1.3590443686006826e-05, |
|
"loss": 0.307, |
|
"step": 9930 |
|
}, |
|
{ |
|
"epoch": 0.8339188523128873, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 1.3522184300341298e-05, |
|
"loss": 0.3352, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 0.8347578048806066, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 1.345392491467577e-05, |
|
"loss": 0.3126, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 0.8355967574483257, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 1.3385665529010242e-05, |
|
"loss": 0.3046, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 0.836435710016045, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 1.331740614334471e-05, |
|
"loss": 0.3024, |
|
"step": 9970 |
|
}, |
|
{ |
|
"epoch": 0.8372746625837642, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 1.3249146757679182e-05, |
|
"loss": 0.3161, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 0.8381136151514834, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 1.3180887372013654e-05, |
|
"loss": 0.3139, |
|
"step": 9990 |
|
}, |
|
{ |
|
"epoch": 0.8389525677192026, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 1.3112627986348123e-05, |
|
"loss": 0.3115, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.8397915202869218, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 1.3044368600682595e-05, |
|
"loss": 0.3043, |
|
"step": 10010 |
|
}, |
|
{ |
|
"epoch": 0.840630472854641, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 1.2976109215017065e-05, |
|
"loss": 0.3056, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 0.8414694254223601, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 1.2907849829351537e-05, |
|
"loss": 0.2873, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 0.8423083779900794, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 1.2839590443686008e-05, |
|
"loss": 0.2969, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 0.8431473305577986, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 1.277133105802048e-05, |
|
"loss": 0.2774, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 0.8439862831255178, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 1.270307167235495e-05, |
|
"loss": 0.276, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 0.844825235693237, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 1.2634812286689422e-05, |
|
"loss": 0.3276, |
|
"step": 10070 |
|
}, |
|
{ |
|
"epoch": 0.8456641882609562, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 1.2566552901023892e-05, |
|
"loss": 0.3473, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 0.8465031408286754, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 1.2498293515358362e-05, |
|
"loss": 0.297, |
|
"step": 10090 |
|
}, |
|
{ |
|
"epoch": 0.8473420933963945, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 1.2430034129692834e-05, |
|
"loss": 0.2876, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.8481810459641138, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 1.2361774744027305e-05, |
|
"loss": 0.277, |
|
"step": 10110 |
|
}, |
|
{ |
|
"epoch": 0.849019998531833, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 1.2293515358361776e-05, |
|
"loss": 0.3034, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 0.8498589510995522, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 1.2225255972696246e-05, |
|
"loss": 0.2847, |
|
"step": 10130 |
|
}, |
|
{ |
|
"epoch": 0.8506979036672714, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 1.2156996587030719e-05, |
|
"loss": 0.3247, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 0.8515368562349906, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 1.2088737201365188e-05, |
|
"loss": 0.3008, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 0.8523758088027098, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 1.202047781569966e-05, |
|
"loss": 0.3411, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 0.8532147613704291, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 1.1952218430034131e-05, |
|
"loss": 0.3085, |
|
"step": 10170 |
|
}, |
|
{ |
|
"epoch": 0.8540537139381482, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 1.1883959044368602e-05, |
|
"loss": 0.3263, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 0.8548926665058674, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 1.1815699658703073e-05, |
|
"loss": 0.3052, |
|
"step": 10190 |
|
}, |
|
{ |
|
"epoch": 0.8557316190735866, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 1.1747440273037543e-05, |
|
"loss": 0.266, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.8565705716413058, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 1.1679180887372014e-05, |
|
"loss": 0.3017, |
|
"step": 10210 |
|
}, |
|
{ |
|
"epoch": 0.857409524209025, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 1.1610921501706485e-05, |
|
"loss": 0.3197, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 0.8582484767767442, |
|
"grad_norm": 0.375, |
|
"learning_rate": 1.1542662116040957e-05, |
|
"loss": 0.3272, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 0.8590874293444635, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 1.1474402730375426e-05, |
|
"loss": 0.2973, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 0.8599263819121826, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 1.1406143344709899e-05, |
|
"loss": 0.3108, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 0.8607653344799019, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 1.133788395904437e-05, |
|
"loss": 0.2892, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 0.861604287047621, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 1.1269624573378842e-05, |
|
"loss": 0.31, |
|
"step": 10270 |
|
}, |
|
{ |
|
"epoch": 0.8624432396153402, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 1.1201365187713311e-05, |
|
"loss": 0.3138, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 0.8632821921830595, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 1.1133105802047784e-05, |
|
"loss": 0.302, |
|
"step": 10290 |
|
}, |
|
{ |
|
"epoch": 0.8641211447507786, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 1.1064846416382253e-05, |
|
"loss": 0.303, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.8649600973184979, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 1.0996587030716725e-05, |
|
"loss": 0.3183, |
|
"step": 10310 |
|
}, |
|
{ |
|
"epoch": 0.865799049886217, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 1.0928327645051196e-05, |
|
"loss": 0.3013, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 0.8666380024539363, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 1.0860068259385665e-05, |
|
"loss": 0.2986, |
|
"step": 10330 |
|
}, |
|
{ |
|
"epoch": 0.8674769550216554, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 1.0791808873720137e-05, |
|
"loss": 0.3251, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 0.8683159075893747, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 1.0723549488054608e-05, |
|
"loss": 0.3291, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 0.8691548601570939, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 1.065529010238908e-05, |
|
"loss": 0.2905, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 0.869993812724813, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 1.058703071672355e-05, |
|
"loss": 0.3071, |
|
"step": 10370 |
|
}, |
|
{ |
|
"epoch": 0.8708327652925323, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 1.0518771331058022e-05, |
|
"loss": 0.3146, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 0.8716717178602514, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 1.0450511945392491e-05, |
|
"loss": 0.2967, |
|
"step": 10390 |
|
}, |
|
{ |
|
"epoch": 0.8725106704279707, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 1.0382252559726964e-05, |
|
"loss": 0.2679, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.8733496229956899, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 1.0313993174061434e-05, |
|
"loss": 0.3069, |
|
"step": 10410 |
|
}, |
|
{ |
|
"epoch": 0.8741885755634091, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 1.0245733788395907e-05, |
|
"loss": 0.2903, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 0.8750275281311283, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 1.0177474402730376e-05, |
|
"loss": 0.2657, |
|
"step": 10430 |
|
}, |
|
{ |
|
"epoch": 0.8758664806988475, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 1.0109215017064847e-05, |
|
"loss": 0.3163, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 0.8767054332665667, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 1.0040955631399319e-05, |
|
"loss": 0.3044, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 0.8775443858342858, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 9.97269624573379e-06, |
|
"loss": 0.3208, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 0.8783833384020051, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 9.90443686006826e-06, |
|
"loss": 0.2919, |
|
"step": 10470 |
|
}, |
|
{ |
|
"epoch": 0.8792222909697243, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 9.836177474402731e-06, |
|
"loss": 0.309, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 0.8800612435374435, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 9.767918088737202e-06, |
|
"loss": 0.3021, |
|
"step": 10490 |
|
}, |
|
{ |
|
"epoch": 0.8809001961051627, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 9.699658703071673e-06, |
|
"loss": 0.3156, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.881739148672882, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 9.631399317406144e-06, |
|
"loss": 0.2878, |
|
"step": 10510 |
|
}, |
|
{ |
|
"epoch": 0.8825781012406011, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 9.563139931740614e-06, |
|
"loss": 0.3264, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 0.8834170538083204, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 9.494880546075085e-06, |
|
"loss": 0.3046, |
|
"step": 10530 |
|
}, |
|
{ |
|
"epoch": 0.8842560063760395, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 9.426621160409558e-06, |
|
"loss": 0.3141, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 0.8850949589437587, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 9.358361774744028e-06, |
|
"loss": 0.3095, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 0.8859339115114779, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 9.290102389078499e-06, |
|
"loss": 0.3005, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 0.8867728640791971, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 9.22184300341297e-06, |
|
"loss": 0.2936, |
|
"step": 10570 |
|
}, |
|
{ |
|
"epoch": 0.8876118166469164, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 9.15358361774744e-06, |
|
"loss": 0.3146, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 0.8884507692146355, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 9.085324232081913e-06, |
|
"loss": 0.2872, |
|
"step": 10590 |
|
}, |
|
{ |
|
"epoch": 0.8892897217823548, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 9.017064846416384e-06, |
|
"loss": 0.297, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.8901286743500739, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 8.948805460750853e-06, |
|
"loss": 0.2811, |
|
"step": 10610 |
|
}, |
|
{ |
|
"epoch": 0.8909676269177932, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 8.880546075085325e-06, |
|
"loss": 0.3066, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 0.8918065794855123, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 8.812286689419796e-06, |
|
"loss": 0.2821, |
|
"step": 10630 |
|
}, |
|
{ |
|
"epoch": 0.8926455320532315, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 8.744027303754267e-06, |
|
"loss": 0.311, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 0.8934844846209508, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 8.675767918088738e-06, |
|
"loss": 0.2785, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 0.8943234371886699, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 8.607508532423208e-06, |
|
"loss": 0.3074, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 0.8951623897563892, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 8.539249146757679e-06, |
|
"loss": 0.2939, |
|
"step": 10670 |
|
}, |
|
{ |
|
"epoch": 0.8960013423241083, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 8.470989761092152e-06, |
|
"loss": 0.3236, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 0.8968402948918276, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 8.402730375426622e-06, |
|
"loss": 0.2945, |
|
"step": 10690 |
|
}, |
|
{ |
|
"epoch": 0.8976792474595467, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 8.334470989761093e-06, |
|
"loss": 0.3205, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.898518200027266, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 8.266211604095564e-06, |
|
"loss": 0.3004, |
|
"step": 10710 |
|
}, |
|
{ |
|
"epoch": 0.8993571525949852, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 8.197952218430035e-06, |
|
"loss": 0.2815, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 0.9001961051627043, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 8.129692832764505e-06, |
|
"loss": 0.3134, |
|
"step": 10730 |
|
}, |
|
{ |
|
"epoch": 0.9010350577304236, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 8.061433447098976e-06, |
|
"loss": 0.299, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 0.9018740102981427, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 7.993174061433447e-06, |
|
"loss": 0.3134, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 0.902712962865862, |
|
"grad_norm": 0.30078125, |
|
"learning_rate": 7.92491467576792e-06, |
|
"loss": 0.3037, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 0.9035519154335812, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 7.85665529010239e-06, |
|
"loss": 0.3105, |
|
"step": 10770 |
|
}, |
|
{ |
|
"epoch": 0.9043908680013004, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 7.788395904436861e-06, |
|
"loss": 0.2843, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 0.9052298205690196, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 7.720136518771332e-06, |
|
"loss": 0.309, |
|
"step": 10790 |
|
}, |
|
{ |
|
"epoch": 0.9060687731367388, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 7.651877133105802e-06, |
|
"loss": 0.3122, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.906907725704458, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 7.583617747440274e-06, |
|
"loss": 0.2958, |
|
"step": 10810 |
|
}, |
|
{ |
|
"epoch": 0.9077466782721771, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 7.515358361774745e-06, |
|
"loss": 0.2904, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 0.9085856308398964, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 7.447098976109216e-06, |
|
"loss": 0.314, |
|
"step": 10830 |
|
}, |
|
{ |
|
"epoch": 0.9094245834076156, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 7.378839590443687e-06, |
|
"loss": 0.2947, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 0.9102635359753348, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 7.310580204778157e-06, |
|
"loss": 0.2921, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 0.911102488543054, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 7.242320819112628e-06, |
|
"loss": 0.3032, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 0.9119414411107732, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 7.174061433447099e-06, |
|
"loss": 0.2936, |
|
"step": 10870 |
|
}, |
|
{ |
|
"epoch": 0.9127803936784924, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 7.10580204778157e-06, |
|
"loss": 0.3046, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 0.9136193462462115, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 7.037542662116042e-06, |
|
"loss": 0.3087, |
|
"step": 10890 |
|
}, |
|
{ |
|
"epoch": 0.9144582988139308, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 6.9692832764505125e-06, |
|
"loss": 0.2993, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.91529725138165, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 6.901023890784983e-06, |
|
"loss": 0.3018, |
|
"step": 10910 |
|
}, |
|
{ |
|
"epoch": 0.9161362039493692, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 6.832764505119455e-06, |
|
"loss": 0.2988, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 0.9169751565170884, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 6.764505119453926e-06, |
|
"loss": 0.2871, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 0.9178141090848076, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 6.696245733788397e-06, |
|
"loss": 0.3117, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 0.9186530616525268, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 6.627986348122868e-06, |
|
"loss": 0.2891, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 0.9194920142202461, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 6.559726962457338e-06, |
|
"loss": 0.305, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 0.9203309667879652, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 6.491467576791809e-06, |
|
"loss": 0.323, |
|
"step": 10970 |
|
}, |
|
{ |
|
"epoch": 0.9211699193556844, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 6.42320819112628e-06, |
|
"loss": 0.2954, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 0.9220088719234036, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 6.354948805460751e-06, |
|
"loss": 0.2797, |
|
"step": 10990 |
|
}, |
|
{ |
|
"epoch": 0.9228478244911228, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 6.286689419795222e-06, |
|
"loss": 0.2958, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.9236867770588421, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 6.218430034129693e-06, |
|
"loss": 0.3124, |
|
"step": 11010 |
|
}, |
|
{ |
|
"epoch": 0.9245257296265612, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 6.150170648464164e-06, |
|
"loss": 0.2913, |
|
"step": 11020 |
|
}, |
|
{ |
|
"epoch": 0.9253646821942805, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 6.081911262798636e-06, |
|
"loss": 0.2895, |
|
"step": 11030 |
|
}, |
|
{ |
|
"epoch": 0.9262036347619996, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 6.0136518771331065e-06, |
|
"loss": 0.3067, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 0.9270425873297189, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 5.945392491467577e-06, |
|
"loss": 0.3171, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 0.927881539897438, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 5.877133105802049e-06, |
|
"loss": 0.3025, |
|
"step": 11060 |
|
}, |
|
{ |
|
"epoch": 0.9287204924651572, |
|
"grad_norm": 0.375, |
|
"learning_rate": 5.80887372013652e-06, |
|
"loss": 0.2783, |
|
"step": 11070 |
|
}, |
|
{ |
|
"epoch": 0.9295594450328765, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 5.7406143344709895e-06, |
|
"loss": 0.3012, |
|
"step": 11080 |
|
}, |
|
{ |
|
"epoch": 0.9303983976005956, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 5.672354948805461e-06, |
|
"loss": 0.3142, |
|
"step": 11090 |
|
}, |
|
{ |
|
"epoch": 0.9312373501683149, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 5.604095563139932e-06, |
|
"loss": 0.3134, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.932076302736034, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 5.535836177474403e-06, |
|
"loss": 0.2859, |
|
"step": 11110 |
|
}, |
|
{ |
|
"epoch": 0.9329152553037533, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 5.467576791808874e-06, |
|
"loss": 0.3346, |
|
"step": 11120 |
|
}, |
|
{ |
|
"epoch": 0.9337542078714725, |
|
"grad_norm": 0.375, |
|
"learning_rate": 5.399317406143345e-06, |
|
"loss": 0.3141, |
|
"step": 11130 |
|
}, |
|
{ |
|
"epoch": 0.9345931604391917, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 5.331058020477816e-06, |
|
"loss": 0.3127, |
|
"step": 11140 |
|
}, |
|
{ |
|
"epoch": 0.9354321130069109, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 5.262798634812287e-06, |
|
"loss": 0.2821, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 0.93627106557463, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 5.194539249146758e-06, |
|
"loss": 0.2895, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 0.9371100181423493, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 5.12627986348123e-06, |
|
"loss": 0.3041, |
|
"step": 11170 |
|
}, |
|
{ |
|
"epoch": 0.9379489707100684, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 5.0580204778157005e-06, |
|
"loss": 0.3035, |
|
"step": 11180 |
|
}, |
|
{ |
|
"epoch": 0.9387879232777877, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 4.989761092150171e-06, |
|
"loss": 0.2749, |
|
"step": 11190 |
|
}, |
|
{ |
|
"epoch": 0.9396268758455069, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 4.921501706484642e-06, |
|
"loss": 0.3236, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.9404658284132261, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 4.853242320819113e-06, |
|
"loss": 0.3138, |
|
"step": 11210 |
|
}, |
|
{ |
|
"epoch": 0.9413047809809453, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 4.784982935153584e-06, |
|
"loss": 0.2989, |
|
"step": 11220 |
|
}, |
|
{ |
|
"epoch": 0.9421437335486645, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 4.716723549488055e-06, |
|
"loss": 0.3034, |
|
"step": 11230 |
|
}, |
|
{ |
|
"epoch": 0.9429826861163837, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 4.648464163822526e-06, |
|
"loss": 0.3193, |
|
"step": 11240 |
|
}, |
|
{ |
|
"epoch": 0.9438216386841028, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 4.580204778156997e-06, |
|
"loss": 0.2978, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 0.9446605912518221, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 4.511945392491468e-06, |
|
"loss": 0.3112, |
|
"step": 11260 |
|
}, |
|
{ |
|
"epoch": 0.9454995438195413, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 4.443686006825939e-06, |
|
"loss": 0.2959, |
|
"step": 11270 |
|
}, |
|
{ |
|
"epoch": 0.9463384963872605, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 4.37542662116041e-06, |
|
"loss": 0.2823, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 0.9471774489549797, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 4.3071672354948806e-06, |
|
"loss": 0.326, |
|
"step": 11290 |
|
}, |
|
{ |
|
"epoch": 0.948016401522699, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 4.238907849829352e-06, |
|
"loss": 0.3031, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.9488553540904181, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 4.170648464163823e-06, |
|
"loss": 0.3171, |
|
"step": 11310 |
|
}, |
|
{ |
|
"epoch": 0.9496943066581374, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 4.102389078498294e-06, |
|
"loss": 0.3, |
|
"step": 11320 |
|
}, |
|
{ |
|
"epoch": 0.9505332592258565, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 4.034129692832765e-06, |
|
"loss": 0.3284, |
|
"step": 11330 |
|
}, |
|
{ |
|
"epoch": 0.9513722117935757, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 3.965870307167236e-06, |
|
"loss": 0.291, |
|
"step": 11340 |
|
}, |
|
{ |
|
"epoch": 0.9522111643612949, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 3.897610921501707e-06, |
|
"loss": 0.3076, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 0.9530501169290141, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 3.8293515358361776e-06, |
|
"loss": 0.3199, |
|
"step": 11360 |
|
}, |
|
{ |
|
"epoch": 0.9538890694967334, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 3.7610921501706487e-06, |
|
"loss": 0.2884, |
|
"step": 11370 |
|
}, |
|
{ |
|
"epoch": 0.9547280220644525, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 3.69283276450512e-06, |
|
"loss": 0.2989, |
|
"step": 11380 |
|
}, |
|
{ |
|
"epoch": 0.9555669746321718, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 3.624573378839591e-06, |
|
"loss": 0.3027, |
|
"step": 11390 |
|
}, |
|
{ |
|
"epoch": 0.9564059271998909, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 3.5563139931740614e-06, |
|
"loss": 0.291, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.9572448797676102, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 3.4880546075085326e-06, |
|
"loss": 0.3165, |
|
"step": 11410 |
|
}, |
|
{ |
|
"epoch": 0.9580838323353293, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 3.419795221843004e-06, |
|
"loss": 0.3191, |
|
"step": 11420 |
|
}, |
|
{ |
|
"epoch": 0.9589227849030485, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 3.3515358361774746e-06, |
|
"loss": 0.3036, |
|
"step": 11430 |
|
}, |
|
{ |
|
"epoch": 0.9597617374707678, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 3.2832764505119457e-06, |
|
"loss": 0.3068, |
|
"step": 11440 |
|
}, |
|
{ |
|
"epoch": 0.9606006900384869, |
|
"grad_norm": 0.375, |
|
"learning_rate": 3.215017064846417e-06, |
|
"loss": 0.294, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 0.9614396426062062, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 3.1467576791808873e-06, |
|
"loss": 0.2933, |
|
"step": 11460 |
|
}, |
|
{ |
|
"epoch": 0.9622785951739253, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 3.0784982935153584e-06, |
|
"loss": 0.3109, |
|
"step": 11470 |
|
}, |
|
{ |
|
"epoch": 0.9631175477416446, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 3.0102389078498296e-06, |
|
"loss": 0.3117, |
|
"step": 11480 |
|
}, |
|
{ |
|
"epoch": 0.9639565003093638, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 2.941979522184301e-06, |
|
"loss": 0.2911, |
|
"step": 11490 |
|
}, |
|
{ |
|
"epoch": 0.964795452877083, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 2.8737201365187716e-06, |
|
"loss": 0.283, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.9656344054448022, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 2.8054607508532428e-06, |
|
"loss": 0.2966, |
|
"step": 11510 |
|
}, |
|
{ |
|
"epoch": 0.9664733580125213, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 2.737201365187713e-06, |
|
"loss": 0.298, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 0.9673123105802406, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 2.6689419795221843e-06, |
|
"loss": 0.3066, |
|
"step": 11530 |
|
}, |
|
{ |
|
"epoch": 0.9681512631479597, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 2.6006825938566555e-06, |
|
"loss": 0.3095, |
|
"step": 11540 |
|
}, |
|
{ |
|
"epoch": 0.968990215715679, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 2.5324232081911266e-06, |
|
"loss": 0.3367, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 0.9698291682833982, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 2.4641638225255974e-06, |
|
"loss": 0.3166, |
|
"step": 11560 |
|
}, |
|
{ |
|
"epoch": 0.9706681208511174, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 2.3959044368600686e-06, |
|
"loss": 0.2939, |
|
"step": 11570 |
|
}, |
|
{ |
|
"epoch": 0.9715070734188366, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 2.3276450511945393e-06, |
|
"loss": 0.3145, |
|
"step": 11580 |
|
}, |
|
{ |
|
"epoch": 0.9723460259865558, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 2.25938566552901e-06, |
|
"loss": 0.2965, |
|
"step": 11590 |
|
}, |
|
{ |
|
"epoch": 0.973184978554275, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 2.1911262798634813e-06, |
|
"loss": 0.3126, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.9740239311219941, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 2.1228668941979525e-06, |
|
"loss": 0.311, |
|
"step": 11610 |
|
}, |
|
{ |
|
"epoch": 0.9748628836897134, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 2.0546075085324232e-06, |
|
"loss": 0.2994, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 0.9757018362574326, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 1.9863481228668944e-06, |
|
"loss": 0.3164, |
|
"step": 11630 |
|
}, |
|
{ |
|
"epoch": 0.9765407888251518, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 1.9180887372013656e-06, |
|
"loss": 0.3058, |
|
"step": 11640 |
|
}, |
|
{ |
|
"epoch": 0.977379741392871, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 1.8498293515358361e-06, |
|
"loss": 0.3194, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 0.9782186939605902, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 1.7815699658703073e-06, |
|
"loss": 0.3018, |
|
"step": 11660 |
|
}, |
|
{ |
|
"epoch": 0.9790576465283094, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 1.7133105802047783e-06, |
|
"loss": 0.2752, |
|
"step": 11670 |
|
}, |
|
{ |
|
"epoch": 0.9798965990960286, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 1.6450511945392493e-06, |
|
"loss": 0.3122, |
|
"step": 11680 |
|
}, |
|
{ |
|
"epoch": 0.9807355516637478, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 1.5767918088737202e-06, |
|
"loss": 0.3016, |
|
"step": 11690 |
|
}, |
|
{ |
|
"epoch": 0.981574504231467, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 1.5085324232081914e-06, |
|
"loss": 0.3003, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.9824134567991862, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 1.4402730375426622e-06, |
|
"loss": 0.2934, |
|
"step": 11710 |
|
}, |
|
{ |
|
"epoch": 0.9832524093669054, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 1.3720136518771331e-06, |
|
"loss": 0.2947, |
|
"step": 11720 |
|
}, |
|
{ |
|
"epoch": 0.9840913619346247, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 1.3037542662116043e-06, |
|
"loss": 0.2875, |
|
"step": 11730 |
|
}, |
|
{ |
|
"epoch": 0.9849303145023438, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 1.2354948805460753e-06, |
|
"loss": 0.3108, |
|
"step": 11740 |
|
}, |
|
{ |
|
"epoch": 0.9857692670700631, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 1.1672354948805463e-06, |
|
"loss": 0.3044, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 0.9866082196377822, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 1.098976109215017e-06, |
|
"loss": 0.3012, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 0.9874471722055014, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 1.0307167235494882e-06, |
|
"loss": 0.3231, |
|
"step": 11770 |
|
}, |
|
{ |
|
"epoch": 0.9882861247732206, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 9.624573378839592e-07, |
|
"loss": 0.3188, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 0.9891250773409398, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 8.9419795221843e-07, |
|
"loss": 0.3022, |
|
"step": 11790 |
|
}, |
|
{ |
|
"epoch": 0.9899640299086591, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 8.259385665529011e-07, |
|
"loss": 0.3228, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.9908029824763782, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 7.576791808873721e-07, |
|
"loss": 0.2862, |
|
"step": 11810 |
|
}, |
|
{ |
|
"epoch": 0.9916419350440975, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 6.89419795221843e-07, |
|
"loss": 0.3241, |
|
"step": 11820 |
|
}, |
|
{ |
|
"epoch": 0.9924808876118166, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 6.21160409556314e-07, |
|
"loss": 0.3101, |
|
"step": 11830 |
|
}, |
|
{ |
|
"epoch": 0.9933198401795359, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 5.52901023890785e-07, |
|
"loss": 0.286, |
|
"step": 11840 |
|
}, |
|
{ |
|
"epoch": 0.994158792747255, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 4.84641638225256e-07, |
|
"loss": 0.3094, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 0.9949977453149742, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 4.16382252559727e-07, |
|
"loss": 0.2974, |
|
"step": 11860 |
|
}, |
|
{ |
|
"epoch": 0.9958366978826935, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 3.48122866894198e-07, |
|
"loss": 0.3095, |
|
"step": 11870 |
|
}, |
|
{ |
|
"epoch": 0.9966756504504126, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 2.79863481228669e-07, |
|
"loss": 0.299, |
|
"step": 11880 |
|
}, |
|
{ |
|
"epoch": 0.9975146030181319, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 2.1160409556313995e-07, |
|
"loss": 0.2975, |
|
"step": 11890 |
|
}, |
|
{ |
|
"epoch": 0.998353555585851, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 1.4334470989761095e-07, |
|
"loss": 0.3266, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.9991925081535703, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 7.508532423208192e-08, |
|
"loss": 0.2895, |
|
"step": 11910 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 6.825938566552902e-09, |
|
"loss": 0.3005, |
|
"step": 11920 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 11920, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.635149811167468e+19, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|