youssefkhalil320's picture
Upload folder using huggingface_hub
eb841f1 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.8919350295746878,
"eval_steps": 5000,
"global_step": 9500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.009388789784996713,
"grad_norm": 87.77811431884766,
"learning_rate": 9.009854528390429e-07,
"loss": 16.2337,
"step": 100
},
{
"epoch": 0.018777579569993427,
"grad_norm": 85.86430358886719,
"learning_rate": 1.8395119662130456e-06,
"loss": 13.5901,
"step": 200
},
{
"epoch": 0.02816636935499014,
"grad_norm": 14.885255813598633,
"learning_rate": 2.7592679493195683e-06,
"loss": 9.8565,
"step": 300
},
{
"epoch": 0.03755515913998685,
"grad_norm": 6.9691972732543945,
"learning_rate": 3.6977944626935713e-06,
"loss": 8.3332,
"step": 400
},
{
"epoch": 0.04694394892498357,
"grad_norm": 5.612818241119385,
"learning_rate": 4.6363209760675744e-06,
"loss": 8.1261,
"step": 500
},
{
"epoch": 0.05633273870998028,
"grad_norm": 4.705409526824951,
"learning_rate": 5.574847489441577e-06,
"loss": 8.0697,
"step": 600
},
{
"epoch": 0.06572152849497699,
"grad_norm": 4.337332725524902,
"learning_rate": 6.51337400281558e-06,
"loss": 8.0298,
"step": 700
},
{
"epoch": 0.0751103182799737,
"grad_norm": 3.6314213275909424,
"learning_rate": 7.451900516189583e-06,
"loss": 8.033,
"step": 800
},
{
"epoch": 0.08449910806497042,
"grad_norm": 3.4845075607299805,
"learning_rate": 8.390427029563585e-06,
"loss": 7.9858,
"step": 900
},
{
"epoch": 0.09388789784996714,
"grad_norm": 5.188210487365723,
"learning_rate": 9.328953542937589e-06,
"loss": 8.012,
"step": 1000
},
{
"epoch": 0.10327668763496385,
"grad_norm": 3.0830442905426025,
"learning_rate": 1.0267480056311592e-05,
"loss": 7.9745,
"step": 1100
},
{
"epoch": 0.11266547741996057,
"grad_norm": 3.4729278087615967,
"learning_rate": 1.1206006569685594e-05,
"loss": 8.0091,
"step": 1200
},
{
"epoch": 0.12205426720495728,
"grad_norm": 2.329235076904297,
"learning_rate": 1.2144533083059597e-05,
"loss": 8.0221,
"step": 1300
},
{
"epoch": 0.13144305698995398,
"grad_norm": 2.7225279808044434,
"learning_rate": 1.3083059596433601e-05,
"loss": 7.9583,
"step": 1400
},
{
"epoch": 0.1408318467749507,
"grad_norm": 2.012805938720703,
"learning_rate": 1.4021586109807603e-05,
"loss": 8.0031,
"step": 1500
},
{
"epoch": 0.1502206365599474,
"grad_norm": 2.9397523403167725,
"learning_rate": 1.4960112623181606e-05,
"loss": 7.9985,
"step": 1600
},
{
"epoch": 0.15960942634494413,
"grad_norm": 2.356337308883667,
"learning_rate": 1.589863913655561e-05,
"loss": 7.9647,
"step": 1700
},
{
"epoch": 0.16899821612994084,
"grad_norm": 2.6846818923950195,
"learning_rate": 1.6837165649929613e-05,
"loss": 7.9857,
"step": 1800
},
{
"epoch": 0.17838700591493756,
"grad_norm": 2.0188565254211426,
"learning_rate": 1.7775692163303613e-05,
"loss": 7.9806,
"step": 1900
},
{
"epoch": 0.18777579569993427,
"grad_norm": 4.030488014221191,
"learning_rate": 1.8714218676677617e-05,
"loss": 7.9761,
"step": 2000
},
{
"epoch": 0.197164585484931,
"grad_norm": 4.183101654052734,
"learning_rate": 1.965274519005162e-05,
"loss": 7.9696,
"step": 2100
},
{
"epoch": 0.2065533752699277,
"grad_norm": 1.4769889116287231,
"learning_rate": 1.9934275728965626e-05,
"loss": 8.0014,
"step": 2200
},
{
"epoch": 0.21594216505492442,
"grad_norm": 2.1914358139038086,
"learning_rate": 1.9829951489228525e-05,
"loss": 7.9546,
"step": 2300
},
{
"epoch": 0.22533095483992113,
"grad_norm": 22.55516815185547,
"learning_rate": 1.972562724949142e-05,
"loss": 7.9874,
"step": 2400
},
{
"epoch": 0.23471974462491785,
"grad_norm": 1.635116457939148,
"learning_rate": 1.962130300975432e-05,
"loss": 7.9846,
"step": 2500
},
{
"epoch": 0.24410853440991456,
"grad_norm": 5.707275390625,
"learning_rate": 1.9516978770017215e-05,
"loss": 7.9664,
"step": 2600
},
{
"epoch": 0.2534973241949113,
"grad_norm": 4.194604396820068,
"learning_rate": 1.9412654530280113e-05,
"loss": 7.9725,
"step": 2700
},
{
"epoch": 0.26288611397990796,
"grad_norm": 2.0074055194854736,
"learning_rate": 1.930833029054301e-05,
"loss": 7.9419,
"step": 2800
},
{
"epoch": 0.2722749037649047,
"grad_norm": 26.4300479888916,
"learning_rate": 1.9204006050805904e-05,
"loss": 7.9786,
"step": 2900
},
{
"epoch": 0.2816636935499014,
"grad_norm": 2.5870931148529053,
"learning_rate": 1.9099681811068803e-05,
"loss": 7.9479,
"step": 3000
},
{
"epoch": 0.29105248333489814,
"grad_norm": 1.6209933757781982,
"learning_rate": 1.8995357571331702e-05,
"loss": 7.9526,
"step": 3100
},
{
"epoch": 0.3004412731198948,
"grad_norm": 19.398080825805664,
"learning_rate": 1.8891033331594598e-05,
"loss": 7.9613,
"step": 3200
},
{
"epoch": 0.30983006290489157,
"grad_norm": 2.124729871749878,
"learning_rate": 1.8786709091857496e-05,
"loss": 7.9994,
"step": 3300
},
{
"epoch": 0.31921885268988826,
"grad_norm": 2.701019763946533,
"learning_rate": 1.8682384852120392e-05,
"loss": 7.9464,
"step": 3400
},
{
"epoch": 0.328607642474885,
"grad_norm": 1.944600224494934,
"learning_rate": 1.8578060612383287e-05,
"loss": 7.9429,
"step": 3500
},
{
"epoch": 0.3379964322598817,
"grad_norm": 41.85493087768555,
"learning_rate": 1.8473736372646186e-05,
"loss": 7.9539,
"step": 3600
},
{
"epoch": 0.34738522204487843,
"grad_norm": 2.024019479751587,
"learning_rate": 1.8369412132909085e-05,
"loss": 7.9699,
"step": 3700
},
{
"epoch": 0.3567740118298751,
"grad_norm": 2.2833781242370605,
"learning_rate": 1.826508789317198e-05,
"loss": 7.9144,
"step": 3800
},
{
"epoch": 0.36616280161487186,
"grad_norm": 3.1666574478149414,
"learning_rate": 1.8160763653434876e-05,
"loss": 7.9424,
"step": 3900
},
{
"epoch": 0.37555159139986855,
"grad_norm": 1.7259443998336792,
"learning_rate": 1.8056439413697775e-05,
"loss": 7.9361,
"step": 4000
},
{
"epoch": 0.3849403811848653,
"grad_norm": 2.5588905811309814,
"learning_rate": 1.795211517396067e-05,
"loss": 7.9144,
"step": 4100
},
{
"epoch": 0.394329170969862,
"grad_norm": 2.542963743209839,
"learning_rate": 1.784779093422357e-05,
"loss": 7.907,
"step": 4200
},
{
"epoch": 0.4037179607548587,
"grad_norm": 2.755725622177124,
"learning_rate": 1.7743466694486468e-05,
"loss": 7.9049,
"step": 4300
},
{
"epoch": 0.4131067505398554,
"grad_norm": 2.6067683696746826,
"learning_rate": 1.7639142454749364e-05,
"loss": 7.939,
"step": 4400
},
{
"epoch": 0.42249554032485215,
"grad_norm": 2.614475965499878,
"learning_rate": 1.753481821501226e-05,
"loss": 7.9067,
"step": 4500
},
{
"epoch": 0.43188433010984884,
"grad_norm": 2.172943353652954,
"learning_rate": 1.7430493975275155e-05,
"loss": 7.9149,
"step": 4600
},
{
"epoch": 0.4412731198948456,
"grad_norm": 6.292716979980469,
"learning_rate": 1.7326169735538053e-05,
"loss": 7.9705,
"step": 4700
},
{
"epoch": 0.45066190967984227,
"grad_norm": 2.595337152481079,
"learning_rate": 1.7221845495800952e-05,
"loss": 7.8992,
"step": 4800
},
{
"epoch": 0.460050699464839,
"grad_norm": 5.529814720153809,
"learning_rate": 1.7117521256063848e-05,
"loss": 7.9077,
"step": 4900
},
{
"epoch": 0.4694394892498357,
"grad_norm": 6.696155548095703,
"learning_rate": 1.7013197016326747e-05,
"loss": 7.8992,
"step": 5000
},
{
"epoch": 0.4694394892498357,
"eval_loss": 7.916718006134033,
"eval_runtime": 334.3618,
"eval_samples_per_second": 1019.336,
"eval_steps_per_second": 7.964,
"step": 5000
},
{
"epoch": 0.4788282790348324,
"grad_norm": 26.500988006591797,
"learning_rate": 1.6908872776589642e-05,
"loss": 7.914,
"step": 5100
},
{
"epoch": 0.4882170688198291,
"grad_norm": 2.454939126968384,
"learning_rate": 1.6804548536852537e-05,
"loss": 7.8913,
"step": 5200
},
{
"epoch": 0.4976058586048258,
"grad_norm": 1.8935959339141846,
"learning_rate": 1.6700224297115436e-05,
"loss": 7.8999,
"step": 5300
},
{
"epoch": 0.5069946483898226,
"grad_norm": 2.2066667079925537,
"learning_rate": 1.6595900057378335e-05,
"loss": 7.8818,
"step": 5400
},
{
"epoch": 0.5163834381748192,
"grad_norm": 2.7672908306121826,
"learning_rate": 1.649157581764123e-05,
"loss": 7.9383,
"step": 5500
},
{
"epoch": 0.5257722279598159,
"grad_norm": 4.411477088928223,
"learning_rate": 1.6387251577904126e-05,
"loss": 7.9094,
"step": 5600
},
{
"epoch": 0.5351610177448127,
"grad_norm": 50.536991119384766,
"learning_rate": 1.6282927338167025e-05,
"loss": 7.8986,
"step": 5700
},
{
"epoch": 0.5445498075298094,
"grad_norm": 2.9657583236694336,
"learning_rate": 1.617860309842992e-05,
"loss": 7.9015,
"step": 5800
},
{
"epoch": 0.5539385973148061,
"grad_norm": 1.910575270652771,
"learning_rate": 1.607427885869282e-05,
"loss": 7.9059,
"step": 5900
},
{
"epoch": 0.5633273870998028,
"grad_norm": 3.9665298461914062,
"learning_rate": 1.5969954618955715e-05,
"loss": 7.8524,
"step": 6000
},
{
"epoch": 0.5727161768847996,
"grad_norm": 2.0286688804626465,
"learning_rate": 1.5865630379218614e-05,
"loss": 7.8788,
"step": 6100
},
{
"epoch": 0.5821049666697963,
"grad_norm": 3.200676202774048,
"learning_rate": 1.576130613948151e-05,
"loss": 7.8712,
"step": 6200
},
{
"epoch": 0.591493756454793,
"grad_norm": 2.0278351306915283,
"learning_rate": 1.5656981899744408e-05,
"loss": 7.8967,
"step": 6300
},
{
"epoch": 0.6008825462397896,
"grad_norm": 3.4500420093536377,
"learning_rate": 1.5553700902404677e-05,
"loss": 7.8677,
"step": 6400
},
{
"epoch": 0.6102713360247864,
"grad_norm": 2.8257575035095215,
"learning_rate": 1.5450419905064945e-05,
"loss": 7.9132,
"step": 6500
},
{
"epoch": 0.6196601258097831,
"grad_norm": 2.537797212600708,
"learning_rate": 1.534609566532784e-05,
"loss": 7.853,
"step": 6600
},
{
"epoch": 0.6290489155947798,
"grad_norm": 3.010673999786377,
"learning_rate": 1.5241771425590736e-05,
"loss": 7.8968,
"step": 6700
},
{
"epoch": 0.6384377053797765,
"grad_norm": 2.7398829460144043,
"learning_rate": 1.5137447185853635e-05,
"loss": 7.8656,
"step": 6800
},
{
"epoch": 0.6478264951647733,
"grad_norm": 4.402348518371582,
"learning_rate": 1.5033122946116532e-05,
"loss": 7.8801,
"step": 6900
},
{
"epoch": 0.65721528494977,
"grad_norm": 2.982344388961792,
"learning_rate": 1.4928798706379428e-05,
"loss": 7.8378,
"step": 7000
},
{
"epoch": 0.6666040747347667,
"grad_norm": 12.262134552001953,
"learning_rate": 1.4824474466642325e-05,
"loss": 7.8554,
"step": 7100
},
{
"epoch": 0.6759928645197634,
"grad_norm": 3.8007514476776123,
"learning_rate": 1.4720150226905224e-05,
"loss": 7.8305,
"step": 7200
},
{
"epoch": 0.6853816543047602,
"grad_norm": 3.8391048908233643,
"learning_rate": 1.461582598716812e-05,
"loss": 7.8613,
"step": 7300
},
{
"epoch": 0.6947704440897569,
"grad_norm": 3.107123374938965,
"learning_rate": 1.4511501747431017e-05,
"loss": 7.8554,
"step": 7400
},
{
"epoch": 0.7041592338747535,
"grad_norm": 2.5167956352233887,
"learning_rate": 1.4407177507693915e-05,
"loss": 7.8653,
"step": 7500
},
{
"epoch": 0.7135480236597502,
"grad_norm": 5.4338765144348145,
"learning_rate": 1.430285326795681e-05,
"loss": 7.8387,
"step": 7600
},
{
"epoch": 0.7229368134447469,
"grad_norm": 3.295238733291626,
"learning_rate": 1.4198529028219708e-05,
"loss": 7.8513,
"step": 7700
},
{
"epoch": 0.7323256032297437,
"grad_norm": 3.02062726020813,
"learning_rate": 1.4094204788482607e-05,
"loss": 7.8496,
"step": 7800
},
{
"epoch": 0.7417143930147404,
"grad_norm": 2.2482168674468994,
"learning_rate": 1.3989880548745502e-05,
"loss": 7.8276,
"step": 7900
},
{
"epoch": 0.7511031827997371,
"grad_norm": 3.394895315170288,
"learning_rate": 1.38855563090084e-05,
"loss": 7.8353,
"step": 8000
},
{
"epoch": 0.7604919725847338,
"grad_norm": 5.990811824798584,
"learning_rate": 1.3781232069271295e-05,
"loss": 7.8103,
"step": 8100
},
{
"epoch": 0.7698807623697306,
"grad_norm": 6.237292289733887,
"learning_rate": 1.3676907829534194e-05,
"loss": 7.8622,
"step": 8200
},
{
"epoch": 0.7792695521547273,
"grad_norm": 3.848353147506714,
"learning_rate": 1.3572583589797091e-05,
"loss": 7.832,
"step": 8300
},
{
"epoch": 0.788658341939724,
"grad_norm": 3.6994402408599854,
"learning_rate": 1.3468259350059986e-05,
"loss": 7.8349,
"step": 8400
},
{
"epoch": 0.7980471317247206,
"grad_norm": 2.945556163787842,
"learning_rate": 1.3363935110322885e-05,
"loss": 7.855,
"step": 8500
},
{
"epoch": 0.8074359215097174,
"grad_norm": 3.0218355655670166,
"learning_rate": 1.3259610870585782e-05,
"loss": 7.8316,
"step": 8600
},
{
"epoch": 0.8168247112947141,
"grad_norm": 3.8162949085235596,
"learning_rate": 1.3155286630848678e-05,
"loss": 7.8066,
"step": 8700
},
{
"epoch": 0.8262135010797108,
"grad_norm": 4.187964916229248,
"learning_rate": 1.3050962391111577e-05,
"loss": 7.8166,
"step": 8800
},
{
"epoch": 0.8356022908647075,
"grad_norm": 3.577139139175415,
"learning_rate": 1.2946638151374474e-05,
"loss": 7.8588,
"step": 8900
},
{
"epoch": 0.8449910806497043,
"grad_norm": 2.923499345779419,
"learning_rate": 1.284231391163737e-05,
"loss": 7.8042,
"step": 9000
},
{
"epoch": 0.854379870434701,
"grad_norm": 6.15308952331543,
"learning_rate": 1.2737989671900267e-05,
"loss": 7.8431,
"step": 9100
},
{
"epoch": 0.8637686602196977,
"grad_norm": 5.1844482421875,
"learning_rate": 1.2633665432163165e-05,
"loss": 7.7947,
"step": 9200
},
{
"epoch": 0.8731574500046944,
"grad_norm": 3.3482980728149414,
"learning_rate": 1.2530384434823432e-05,
"loss": 7.8175,
"step": 9300
},
{
"epoch": 0.8825462397896912,
"grad_norm": 3.3958215713500977,
"learning_rate": 1.242606019508633e-05,
"loss": 7.8299,
"step": 9400
},
{
"epoch": 0.8919350295746878,
"grad_norm": 37.65146255493164,
"learning_rate": 1.2321735955349225e-05,
"loss": 7.8455,
"step": 9500
}
],
"logging_steps": 100,
"max_steps": 21302,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 128,
"trial_name": null,
"trial_params": null
}