youssefkhalil320's picture
Upload folder using huggingface_hub
ac73f49 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.4694394892498357,
"eval_steps": 5000,
"global_step": 5000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.009388789784996713,
"grad_norm": 87.77811431884766,
"learning_rate": 9.009854528390429e-07,
"loss": 16.2337,
"step": 100
},
{
"epoch": 0.018777579569993427,
"grad_norm": 85.86430358886719,
"learning_rate": 1.8395119662130456e-06,
"loss": 13.5901,
"step": 200
},
{
"epoch": 0.02816636935499014,
"grad_norm": 14.885255813598633,
"learning_rate": 2.7592679493195683e-06,
"loss": 9.8565,
"step": 300
},
{
"epoch": 0.03755515913998685,
"grad_norm": 6.9691972732543945,
"learning_rate": 3.6977944626935713e-06,
"loss": 8.3332,
"step": 400
},
{
"epoch": 0.04694394892498357,
"grad_norm": 5.612818241119385,
"learning_rate": 4.6363209760675744e-06,
"loss": 8.1261,
"step": 500
},
{
"epoch": 0.05633273870998028,
"grad_norm": 4.705409526824951,
"learning_rate": 5.574847489441577e-06,
"loss": 8.0697,
"step": 600
},
{
"epoch": 0.06572152849497699,
"grad_norm": 4.337332725524902,
"learning_rate": 6.51337400281558e-06,
"loss": 8.0298,
"step": 700
},
{
"epoch": 0.0751103182799737,
"grad_norm": 3.6314213275909424,
"learning_rate": 7.451900516189583e-06,
"loss": 8.033,
"step": 800
},
{
"epoch": 0.08449910806497042,
"grad_norm": 3.4845075607299805,
"learning_rate": 8.390427029563585e-06,
"loss": 7.9858,
"step": 900
},
{
"epoch": 0.09388789784996714,
"grad_norm": 5.188210487365723,
"learning_rate": 9.328953542937589e-06,
"loss": 8.012,
"step": 1000
},
{
"epoch": 0.10327668763496385,
"grad_norm": 3.0830442905426025,
"learning_rate": 1.0267480056311592e-05,
"loss": 7.9745,
"step": 1100
},
{
"epoch": 0.11266547741996057,
"grad_norm": 3.4729278087615967,
"learning_rate": 1.1206006569685594e-05,
"loss": 8.0091,
"step": 1200
},
{
"epoch": 0.12205426720495728,
"grad_norm": 2.329235076904297,
"learning_rate": 1.2144533083059597e-05,
"loss": 8.0221,
"step": 1300
},
{
"epoch": 0.13144305698995398,
"grad_norm": 2.7225279808044434,
"learning_rate": 1.3083059596433601e-05,
"loss": 7.9583,
"step": 1400
},
{
"epoch": 0.1408318467749507,
"grad_norm": 2.012805938720703,
"learning_rate": 1.4021586109807603e-05,
"loss": 8.0031,
"step": 1500
},
{
"epoch": 0.1502206365599474,
"grad_norm": 2.9397523403167725,
"learning_rate": 1.4960112623181606e-05,
"loss": 7.9985,
"step": 1600
},
{
"epoch": 0.15960942634494413,
"grad_norm": 2.356337308883667,
"learning_rate": 1.589863913655561e-05,
"loss": 7.9647,
"step": 1700
},
{
"epoch": 0.16899821612994084,
"grad_norm": 2.6846818923950195,
"learning_rate": 1.6837165649929613e-05,
"loss": 7.9857,
"step": 1800
},
{
"epoch": 0.17838700591493756,
"grad_norm": 2.0188565254211426,
"learning_rate": 1.7775692163303613e-05,
"loss": 7.9806,
"step": 1900
},
{
"epoch": 0.18777579569993427,
"grad_norm": 4.030488014221191,
"learning_rate": 1.8714218676677617e-05,
"loss": 7.9761,
"step": 2000
},
{
"epoch": 0.197164585484931,
"grad_norm": 4.183101654052734,
"learning_rate": 1.965274519005162e-05,
"loss": 7.9696,
"step": 2100
},
{
"epoch": 0.2065533752699277,
"grad_norm": 1.4769889116287231,
"learning_rate": 1.9934275728965626e-05,
"loss": 8.0014,
"step": 2200
},
{
"epoch": 0.21594216505492442,
"grad_norm": 2.1914358139038086,
"learning_rate": 1.9829951489228525e-05,
"loss": 7.9546,
"step": 2300
},
{
"epoch": 0.22533095483992113,
"grad_norm": 22.55516815185547,
"learning_rate": 1.972562724949142e-05,
"loss": 7.9874,
"step": 2400
},
{
"epoch": 0.23471974462491785,
"grad_norm": 1.635116457939148,
"learning_rate": 1.962130300975432e-05,
"loss": 7.9846,
"step": 2500
},
{
"epoch": 0.24410853440991456,
"grad_norm": 5.707275390625,
"learning_rate": 1.9516978770017215e-05,
"loss": 7.9664,
"step": 2600
},
{
"epoch": 0.2534973241949113,
"grad_norm": 4.194604396820068,
"learning_rate": 1.9412654530280113e-05,
"loss": 7.9725,
"step": 2700
},
{
"epoch": 0.26288611397990796,
"grad_norm": 2.0074055194854736,
"learning_rate": 1.930833029054301e-05,
"loss": 7.9419,
"step": 2800
},
{
"epoch": 0.2722749037649047,
"grad_norm": 26.4300479888916,
"learning_rate": 1.9204006050805904e-05,
"loss": 7.9786,
"step": 2900
},
{
"epoch": 0.2816636935499014,
"grad_norm": 2.5870931148529053,
"learning_rate": 1.9099681811068803e-05,
"loss": 7.9479,
"step": 3000
},
{
"epoch": 0.29105248333489814,
"grad_norm": 1.6209933757781982,
"learning_rate": 1.8995357571331702e-05,
"loss": 7.9526,
"step": 3100
},
{
"epoch": 0.3004412731198948,
"grad_norm": 19.398080825805664,
"learning_rate": 1.8891033331594598e-05,
"loss": 7.9613,
"step": 3200
},
{
"epoch": 0.30983006290489157,
"grad_norm": 2.124729871749878,
"learning_rate": 1.8786709091857496e-05,
"loss": 7.9994,
"step": 3300
},
{
"epoch": 0.31921885268988826,
"grad_norm": 2.701019763946533,
"learning_rate": 1.8682384852120392e-05,
"loss": 7.9464,
"step": 3400
},
{
"epoch": 0.328607642474885,
"grad_norm": 1.944600224494934,
"learning_rate": 1.8578060612383287e-05,
"loss": 7.9429,
"step": 3500
},
{
"epoch": 0.3379964322598817,
"grad_norm": 41.85493087768555,
"learning_rate": 1.8473736372646186e-05,
"loss": 7.9539,
"step": 3600
},
{
"epoch": 0.34738522204487843,
"grad_norm": 2.024019479751587,
"learning_rate": 1.8369412132909085e-05,
"loss": 7.9699,
"step": 3700
},
{
"epoch": 0.3567740118298751,
"grad_norm": 2.2833781242370605,
"learning_rate": 1.826508789317198e-05,
"loss": 7.9144,
"step": 3800
},
{
"epoch": 0.36616280161487186,
"grad_norm": 3.1666574478149414,
"learning_rate": 1.8160763653434876e-05,
"loss": 7.9424,
"step": 3900
},
{
"epoch": 0.37555159139986855,
"grad_norm": 1.7259443998336792,
"learning_rate": 1.8056439413697775e-05,
"loss": 7.9361,
"step": 4000
},
{
"epoch": 0.3849403811848653,
"grad_norm": 2.5588905811309814,
"learning_rate": 1.795211517396067e-05,
"loss": 7.9144,
"step": 4100
},
{
"epoch": 0.394329170969862,
"grad_norm": 2.542963743209839,
"learning_rate": 1.784779093422357e-05,
"loss": 7.907,
"step": 4200
},
{
"epoch": 0.4037179607548587,
"grad_norm": 2.755725622177124,
"learning_rate": 1.7743466694486468e-05,
"loss": 7.9049,
"step": 4300
},
{
"epoch": 0.4131067505398554,
"grad_norm": 2.6067683696746826,
"learning_rate": 1.7639142454749364e-05,
"loss": 7.939,
"step": 4400
},
{
"epoch": 0.42249554032485215,
"grad_norm": 2.614475965499878,
"learning_rate": 1.753481821501226e-05,
"loss": 7.9067,
"step": 4500
},
{
"epoch": 0.43188433010984884,
"grad_norm": 2.172943353652954,
"learning_rate": 1.7430493975275155e-05,
"loss": 7.9149,
"step": 4600
},
{
"epoch": 0.4412731198948456,
"grad_norm": 6.292716979980469,
"learning_rate": 1.7326169735538053e-05,
"loss": 7.9705,
"step": 4700
},
{
"epoch": 0.45066190967984227,
"grad_norm": 2.595337152481079,
"learning_rate": 1.7221845495800952e-05,
"loss": 7.8992,
"step": 4800
},
{
"epoch": 0.460050699464839,
"grad_norm": 5.529814720153809,
"learning_rate": 1.7117521256063848e-05,
"loss": 7.9077,
"step": 4900
},
{
"epoch": 0.4694394892498357,
"grad_norm": 6.696155548095703,
"learning_rate": 1.7013197016326747e-05,
"loss": 7.8992,
"step": 5000
},
{
"epoch": 0.4694394892498357,
"eval_loss": 7.916718006134033,
"eval_runtime": 334.3618,
"eval_samples_per_second": 1019.336,
"eval_steps_per_second": 7.964,
"step": 5000
}
],
"logging_steps": 100,
"max_steps": 21302,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 128,
"trial_name": null,
"trial_params": null
}