Jiazheng Li
commited on
Commit
·
7d35748
1
Parent(s):
9417bd0
init push
Browse files- README.md +69 -0
- adapter_config.json +35 -0
- adapter_model.safetensors +3 -0
- all_results.json +11 -0
- eval_results.json +7 -0
- special_tokens_map.json +24 -0
- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer_config.json +45 -0
- train_results.json +7 -0
- trainer_log.jsonl +38 -0
- trainer_state.json +295 -0
- training_args.bin +3 -0
- training_eval_loss.png +0 -0
- training_loss.png +0 -0
README.md
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: other
|
| 3 |
+
library_name: peft
|
| 4 |
+
tags:
|
| 5 |
+
- llama-factory
|
| 6 |
+
- lora
|
| 7 |
+
- generated_from_trainer
|
| 8 |
+
base_model: mistralai/Mixtral-8x7B-Instruct-v0.1
|
| 9 |
+
model-index:
|
| 10 |
+
- name: sft_trained_woaqa_mixtral
|
| 11 |
+
results: []
|
| 12 |
+
---
|
| 13 |
+
|
| 14 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 15 |
+
should probably proofread and complete it, then remove this comment. -->
|
| 16 |
+
|
| 17 |
+
# sft_trained_woaqa_mixtral
|
| 18 |
+
|
| 19 |
+
This model is a fine-tuned version of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) on the sft_wo_aqa_mistral dataset.
|
| 20 |
+
It achieves the following results on the evaluation set:
|
| 21 |
+
- Loss: 0.8062
|
| 22 |
+
|
| 23 |
+
## Model description
|
| 24 |
+
|
| 25 |
+
More information needed
|
| 26 |
+
|
| 27 |
+
## Intended uses & limitations
|
| 28 |
+
|
| 29 |
+
More information needed
|
| 30 |
+
|
| 31 |
+
## Training and evaluation data
|
| 32 |
+
|
| 33 |
+
More information needed
|
| 34 |
+
|
| 35 |
+
## Training procedure
|
| 36 |
+
|
| 37 |
+
### Training hyperparameters
|
| 38 |
+
|
| 39 |
+
The following hyperparameters were used during training:
|
| 40 |
+
- learning_rate: 5e-05
|
| 41 |
+
- train_batch_size: 8
|
| 42 |
+
- eval_batch_size: 8
|
| 43 |
+
- seed: 42
|
| 44 |
+
- gradient_accumulation_steps: 8
|
| 45 |
+
- total_train_batch_size: 64
|
| 46 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 47 |
+
- lr_scheduler_type: cosine
|
| 48 |
+
- num_epochs: 4.0
|
| 49 |
+
- mixed_precision_training: Native AMP
|
| 50 |
+
|
| 51 |
+
### Training results
|
| 52 |
+
|
| 53 |
+
| Training Loss | Epoch | Step | Validation Loss |
|
| 54 |
+
|:-------------:|:-----:|:----:|:---------------:|
|
| 55 |
+
| 0.8668 | 0.63 | 100 | 0.8571 |
|
| 56 |
+
| 0.7837 | 1.26 | 200 | 0.8230 |
|
| 57 |
+
| 0.7824 | 1.9 | 300 | 0.8058 |
|
| 58 |
+
| 0.7401 | 2.53 | 400 | 0.8059 |
|
| 59 |
+
| 0.7101 | 3.16 | 500 | 0.8072 |
|
| 60 |
+
| 0.7037 | 3.79 | 600 | 0.8062 |
|
| 61 |
+
|
| 62 |
+
|
| 63 |
+
### Framework versions
|
| 64 |
+
|
| 65 |
+
- PEFT 0.10.0
|
| 66 |
+
- Transformers 4.38.2
|
| 67 |
+
- Pytorch 2.2.1+cu121
|
| 68 |
+
- Datasets 2.18.0
|
| 69 |
+
- Tokenizers 0.15.2
|
adapter_config.json
ADDED
|
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alpha_pattern": {},
|
| 3 |
+
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "mistralai/Mixtral-8x7B-Instruct-v0.1",
|
| 5 |
+
"bias": "none",
|
| 6 |
+
"fan_in_fan_out": false,
|
| 7 |
+
"inference_mode": true,
|
| 8 |
+
"init_lora_weights": true,
|
| 9 |
+
"layer_replication": null,
|
| 10 |
+
"layers_pattern": null,
|
| 11 |
+
"layers_to_transform": null,
|
| 12 |
+
"loftq_config": {},
|
| 13 |
+
"lora_alpha": 16,
|
| 14 |
+
"lora_dropout": 0.0,
|
| 15 |
+
"megatron_config": null,
|
| 16 |
+
"megatron_core": "megatron.core",
|
| 17 |
+
"modules_to_save": null,
|
| 18 |
+
"peft_type": "LORA",
|
| 19 |
+
"r": 8,
|
| 20 |
+
"rank_pattern": {},
|
| 21 |
+
"revision": null,
|
| 22 |
+
"target_modules": [
|
| 23 |
+
"w3",
|
| 24 |
+
"w1",
|
| 25 |
+
"w2",
|
| 26 |
+
"v_proj",
|
| 27 |
+
"o_proj",
|
| 28 |
+
"q_proj",
|
| 29 |
+
"gate",
|
| 30 |
+
"k_proj"
|
| 31 |
+
],
|
| 32 |
+
"task_type": "CAUSAL_LM",
|
| 33 |
+
"use_dora": false,
|
| 34 |
+
"use_rslora": false
|
| 35 |
+
}
|
adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:95394e419230e4e32bc7a3243d358ebc852aaf01d6f99d31e9fe116b41372fe9
|
| 3 |
+
size 484722304
|
all_results.json
ADDED
|
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 3.99,
|
| 3 |
+
"eval_loss": 0.8061766028404236,
|
| 4 |
+
"eval_runtime": 867.2475,
|
| 5 |
+
"eval_samples_per_second": 3.195,
|
| 6 |
+
"eval_steps_per_second": 0.4,
|
| 7 |
+
"train_loss": 0.7920433780815028,
|
| 8 |
+
"train_runtime": 47185.3474,
|
| 9 |
+
"train_samples_per_second": 0.858,
|
| 10 |
+
"train_steps_per_second": 0.013
|
| 11 |
+
}
|
eval_results.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 3.99,
|
| 3 |
+
"eval_loss": 0.8061766028404236,
|
| 4 |
+
"eval_runtime": 867.2475,
|
| 5 |
+
"eval_samples_per_second": 3.195,
|
| 6 |
+
"eval_steps_per_second": 0.4
|
| 7 |
+
}
|
special_tokens_map.json
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "</s>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"pad_token": "</s>",
|
| 17 |
+
"unk_token": {
|
| 18 |
+
"content": "<unk>",
|
| 19 |
+
"lstrip": false,
|
| 20 |
+
"normalized": false,
|
| 21 |
+
"rstrip": false,
|
| 22 |
+
"single_word": false
|
| 23 |
+
}
|
| 24 |
+
}
|
tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
| 3 |
+
size 493443
|
tokenizer_config.json
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_bos_token": true,
|
| 3 |
+
"add_eos_token": false,
|
| 4 |
+
"added_tokens_decoder": {
|
| 5 |
+
"0": {
|
| 6 |
+
"content": "<unk>",
|
| 7 |
+
"lstrip": false,
|
| 8 |
+
"normalized": false,
|
| 9 |
+
"rstrip": false,
|
| 10 |
+
"single_word": false,
|
| 11 |
+
"special": true
|
| 12 |
+
},
|
| 13 |
+
"1": {
|
| 14 |
+
"content": "<s>",
|
| 15 |
+
"lstrip": false,
|
| 16 |
+
"normalized": false,
|
| 17 |
+
"rstrip": false,
|
| 18 |
+
"single_word": false,
|
| 19 |
+
"special": true
|
| 20 |
+
},
|
| 21 |
+
"2": {
|
| 22 |
+
"content": "</s>",
|
| 23 |
+
"lstrip": false,
|
| 24 |
+
"normalized": false,
|
| 25 |
+
"rstrip": false,
|
| 26 |
+
"single_word": false,
|
| 27 |
+
"special": true
|
| 28 |
+
}
|
| 29 |
+
},
|
| 30 |
+
"additional_special_tokens": [],
|
| 31 |
+
"bos_token": "<s>",
|
| 32 |
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{{ '<s>' + system_message }}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ ' [INST] ' + content + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ content + '</s>' }}{% endif %}{% endfor %}",
|
| 33 |
+
"clean_up_tokenization_spaces": false,
|
| 34 |
+
"eos_token": "</s>",
|
| 35 |
+
"legacy": true,
|
| 36 |
+
"model_max_length": 1000000000000000019884624838656,
|
| 37 |
+
"pad_token": "</s>",
|
| 38 |
+
"padding_side": "right",
|
| 39 |
+
"sp_model_kwargs": {},
|
| 40 |
+
"spaces_between_special_tokens": false,
|
| 41 |
+
"split_special_tokens": false,
|
| 42 |
+
"tokenizer_class": "LlamaTokenizer",
|
| 43 |
+
"unk_token": "<unk>",
|
| 44 |
+
"use_default_system_prompt": false
|
| 45 |
+
}
|
train_results.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"epoch": 3.99,
|
| 3 |
+
"train_loss": 0.7920433780815028,
|
| 4 |
+
"train_runtime": 47185.3474,
|
| 5 |
+
"train_samples_per_second": 0.858,
|
| 6 |
+
"train_steps_per_second": 0.013
|
| 7 |
+
}
|
trainer_log.jsonl
ADDED
|
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"current_steps": 20, "total_steps": 632, "loss": 1.3016, "learning_rate": 4.9876553763060684e-05, "epoch": 0.13, "percentage": 3.16, "elapsed_time": "0:22:10", "remaining_time": "11:18:44"}
|
| 2 |
+
{"current_steps": 40, "total_steps": 632, "loss": 0.9953, "learning_rate": 4.950743417011591e-05, "epoch": 0.25, "percentage": 6.33, "elapsed_time": "0:44:06", "remaining_time": "10:52:50"}
|
| 3 |
+
{"current_steps": 60, "total_steps": 632, "loss": 0.925, "learning_rate": 4.889628653514402e-05, "epoch": 0.38, "percentage": 9.49, "elapsed_time": "1:09:28", "remaining_time": "11:02:22"}
|
| 4 |
+
{"current_steps": 80, "total_steps": 632, "loss": 0.8788, "learning_rate": 4.804914636820517e-05, "epoch": 0.51, "percentage": 12.66, "elapsed_time": "1:31:29", "remaining_time": "10:31:14"}
|
| 5 |
+
{"current_steps": 100, "total_steps": 632, "loss": 0.8668, "learning_rate": 4.6974379770560846e-05, "epoch": 0.63, "percentage": 15.82, "elapsed_time": "1:53:23", "remaining_time": "10:03:12"}
|
| 6 |
+
{"current_steps": 100, "total_steps": 632, "eval_loss": 0.8571113348007202, "epoch": 0.63, "percentage": 15.82, "elapsed_time": "2:07:51", "remaining_time": "11:20:11"}
|
| 7 |
+
{"current_steps": 120, "total_steps": 632, "loss": 0.8488, "learning_rate": 4.5682600813576435e-05, "epoch": 0.76, "percentage": 18.99, "elapsed_time": "2:29:50", "remaining_time": "10:39:21"}
|
| 8 |
+
{"current_steps": 140, "total_steps": 632, "loss": 0.8462, "learning_rate": 4.41865667173477e-05, "epoch": 0.88, "percentage": 22.15, "elapsed_time": "2:51:49", "remaining_time": "10:03:51"}
|
| 9 |
+
{"current_steps": 160, "total_steps": 632, "loss": 0.827, "learning_rate": 4.2501051864235636e-05, "epoch": 1.01, "percentage": 25.32, "elapsed_time": "3:14:53", "remaining_time": "9:34:54"}
|
| 10 |
+
{"current_steps": 180, "total_steps": 632, "loss": 0.7885, "learning_rate": 4.0642701891514e-05, "epoch": 1.14, "percentage": 28.48, "elapsed_time": "3:37:46", "remaining_time": "9:06:50"}
|
| 11 |
+
{"current_steps": 200, "total_steps": 632, "loss": 0.7837, "learning_rate": 3.862986930406669e-05, "epoch": 1.26, "percentage": 31.65, "elapsed_time": "3:59:48", "remaining_time": "8:37:58"}
|
| 12 |
+
{"current_steps": 200, "total_steps": 632, "eval_loss": 0.8230095505714417, "epoch": 1.26, "percentage": 31.65, "elapsed_time": "4:14:17", "remaining_time": "9:09:16"}
|
| 13 |
+
{"current_steps": 220, "total_steps": 632, "loss": 0.7938, "learning_rate": 3.6482432230574446e-05, "epoch": 1.39, "percentage": 34.81, "elapsed_time": "4:36:16", "remaining_time": "8:37:23"}
|
| 14 |
+
{"current_steps": 240, "total_steps": 632, "loss": 0.7819, "learning_rate": 3.4221598113100195e-05, "epoch": 1.52, "percentage": 37.97, "elapsed_time": "4:58:13", "remaining_time": "8:07:06"}
|
| 15 |
+
{"current_steps": 260, "total_steps": 632, "loss": 0.7877, "learning_rate": 3.186969426877563e-05, "epoch": 1.64, "percentage": 41.14, "elapsed_time": "5:20:11", "remaining_time": "7:38:07"}
|
| 16 |
+
{"current_steps": 280, "total_steps": 632, "loss": 0.7918, "learning_rate": 2.9449947391938766e-05, "epoch": 1.77, "percentage": 44.3, "elapsed_time": "5:42:00", "remaining_time": "7:09:57"}
|
| 17 |
+
{"current_steps": 300, "total_steps": 632, "loss": 0.7824, "learning_rate": 2.6986254174292862e-05, "epoch": 1.9, "percentage": 47.47, "elapsed_time": "6:04:00", "remaining_time": "6:42:49"}
|
| 18 |
+
{"current_steps": 300, "total_steps": 632, "eval_loss": 0.8058096766471863, "epoch": 1.9, "percentage": 47.47, "elapsed_time": "6:18:27", "remaining_time": "6:58:50"}
|
| 19 |
+
{"current_steps": 320, "total_steps": 632, "loss": 0.7741, "learning_rate": 2.4502945308373246e-05, "epoch": 2.02, "percentage": 50.63, "elapsed_time": "6:40:24", "remaining_time": "6:30:24"}
|
| 20 |
+
{"current_steps": 340, "total_steps": 632, "loss": 0.7369, "learning_rate": 2.2024545204952383e-05, "epoch": 2.15, "percentage": 53.8, "elapsed_time": "7:02:25", "remaining_time": "6:02:47"}
|
| 21 |
+
{"current_steps": 360, "total_steps": 632, "loss": 0.7365, "learning_rate": 1.957552979734205e-05, "epoch": 2.27, "percentage": 56.96, "elapsed_time": "7:24:19", "remaining_time": "5:35:42"}
|
| 22 |
+
{"current_steps": 380, "total_steps": 632, "loss": 0.7463, "learning_rate": 1.7180084824444325e-05, "epoch": 2.4, "percentage": 60.13, "elapsed_time": "7:46:14", "remaining_time": "5:09:11"}
|
| 23 |
+
{"current_steps": 400, "total_steps": 632, "loss": 0.7401, "learning_rate": 1.4861866979675154e-05, "epoch": 2.53, "percentage": 63.29, "elapsed_time": "8:08:15", "remaining_time": "4:43:11"}
|
| 24 |
+
{"current_steps": 400, "total_steps": 632, "eval_loss": 0.8059037923812866, "epoch": 2.53, "percentage": 63.29, "elapsed_time": "8:22:49", "remaining_time": "4:51:38"}
|
| 25 |
+
{"current_steps": 420, "total_steps": 632, "loss": 0.7332, "learning_rate": 1.2643770284581929e-05, "epoch": 2.65, "percentage": 66.46, "elapsed_time": "8:44:49", "remaining_time": "4:24:54"}
|
| 26 |
+
{"current_steps": 440, "total_steps": 632, "loss": 0.7364, "learning_rate": 1.0547699994378787e-05, "epoch": 2.78, "percentage": 69.62, "elapsed_time": "9:06:48", "remaining_time": "3:58:36"}
|
| 27 |
+
{"current_steps": 460, "total_steps": 632, "loss": 0.7318, "learning_rate": 8.594356268240616e-06, "epoch": 2.91, "percentage": 72.78, "elapsed_time": "9:28:46", "remaining_time": "3:32:40"}
|
| 28 |
+
{"current_steps": 480, "total_steps": 632, "loss": 0.7222, "learning_rate": 6.803029740762648e-06, "epoch": 3.03, "percentage": 75.95, "elapsed_time": "9:50:39", "remaining_time": "3:07:02"}
|
| 29 |
+
{"current_steps": 500, "total_steps": 632, "loss": 0.7101, "learning_rate": 5.191411013460645e-06, "epoch": 3.16, "percentage": 79.11, "elapsed_time": "10:12:38", "remaining_time": "2:41:44"}
|
| 30 |
+
{"current_steps": 500, "total_steps": 632, "eval_loss": 0.807178795337677, "epoch": 3.16, "percentage": 79.11, "elapsed_time": "10:27:05", "remaining_time": "2:45:33"}
|
| 31 |
+
{"current_steps": 520, "total_steps": 632, "loss": 0.7135, "learning_rate": 3.775415947715899e-06, "epoch": 3.29, "percentage": 82.28, "elapsed_time": "10:49:03", "remaining_time": "2:19:47"}
|
| 32 |
+
{"current_steps": 540, "total_steps": 632, "loss": 0.7045, "learning_rate": 2.5690284845196923e-06, "epoch": 3.41, "percentage": 85.44, "elapsed_time": "11:10:58", "remaining_time": "1:54:18"}
|
| 33 |
+
{"current_steps": 560, "total_steps": 632, "loss": 0.7057, "learning_rate": 1.5841625432818057e-06, "epoch": 3.54, "percentage": 88.61, "elapsed_time": "11:32:56", "remaining_time": "1:29:05"}
|
| 34 |
+
{"current_steps": 580, "total_steps": 632, "loss": 0.7044, "learning_rate": 8.305443635490711e-07, "epoch": 3.67, "percentage": 91.77, "elapsed_time": "11:54:45", "remaining_time": "1:04:04"}
|
| 35 |
+
{"current_steps": 600, "total_steps": 632, "loss": 0.7037, "learning_rate": 3.1561645159166597e-07, "epoch": 3.79, "percentage": 94.94, "elapsed_time": "12:16:39", "remaining_time": "0:39:17"}
|
| 36 |
+
{"current_steps": 600, "total_steps": 632, "eval_loss": 0.8061766028404236, "epoch": 3.79, "percentage": 94.94, "elapsed_time": "12:31:07", "remaining_time": "0:40:03"}
|
| 37 |
+
{"current_steps": 620, "total_steps": 632, "loss": 0.7027, "learning_rate": 4.4464080451675494e-08, "epoch": 3.92, "percentage": 98.1, "elapsed_time": "12:53:10", "remaining_time": "0:14:57"}
|
| 38 |
+
{"current_steps": 632, "total_steps": 632, "epoch": 3.99, "percentage": 100.0, "elapsed_time": "13:06:25", "remaining_time": "0:00:00"}
|
trainer_state.json
ADDED
|
@@ -0,0 +1,295 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 3.993680884676145,
|
| 5 |
+
"eval_steps": 100,
|
| 6 |
+
"global_step": 632,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.13,
|
| 13 |
+
"grad_norm": 0.37133172154426575,
|
| 14 |
+
"learning_rate": 4.9876553763060684e-05,
|
| 15 |
+
"loss": 1.3016,
|
| 16 |
+
"step": 20
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.25,
|
| 20 |
+
"grad_norm": 0.3215538263320923,
|
| 21 |
+
"learning_rate": 4.950743417011591e-05,
|
| 22 |
+
"loss": 0.9953,
|
| 23 |
+
"step": 40
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.38,
|
| 27 |
+
"grad_norm": 0.33872345089912415,
|
| 28 |
+
"learning_rate": 4.889628653514402e-05,
|
| 29 |
+
"loss": 0.925,
|
| 30 |
+
"step": 60
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.51,
|
| 34 |
+
"grad_norm": 0.3175918459892273,
|
| 35 |
+
"learning_rate": 4.804914636820517e-05,
|
| 36 |
+
"loss": 0.8788,
|
| 37 |
+
"step": 80
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 0.63,
|
| 41 |
+
"grad_norm": 0.3202904164791107,
|
| 42 |
+
"learning_rate": 4.6974379770560846e-05,
|
| 43 |
+
"loss": 0.8668,
|
| 44 |
+
"step": 100
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 0.63,
|
| 48 |
+
"eval_loss": 0.8571113348007202,
|
| 49 |
+
"eval_runtime": 868.2347,
|
| 50 |
+
"eval_samples_per_second": 3.192,
|
| 51 |
+
"eval_steps_per_second": 0.4,
|
| 52 |
+
"step": 100
|
| 53 |
+
},
|
| 54 |
+
{
|
| 55 |
+
"epoch": 0.76,
|
| 56 |
+
"grad_norm": 0.33093100786209106,
|
| 57 |
+
"learning_rate": 4.5682600813576435e-05,
|
| 58 |
+
"loss": 0.8488,
|
| 59 |
+
"step": 120
|
| 60 |
+
},
|
| 61 |
+
{
|
| 62 |
+
"epoch": 0.88,
|
| 63 |
+
"grad_norm": 0.3297623097896576,
|
| 64 |
+
"learning_rate": 4.41865667173477e-05,
|
| 65 |
+
"loss": 0.8462,
|
| 66 |
+
"step": 140
|
| 67 |
+
},
|
| 68 |
+
{
|
| 69 |
+
"epoch": 1.01,
|
| 70 |
+
"grad_norm": 0.3524036705493927,
|
| 71 |
+
"learning_rate": 4.2501051864235636e-05,
|
| 72 |
+
"loss": 0.827,
|
| 73 |
+
"step": 160
|
| 74 |
+
},
|
| 75 |
+
{
|
| 76 |
+
"epoch": 1.14,
|
| 77 |
+
"grad_norm": 0.3592537045478821,
|
| 78 |
+
"learning_rate": 4.0642701891514e-05,
|
| 79 |
+
"loss": 0.7885,
|
| 80 |
+
"step": 180
|
| 81 |
+
},
|
| 82 |
+
{
|
| 83 |
+
"epoch": 1.26,
|
| 84 |
+
"grad_norm": 0.3888987600803375,
|
| 85 |
+
"learning_rate": 3.862986930406669e-05,
|
| 86 |
+
"loss": 0.7837,
|
| 87 |
+
"step": 200
|
| 88 |
+
},
|
| 89 |
+
{
|
| 90 |
+
"epoch": 1.26,
|
| 91 |
+
"eval_loss": 0.8230095505714417,
|
| 92 |
+
"eval_runtime": 869.2538,
|
| 93 |
+
"eval_samples_per_second": 3.188,
|
| 94 |
+
"eval_steps_per_second": 0.399,
|
| 95 |
+
"step": 200
|
| 96 |
+
},
|
| 97 |
+
{
|
| 98 |
+
"epoch": 1.39,
|
| 99 |
+
"grad_norm": 0.380818247795105,
|
| 100 |
+
"learning_rate": 3.6482432230574446e-05,
|
| 101 |
+
"loss": 0.7938,
|
| 102 |
+
"step": 220
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"epoch": 1.52,
|
| 106 |
+
"grad_norm": 0.3564074635505676,
|
| 107 |
+
"learning_rate": 3.4221598113100195e-05,
|
| 108 |
+
"loss": 0.7819,
|
| 109 |
+
"step": 240
|
| 110 |
+
},
|
| 111 |
+
{
|
| 112 |
+
"epoch": 1.64,
|
| 113 |
+
"grad_norm": 0.3780010938644409,
|
| 114 |
+
"learning_rate": 3.186969426877563e-05,
|
| 115 |
+
"loss": 0.7877,
|
| 116 |
+
"step": 260
|
| 117 |
+
},
|
| 118 |
+
{
|
| 119 |
+
"epoch": 1.77,
|
| 120 |
+
"grad_norm": 0.36975908279418945,
|
| 121 |
+
"learning_rate": 2.9449947391938766e-05,
|
| 122 |
+
"loss": 0.7918,
|
| 123 |
+
"step": 280
|
| 124 |
+
},
|
| 125 |
+
{
|
| 126 |
+
"epoch": 1.9,
|
| 127 |
+
"grad_norm": 0.39148128032684326,
|
| 128 |
+
"learning_rate": 2.6986254174292862e-05,
|
| 129 |
+
"loss": 0.7824,
|
| 130 |
+
"step": 300
|
| 131 |
+
},
|
| 132 |
+
{
|
| 133 |
+
"epoch": 1.9,
|
| 134 |
+
"eval_loss": 0.8058096766471863,
|
| 135 |
+
"eval_runtime": 867.7611,
|
| 136 |
+
"eval_samples_per_second": 3.193,
|
| 137 |
+
"eval_steps_per_second": 0.4,
|
| 138 |
+
"step": 300
|
| 139 |
+
},
|
| 140 |
+
{
|
| 141 |
+
"epoch": 2.02,
|
| 142 |
+
"grad_norm": 0.36495909094810486,
|
| 143 |
+
"learning_rate": 2.4502945308373246e-05,
|
| 144 |
+
"loss": 0.7741,
|
| 145 |
+
"step": 320
|
| 146 |
+
},
|
| 147 |
+
{
|
| 148 |
+
"epoch": 2.15,
|
| 149 |
+
"grad_norm": 0.40916556119918823,
|
| 150 |
+
"learning_rate": 2.2024545204952383e-05,
|
| 151 |
+
"loss": 0.7369,
|
| 152 |
+
"step": 340
|
| 153 |
+
},
|
| 154 |
+
{
|
| 155 |
+
"epoch": 2.27,
|
| 156 |
+
"grad_norm": 0.42436033487319946,
|
| 157 |
+
"learning_rate": 1.957552979734205e-05,
|
| 158 |
+
"loss": 0.7365,
|
| 159 |
+
"step": 360
|
| 160 |
+
},
|
| 161 |
+
{
|
| 162 |
+
"epoch": 2.4,
|
| 163 |
+
"grad_norm": 0.43983975052833557,
|
| 164 |
+
"learning_rate": 1.7180084824444325e-05,
|
| 165 |
+
"loss": 0.7463,
|
| 166 |
+
"step": 380
|
| 167 |
+
},
|
| 168 |
+
{
|
| 169 |
+
"epoch": 2.53,
|
| 170 |
+
"grad_norm": 0.44617146253585815,
|
| 171 |
+
"learning_rate": 1.4861866979675154e-05,
|
| 172 |
+
"loss": 0.7401,
|
| 173 |
+
"step": 400
|
| 174 |
+
},
|
| 175 |
+
{
|
| 176 |
+
"epoch": 2.53,
|
| 177 |
+
"eval_loss": 0.8059037923812866,
|
| 178 |
+
"eval_runtime": 873.5556,
|
| 179 |
+
"eval_samples_per_second": 3.172,
|
| 180 |
+
"eval_steps_per_second": 0.397,
|
| 181 |
+
"step": 400
|
| 182 |
+
},
|
| 183 |
+
{
|
| 184 |
+
"epoch": 2.65,
|
| 185 |
+
"grad_norm": 0.4369719624519348,
|
| 186 |
+
"learning_rate": 1.2643770284581929e-05,
|
| 187 |
+
"loss": 0.7332,
|
| 188 |
+
"step": 420
|
| 189 |
+
},
|
| 190 |
+
{
|
| 191 |
+
"epoch": 2.78,
|
| 192 |
+
"grad_norm": 0.4235495328903198,
|
| 193 |
+
"learning_rate": 1.0547699994378787e-05,
|
| 194 |
+
"loss": 0.7364,
|
| 195 |
+
"step": 440
|
| 196 |
+
},
|
| 197 |
+
{
|
| 198 |
+
"epoch": 2.91,
|
| 199 |
+
"grad_norm": 0.4584214389324188,
|
| 200 |
+
"learning_rate": 8.594356268240616e-06,
|
| 201 |
+
"loss": 0.7318,
|
| 202 |
+
"step": 460
|
| 203 |
+
},
|
| 204 |
+
{
|
| 205 |
+
"epoch": 3.03,
|
| 206 |
+
"grad_norm": 0.4329874813556671,
|
| 207 |
+
"learning_rate": 6.803029740762648e-06,
|
| 208 |
+
"loss": 0.7222,
|
| 209 |
+
"step": 480
|
| 210 |
+
},
|
| 211 |
+
{
|
| 212 |
+
"epoch": 3.16,
|
| 213 |
+
"grad_norm": 0.5470691919326782,
|
| 214 |
+
"learning_rate": 5.191411013460645e-06,
|
| 215 |
+
"loss": 0.7101,
|
| 216 |
+
"step": 500
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"epoch": 3.16,
|
| 220 |
+
"eval_loss": 0.807178795337677,
|
| 221 |
+
"eval_runtime": 867.2257,
|
| 222 |
+
"eval_samples_per_second": 3.195,
|
| 223 |
+
"eval_steps_per_second": 0.4,
|
| 224 |
+
"step": 500
|
| 225 |
+
},
|
| 226 |
+
{
|
| 227 |
+
"epoch": 3.29,
|
| 228 |
+
"grad_norm": 0.4557144343852997,
|
| 229 |
+
"learning_rate": 3.775415947715899e-06,
|
| 230 |
+
"loss": 0.7135,
|
| 231 |
+
"step": 520
|
| 232 |
+
},
|
| 233 |
+
{
|
| 234 |
+
"epoch": 3.41,
|
| 235 |
+
"grad_norm": 0.46527382731437683,
|
| 236 |
+
"learning_rate": 2.5690284845196923e-06,
|
| 237 |
+
"loss": 0.7045,
|
| 238 |
+
"step": 540
|
| 239 |
+
},
|
| 240 |
+
{
|
| 241 |
+
"epoch": 3.54,
|
| 242 |
+
"grad_norm": 0.4702458679676056,
|
| 243 |
+
"learning_rate": 1.5841625432818057e-06,
|
| 244 |
+
"loss": 0.7057,
|
| 245 |
+
"step": 560
|
| 246 |
+
},
|
| 247 |
+
{
|
| 248 |
+
"epoch": 3.67,
|
| 249 |
+
"grad_norm": 0.4915095865726471,
|
| 250 |
+
"learning_rate": 8.305443635490711e-07,
|
| 251 |
+
"loss": 0.7044,
|
| 252 |
+
"step": 580
|
| 253 |
+
},
|
| 254 |
+
{
|
| 255 |
+
"epoch": 3.79,
|
| 256 |
+
"grad_norm": 0.5043957233428955,
|
| 257 |
+
"learning_rate": 3.1561645159166597e-07,
|
| 258 |
+
"loss": 0.7037,
|
| 259 |
+
"step": 600
|
| 260 |
+
},
|
| 261 |
+
{
|
| 262 |
+
"epoch": 3.79,
|
| 263 |
+
"eval_loss": 0.8061766028404236,
|
| 264 |
+
"eval_runtime": 867.9324,
|
| 265 |
+
"eval_samples_per_second": 3.193,
|
| 266 |
+
"eval_steps_per_second": 0.4,
|
| 267 |
+
"step": 600
|
| 268 |
+
},
|
| 269 |
+
{
|
| 270 |
+
"epoch": 3.92,
|
| 271 |
+
"grad_norm": 0.4672119915485382,
|
| 272 |
+
"learning_rate": 4.4464080451675494e-08,
|
| 273 |
+
"loss": 0.7027,
|
| 274 |
+
"step": 620
|
| 275 |
+
},
|
| 276 |
+
{
|
| 277 |
+
"epoch": 3.99,
|
| 278 |
+
"step": 632,
|
| 279 |
+
"total_flos": 1.137224723506643e+19,
|
| 280 |
+
"train_loss": 0.7920433780815028,
|
| 281 |
+
"train_runtime": 47185.3474,
|
| 282 |
+
"train_samples_per_second": 0.858,
|
| 283 |
+
"train_steps_per_second": 0.013
|
| 284 |
+
}
|
| 285 |
+
],
|
| 286 |
+
"logging_steps": 20,
|
| 287 |
+
"max_steps": 632,
|
| 288 |
+
"num_input_tokens_seen": 0,
|
| 289 |
+
"num_train_epochs": 4,
|
| 290 |
+
"save_steps": 100,
|
| 291 |
+
"total_flos": 1.137224723506643e+19,
|
| 292 |
+
"train_batch_size": 8,
|
| 293 |
+
"trial_name": null,
|
| 294 |
+
"trial_params": null
|
| 295 |
+
}
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1b573d48fc5f9488e07fd9eb672eddaf1b96a03a8caa287045c162f31893e3fd
|
| 3 |
+
size 5112
|
training_eval_loss.png
ADDED
|
training_loss.png
ADDED
|