Spaces:
Paused
Paused
task: llm-sft | |
base_model: microsoft/Phi-3.5-mini-instruct | |
# base_model: TheBloke/WizardCoder-33B-V1.1-AWQ | |
project_name: llama | |
log: tensorboard | |
backend: local | |
data: | |
path: Platma/platma-ai | |
train_split: train | |
valid_split: null | |
chat_template: null | |
column_mapping: | |
text_column: text | |
params: | |
block_size: 1024 | |
lr: 1e-4 | |
warmup_ratio: 0.1 | |
weight_decay: 0.01 | |
epochs: 1 | |
batch_size: 2 | |
gradient_accumulation: 8 | |
mixed_precision: fp16 | |
peft: True | |
quantization: null | |
lora_r: 16 | |
lora_alpha: 32 | |
lora_dropout: 0.05 | |
unsloth: False | |
optimizer: paged_adamw_8bit | |
target_modules: all-linear | |
padding: right | |
optimizer: paged_adamw_8bit | |
scheduler: cosine | |
hub: | |
username: Platma | |
token: ${HF_ACCESS_TOKEN} | |
push_to_hub: True |