#!/bin/bash set -x MODEL_PATH=meta-llama/Meta-Llama-3-8B-Instruct llamafactory-cli train \ --model_name_or_path ${MODEL_PATH} \ --trust_remote_code \ --stage sft \ --do_train \ --finetuning_type lora \ --lora_rank 8 \ --lora_target all \ --dataset identity,alpaca_en_demo \ --template llama3 \ --cutoff_len 2048 \ --max_samples 1000 \ --overwrite_cache \ --preprocessing_num_workers 16 \ --dataloader_num_workers 4 \ --output_dir saves/llama3-8b/lora/sft \ --logging_steps 10 \ --save_steps 500 \ --plot_loss \ --overwrite_output_dir \ --save_only_model false \ --report_to none \ --per_device_train_batch_size 1 \ --gradient_accumulation_steps 8 \ --learning_rate 1e-4 \ --num_train_epochs 3.0 \ --lr_scheduler_type cosine \ --warmup_ratio 0.1 \ --bf16 \ --ddp_timeout 180000000