metadata
tags:
- llama
- alpaca
- qlora
- unsloth
- instruction-tuning
- fine-tuned
base_model: unsloth/llama-3.1-8b-bnb-4bit
library_name: peft
license: apache-2.0
datasets:
- tatsu-lab/alpaca
language:
- en
pipeline_tag: text-generation
unsloth/mistral-7b-v0.3-bnb-4bit Fine-tuned with QLoRA (Unsloth) on Alpaca
This model is a fine-tuned version of unsloth/mistral-7b-v0.3-bnb-4bit
using QLoRA and Unsloth for efficient instruction-tuning.
π Training Details
- Dataset:
yahma/alpaca-cleaned
- QLoRA: 4 bit quantization (NF4) using
bitsandbytes
- LoRA Rank: 16 (adjust based on your config)
- LoRA Alpha: 16
- Batch Size: 2 per device
- Gradient Accumulation: 4
- Learning Rate: 2e-4
- Epochs: 1
- Trainer:
trl.SFTTrainer
π‘ Notes
- Optimized for memory-efficient fine-tuning with Unsloth
- No evaluation was run during training β please evaluate separately
π License
Apache 2.0