ParamDev's picture
Update README.md
838be76 verified
---
tags:
- llama
- alpaca
- qlora
- unsloth
- instruction-tuning
- fine-tuned
base_model: unsloth/llama-3.1-8b-bnb-4bit
library_name: peft
license: apache-2.0
datasets:
- tatsu-lab/alpaca
language:
- en
pipeline_tag: text-generation
---
# unsloth/mistral-7b-v0.3-bnb-4bit Fine-tuned with QLoRA (Unsloth) on Alpaca
This model is a fine-tuned version of [`unsloth/mistral-7b-v0.3-bnb-4bit`](https://huggingface.co/unsloth/mistral-7b-v0.3-bnb-4bit) using **QLoRA** and [Unsloth](https://github.com/unslothai/unsloth) for efficient instruction-tuning.
## πŸ“– Training Details
- **Dataset**: [`yahma/alpaca-cleaned`](https://huggingface.co/datasets/yahma/alpaca-cleaned)
- **QLoRA**: 4 bit quantization (NF4) using `bitsandbytes`
- **LoRA Rank**: 16 (adjust based on your config)
- **LoRA Alpha**: 16
- **Batch Size**: 2 per device
- **Gradient Accumulation**: 4
- **Learning Rate**: 2e-4
- **Epochs**: 1
- **Trainer**: `trl.SFTTrainer`
## πŸ’‘ Notes
- Optimized for memory-efficient fine-tuning with Unsloth
- No evaluation was run during training β€” please evaluate separately
## πŸ“ License
Apache 2.0