File size: 1,107 Bytes
b8eba7c
 
838be76
 
 
b8eba7c
838be76
 
 
 
 
 
 
 
 
 
b8eba7c
838be76
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
---
tags:
- llama
- alpaca
- qlora
- unsloth
- instruction-tuning
- fine-tuned
base_model: unsloth/llama-3.1-8b-bnb-4bit
library_name: peft
license: apache-2.0
datasets:
- tatsu-lab/alpaca
language:
- en
pipeline_tag: text-generation
---
# unsloth/mistral-7b-v0.3-bnb-4bit Fine-tuned with QLoRA (Unsloth) on Alpaca

This model is a fine-tuned version of [`unsloth/mistral-7b-v0.3-bnb-4bit`](https://huggingface.co/unsloth/mistral-7b-v0.3-bnb-4bit) using **QLoRA** and [Unsloth](https://github.com/unslothai/unsloth) for efficient instruction-tuning.

## πŸ“– Training Details
- **Dataset**: [`yahma/alpaca-cleaned`](https://huggingface.co/datasets/yahma/alpaca-cleaned)
- **QLoRA**: 4 bit quantization (NF4) using `bitsandbytes`
- **LoRA Rank**: 16 (adjust based on your config)
- **LoRA Alpha**: 16
- **Batch Size**: 2 per device
- **Gradient Accumulation**: 4
- **Learning Rate**: 2e-4
- **Epochs**: 1
- **Trainer**: `trl.SFTTrainer`

## πŸ’‘ Notes
- Optimized for memory-efficient fine-tuning with Unsloth
- No evaluation was run during training β€” please evaluate separately

## πŸ“ License
Apache 2.0