File size: 928 Bytes
fdbf4ad 1d4abd6 26c956a 1d4abd6 fdbf4ad |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 |
---
language:
- en
license: cc-by-nc-4.0
library_name: transformers
tags:
- llama
- math
- reasoning
- fine-tuned
- fine-tuning
pipeline_tag: text-generation
model-index:
- name: Llama-3.1-8B-math-reasoning
results:
- task:
type: text-generation
name: Text Generation
dataset:
name: tulu3_mixture_math_reasoning
type: custom
metrics:
- name: Training Loss
type: loss
value: 0.98
base_model: meta-llama/Llama-3.1-8B
---
# Llama-3.1-8B Math Reasoning Model
Llama-3.1-8B SFT checkpoints for mathematical reasoning—artifacts of https://arxiv.org/abs/2509.11167.
## Model Details
- Base model: Llama-3.1-8B
- Training dataset: tulu3_mixture_math_reasoning
- Learning rate: 5e-06
- Effective batch size: 128
## Export Files
This repository includes export files for state averaging and other advanced techniques.
|