|
|
--- |
|
|
language: |
|
|
- en |
|
|
license: cc-by-nc-4.0 |
|
|
library_name: transformers |
|
|
tags: |
|
|
- llama |
|
|
- math |
|
|
- reasoning |
|
|
- fine-tuned |
|
|
- fine-tuning |
|
|
pipeline_tag: text-generation |
|
|
model-index: |
|
|
- name: Llama-3.1-8B-math-reasoning |
|
|
results: |
|
|
- task: |
|
|
type: text-generation |
|
|
name: Text Generation |
|
|
dataset: |
|
|
name: tulu3_mixture_math_reasoning |
|
|
type: custom |
|
|
metrics: |
|
|
- name: Training Loss |
|
|
type: loss |
|
|
value: 0.98 |
|
|
base_model: meta-llama/Llama-3.1-8B |
|
|
--- |
|
|
|
|
|
# Llama-3.1-8B Math Reasoning Model |
|
|
|
|
|
Llama-3.1-8B SFT checkpoints for mathematical reasoning—artifacts of https://arxiv.org/abs/2509.11167. |
|
|
|
|
|
## Model Details |
|
|
- Base model: Llama-3.1-8B |
|
|
- Training dataset: tulu3_mixture_math_reasoning |
|
|
- Learning rate: 5e-06 |
|
|
- Effective batch size: 128 |
|
|
|
|
|
## Export Files |
|
|
This repository includes export files for state averaging and other advanced techniques. |
|
|
|