File size: 928 Bytes
fdbf4ad
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1d4abd6
 
26c956a
1d4abd6
 
 
 
 
 
 
fdbf4ad
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
---
language:
  - en
license: cc-by-nc-4.0
library_name: transformers
tags:
  - llama
  - math
  - reasoning
  - fine-tuned
  - fine-tuning
pipeline_tag: text-generation
model-index:
  - name: Llama-3.1-8B-math-reasoning
    results:
      - task:
          type: text-generation
          name: Text Generation
        dataset:
          name: tulu3_mixture_math_reasoning
          type: custom
        metrics:
          - name: Training Loss
            type: loss
            value: 0.98
base_model: meta-llama/Llama-3.1-8B
---

# Llama-3.1-8B Math Reasoning Model

Llama-3.1-8B SFT checkpoints for mathematical reasoning—artifacts of https://arxiv.org/abs/2509.11167. 

## Model Details
- Base model: Llama-3.1-8B
- Training dataset: tulu3_mixture_math_reasoning
- Learning rate: 5e-06
- Effective batch size: 128

## Export Files
This repository includes export files for state averaging and other advanced techniques.