ShangZhu-Together commited on
Commit
b76cd90
·
verified ·
1 Parent(s): 38b6210

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -8,7 +8,7 @@ tags: []
8
 
9
  ## Model Description
10
 
11
- This is the SFT model model in our Mixture of Agents Alignment (MoAA) pipeline. This model is tuned on the Gemma-2-9b-it. MoAA is an approach that leverages collective intelligence from open‑source LLMs to advance alignment.
12
 
13
  Two mains stages are involved in our MoAA method. In the first stage, we employ MoA to produce high-quality synthetic data for supervised fine-tuning. In the second stage, we combines multiple LLMs as a reward model to provide preference annotations.
14
 
@@ -64,7 +64,7 @@ Refer to [Paper](https://arxiv.org/abs/2505.03059) for metrics.
64
 
65
 
66
 
67
- ## Citation [optional]
68
  ```
69
  @article{wang2025improving,
70
  title = {Improving Model Alignment Through Collective Intelligence of Open-Source LLMS},
 
8
 
9
  ## Model Description
10
 
11
+ This is the SFT model in our Mixture of Agents Alignment (MoAA) pipeline. This model is tuned on the Gemma-2-9b-it. MoAA is an approach that leverages collective intelligence from open‑source LLMs to advance alignment.
12
 
13
  Two mains stages are involved in our MoAA method. In the first stage, we employ MoA to produce high-quality synthetic data for supervised fine-tuning. In the second stage, we combines multiple LLMs as a reward model to provide preference annotations.
14
 
 
64
 
65
 
66
 
67
+ ## Citation
68
  ```
69
  @article{wang2025improving,
70
  title = {Improving Model Alignment Through Collective Intelligence of Open-Source LLMS},