ALaRM: Align Language Models via Hierarchical Rewards Modeling
Paper
•
2403.06754
•
Published
Trained SFT policy for MT task in the paper "ALaRM: Align Language Models via Hierarchical Rewards Modeling".
Check out our project page for more information.