Introducing GRM2, a powerful 3b parameter model designed for long-term reasoning and high performance in complex tasks.
Even with only 3b of parameters, it outperforms qwen3-32b in several benchmarks.
With only 3b of parameters, it can also generate large and complex code of over 1000 lines, use tools in a way comparable to large models, and is perfect for agentic tasks.
GRM2 is licensed under Apache 2.0, making it perfect as a FineTune base for other tasks.
While not good at complex code, we observed a significant improvement in code generation (especially in Python code), demonstrating that, when trained correctly, small AIs can, in fact, program.
Introducing GRM Family, a family of fine-tuned small models from the Qwen2.5 family for Long Cot and General Reasoning and Agentic Tasks.
GRM is available in 7b and 1.5b parameter sizes, these models being significantly relevant for complex tasks or local inference agents. OrionLLM/GRM-7b OrionLLM/GRM-1.5b