IntelligentEstate/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF a Thinking or Brooding Model

This model Is showing impressive speed and accuracy with situational and complex math reasoning the Group Realative Policy Optimization applied has shown to smooth out some of the overall issues while adding accuracy in the small model when quantized.(So much so the process inspire the creation of berkley's DeepScaleR-1) Showing improvements in my favorite little model and giving it a bit more cohearance and in many areas overtaking Qwen's 2.5 3B original model. A VERY impressive Model by FBLGIT. and out of 10 different Quants this one seems to be the best in local intel based testing.

Screenshot 2024-12-18 at 09-00-31 Ideogram.png

Still the best little guy for it's size, THANKS for the present FBLGIT/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF

This model was converted to GGUF format from fblgit/miniclaus-qw1.5B-UNAMGS-GRPO using llama.cpp via the ggml.ai's GGUF-my-repo space. Refer to the original model card for more details on the model.

Use with llama.cpp

Install llama.cpp through brew (works on Mac and Linux)

brew install llama.cpp

Invoke the llama.cpp server or the CLI.

GPT4ALL/Ollama: use standard qwen templates/prompting opening context window for length may applie qwen "Reasoner" template for analysing function.

Downloads last month
3
GGUF
Model size
1.54B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for IntelligentEstate/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF

Dataset used to train IntelligentEstate/miniclaus-qw1.5B-UNAMGS-GRPO-Q8_0-GGUF