LM Studio compatibility?
Are you working on compatibility with LM Studio so that it can be used easier?
When I search for Apertus it won't show up in LM Studio.
there is a quantization converted to MLX by mlx-community 👉 https://huggingface.co/mlx-community/Apertus-8B-2509-4bit, this can be used in LM Studio
GGUF My Repo don't work too.
could you try again now that is has been merged into MLX-LM officially? https://github.com/ml-explore/mlx-lm/releases/tag/v0.27.1
Tried it since launch and just now again, with the latest update to LMS 0.3.25 (build 2) i still get the same error with the available MLX versions from mlx-community and NexVeridian.
Error when loading model: ValueError: Model type apertus not supported.
yes this seems it's not using mlx 27.1 yet but still 27 internally
in the meantime i guess you have to use MLX-LM directly, or wait till next week's next LMstudio release. i created an issue here: https://github.com/lmstudio-ai/lmstudio-bug-tracker/issues/975
Today when starting my installed LMS 0.3.25 (build 2) came update to the LM Runtime to 0.26.1 with mlx-lm 0.27.1. Now Apertus starts without errors and runs! Great 👍😎😊
Hi,
I tried to load the model Apertus-8B-2509 in LM Studio (Windows, version 0.3.27 build 4), but I get the following error:
Failed to load model
error loading model: error loading model architecture: unknown model architecture: 'apertus'
From what I understand, LM Studio on Windows uses llama.cpp/GGUF under the hood, and it seems the ApertusForCausalLM architecture is not yet supported there. On macOS with MLX Runtime this might work, but Windows doesn’t have MLX.
Questions:
Is there already an official GGUF export or roadmap for Apertus so it can run in LM Studio / llama.cpp on Windows?
If not, would you recommend using transformers + PyTorch/VLLM for now instead of trying LM Studio?
Are there community-converted GGUF weights that you would consider “safe” to use, or is that not recommended yet?
Thanks for the clarification and all your great work on this model!

