gibberish still persists?

#6
by Geximus - opened

And i have problems - model answers is hallucinating - i have very emotional answers with different themes mixed out but i have standart assistent promt. Do you have same problem ?

Owner

Thank you for your input! This might be the result of the active paramsshared_expert being quantized into 4bit, which is due to vllm not being able to load load the model if shared_expert is not quantized.

My apologies for this, but there will be an update to the model soon in the next few days to improve the model accuracy.

thanks for you work looking forward to use your AWQ instruct model!

Thank you for your input! This might be the result of the active paramsshared_expert being quantized into 4bit, which is due to vllm not being able to load load the model if shared_expert is not quantized.

My apologies for this, but there will be an update to the model soon in the next few days to improve the model accuracy.

Hey! Any updates ? What is your plans?

Owner

Hi @Geximus , thank you for your interest. There already has been a vllm update, which should improve every qwen3next models. Regarding this specific model, the update will be available in the next 36 hours.

Hi @Geximus , thank you for downloading my models :)

I have just released a recent update with active params being in bf16 and not quantized. It should match the original bf16 model more closely.

Please let me know if the problem still persists.

Sign up or log in to comment