gibberish still persists?
And i have problems - model answers is hallucinating - i have very emotional answers with different themes mixed out but i have standart assistent promt. Do you have same problem ?
Thank you for your input! This might be the result of the active paramsshared_expert being quantized into 4bit, which is due to vllm not being able to load load the model if shared_expert is not quantized.
My apologies for this, but there will be an update to the model soon in the next few days to improve the model accuracy.
thanks for you work looking forward to use your AWQ instruct model!
Thank you for your input! This might be the result of the active params
shared_expertbeing quantized into 4bit, which is due to vllm not being able to load load the model ifshared_expertis not quantized.My apologies for this, but there will be an update to the model soon in the next few days to improve the model accuracy.
Hey! Any updates ? What is your plans?