Error running ether0-8bit with mlx

#1
by cafierom - opened

Model loads fine, but when running an inference

if tokenizer.chat_template is not None:
    messages = [{"role": "user", "content": prompt}]
    prompt = tokenizer.apply_chat_template(
        conversation=messages, add_generation_prompt=True
    )
response = generate(
    model=model,
    tokenizer=tokenizer,
    prompt=prompt,
)

I get the error:
[metal::device_info] Cannot get device info without metal backend

Sign up or log in to comment