Spaces:
Sleeping
Sleeping
import gradio as gr | |
import ollama | |
# The model name must exactly match what was pulled from Hugging Face | |
MODEL_NAME = 'hf.co/unsloth/gemma-3-4b-it-qat-GGUF:Q4_K_M' | |
def predict(prompt, history): | |
# The history is not used in this simple example, but is required by the ChatInterface | |
response = ollama.chat( | |
model=MODEL_NAME, | |
messages=[{'role': 'user', 'content': prompt}] | |
) | |
return response['message']['content'] | |
# Setup the Gradio Chat Interface | |
iface = gr.ChatInterface( | |
fn=predict, | |
title="Gemma-3 QAT GGUF Chat", | |
description=f"Chat with the {MODEL_NAME} model via Ollama." | |
) | |
# Launch the interface | |
iface.launch(server_name="0.0.0.0", server_port=7860) |