Spaces:
Sleeping
Sleeping
title: docker ollama model | |
short_description: REST API to serve a model using Docker and Ollama | |
emoji: 🔥 | |
colorFrom: green | |
colorTo: blue | |
sdk: docker | |
pinned: false | |
To do inference, use command line | |
curl https://myyim-docker-ollama-model.hf.space/api/generate -d '{ | |
"model": "gemma3:1b", | |
"prompt": "Write me a poem about generative AI.", | |
"stream": false | |
}' | |
(https://medium.com/p/1f5d8f871887) | |
or | |
import base64 | |
import requests | |
image_path = 'your_image.jpg' | |
with open(image_path, 'rb') as image_file: | |
image_data = image_file.read() | |
base64_encoded_data = base64.b64encode(image_data) | |
base64_string = base64_encoded_data.decode('utf-8') | |
headers = { | |
'Content-Type': 'application/x-www-form-urlencoded', | |
} | |
model = "gemma3:4b" | |
prompt = "Describe the image in details and give it a caption." | |
ollamaURL = 'https://myyim-docker-ollama-model.hf.space/api/generate' | |
data = { | |
"model": model, | |
"prompt": prompt, | |
"stream": False, | |
"images": [base64_string] | |
} | |
response = requests.post(ollamaURL, headers=headers, json=data) | |
print(response.json()["response"]) | |
(https://medium.com/@manyi.yim/deploy-ollama-models-on-hugging-face-spaces-with-python-library-requests-503ac6b5ca04) |