Spaces:
Sleeping
title: docker ollama model
short_description: REST API to serve a model using Docker and Ollama
emoji: 🔥
colorFrom: green
colorTo: blue
sdk: docker
pinned: false
To do inference, use command line
curl https://myyim-docker-ollama-model.hf.space/api/generate -d '{ "model": "gemma3:1b", "prompt": "Write me a poem about generative AI.", "stream": false }'
(https://medium.com/p/1f5d8f871887)
or
import base64 import requests
image_path = 'your_image.jpg' with open(image_path, 'rb') as image_file: image_data = image_file.read() base64_encoded_data = base64.b64encode(image_data) base64_string = base64_encoded_data.decode('utf-8')
headers = { 'Content-Type': 'application/x-www-form-urlencoded', }
model = "gemma3:4b" prompt = "Describe the image in details and give it a caption." ollamaURL = 'https://myyim-docker-ollama-model.hf.space/api/generate'
data = { "model": model, "prompt": prompt, "stream": False, "images": [base64_string] }
response = requests.post(ollamaURL, headers=headers, json=data) print(response.json()["response"])