ollamapi / entrypoint.sh
cometapii's picture
Update entrypoint.sh
5292ff7 verified
raw
history blame contribute delete
811 Bytes
#!/bin/bash
set -e
export OLLAMA_HOST=127.0.0.1:11434
export OLLAMA_MODELS=/home/user/.ollama/models
echo "==> Starting Ollama..."
ollama serve &
# Wait for ollama ready
echo "==> Waiting for Ollama to be ready..."
MAX_RETRIES=30
COUNT=0
until curl -s http://127.0.0.1:11434/api/version > /dev/null 2>&1; do
COUNT=$((COUNT + 1))
if [ $COUNT -ge $MAX_RETRIES ]; then
echo "ERROR: Ollama did not start."
exit 1
fi
echo " ... attempt $COUNT/$MAX_RETRIES"
sleep 2
done
echo "==> Ollama ready!"
# Pull models
echo "==> Pulling deepseek-r1:latest..."
ollama pull deepseek-r1:latest
echo "==> Pulling qwen3-vl:latest..."
ollama pull qwen3-vl:latest
echo "==> Models loaded:"
ollama list
echo "==> Starting proxy on :7860..."
exec uvicorn proxy:app --host 0.0.0.0 --port 7860