File size: 598 Bytes
40227eb 8b359a6 f54351d bef8dc6 93f84f9 13ab4d6 8b359a6 d0e266e 13ab4d6 7423a6d 13ab4d6 7423a6d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 |
FROM python:3.11.6-bullseye
ENV LLAMA_FAST=true
RUN apt update && apt install -y libopenblas-dev
RUN git clone --recurse-submodules https://github.com/spuuntries/llava-cpp-server
WORKDIR llava-cpp-server
WORKDIR llama.cpp
RUN git pull origin master
WORKDIR ..
RUN wget https://huggingface.co/mys/ggml_llava-v1.5-7b/resolve/main/mmproj-model-f16.gguf
RUN wget https://huggingface.co/mys/ggml_llava-v1.5-7b/resolve/main/ggml-model-q4_k.gguf
RUN make
CMD ["bin/llava-server", "-m", "ggml-model-q4_k.gguf", "--mmproj", "mmproj-model-f16.gguf", "--port", "7860", "--host", "0.0.0.0", "--log-http"] |