llama-2-gguf / Dockerfile
mvasim's picture
first commit
81cf53b
raw history blame
No virus
559 Bytes
# read the doc: https://huggingface.co/docs/hub/spaces-sdks-docker
# you will also find guides on how best to write your Dockerfile
FROM python:3.9
WORKDIR /code
ENV REPO=TheBloke/Llama-2-7B-Chat-GGUF
ENV MODEL_NAME=llama-2-7b-chat.Q5_K_M.gguf
COPY ./requirements.txt /code/requirements.txt
RUN pip install --no-cache-dir --upgrade -r /code/requirements.txt
COPY . .
RUN !huggingface-cli download \
${REPO} \
${MODEL_NAME}\
--local-dir . \
--local-dir-use-symlinks False
CMD ["uvicorn", "app.main:app", "--host", "0.0.0.0", "--port", "7860"]