Spaces:
Sleeping
Sleeping
| FROM python:3.11.1-bullseye | |
| USER root | |
| ENV PROTOCOL_BUFFERS_PYTHON_IMPLEMENTATION=python | |
| RUN apt install g++ -y | |
| RUN git clone https://github.com/ggerganov/llama.cpp.git | |
| RUN cd llama.cpp | |
| RUN make | |
| RUN wget https://huggingface.co/OpenBuddy/openbuddy-ggml/resolve/main/openbuddy-openllama-3b-v10-q5_0.bin | |
| COPY . ./ | |
| RUN chmod -R 777 ./ | |
| WORKDIR ./ | |
| RUN python3 -m pip install -U --no-cache-dir pip setuptools wheel | |
| RUN pip install --no-cache-dir --upgrade -r /app/requirements.txt | |
| RUN mkdir -p /.cache/huggingface/hub | |
| RUN chown -R root:root /.cache/huggingface/hub | |
| RUN chmod -R 777 /.cache/huggingface/hub | |
| CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"] |