FROM nvidia/cuda:11.3.1-cudnn8-devel-ubuntu18.04 ARG LLAMA_CPP_VERSION="0.1.52" ARG CMAKE_VERSION=3.26 ARG CMAKE_VERSION_PATCH=3.26.3 ARG CMAKE_OS=linux ARG DEBIAN_FRONTEND=noninteractive ENV TZ=UTC RUN apt-get update && \ apt-get install --no-install-recommends -y \ curl git vim build-essential software-properties-common python3 python3-pip python3-dev python3-venv \ libffi-dev libncurses5-dev zlib1g zlib1g-dev libreadline-dev libbz2-dev libsqlite3-dev libssl-dev \ libblas-dev liblapack-dev libopenblas-dev cmake && \ add-apt-repository ppa:ubuntu-toolchain-r/test && \ apt-get update && \ apt install --no-install-recommends -y gcc-10 g++-10 && \ update-alternatives --install /usr/bin/gcc gcc /usr/bin/gcc-10 100 --slave /usr/bin/g++ g++ /usr/bin/g++-10 --slave /usr/bin/gcov gcov /usr/bin/gcov-10 && \ rm -rf /var/lib/apt/lists/* && \ pip3 install scikit-build RUN curl -L https://cmake.org/files/v$CMAKE_VERSION/cmake-$CMAKE_VERSION_PATCH-$CMAKE_OS-x86_64.sh -o /tmp/cmake-$CMAKE_VERSION_PATCH-$CMAKE_OS-x86_64.sh && \ mkdir /opt/cmake && \ sh /tmp/cmake-$CMAKE_VERSION_PATCH-$CMAKE_OS-x86_64.sh --skip-license --prefix=/opt/cmake && \ ln -s /opt/cmake/bin/cmake /usr/local/bin/cmake RUN useradd -m -u 1000 appuser WORKDIR /build RUN chown appuser:appuser /build USER appuser ENV HOME /home/appuser ENV PYENV_ROOT $HOME/.pyenv ENV PATH $PYENV_ROOT/shims:$PYENV_ROOT/bin:$PATH RUN git clone --depth 1 --branch v$LLAMA_CPP_VERSION https://github.com/abetlen/llama-cpp-python.git /build RUN git clone https://github.com/ggerganov/llama.cpp.git /build/vendor/llama.cpp RUN curl https://pyenv.run | bash RUN pyenv install 3.8.9 && \ pyenv global 3.8.9 && \ pyenv rehash && \ pip install --no-cache-dir --upgrade pip==22.3.1 setuptools wheel && \ pip install --no-cache-dir datasets "huggingface-hub>=0.12.1" "protobuf<4" "click<8.1" "scikit-build" && \ CMAKE_ARGS="-DLLAMA_CUBLAS=on -DLLAMA_OPENBLAS=off" FORCE_CMAKE=1 python3 setup.py bdist_wheel && \ mkdir /build/dists/ && \ cp dist/llama_cpp_python-${LLAMA_CPP_VERSION}-cp38-cp38-linux_x86_64.whl dists/llama_cpp_python-gpu-${LLAMA_CPP_VERSION}-cp38-cp38-linux_x86_64.whl && \ CMAKE_ARGS="-DLLAMA_CUBLAS=off -DLLAMA_OPENBLAS=off" FORCE_CMAKE=1 python3 setup.py bdist_wheel && \ cp dist/llama_cpp_python-${LLAMA_CPP_VERSION}-cp38-cp38-linux_x86_64.whl dists/llama_cpp_python-cpu-${LLAMA_CPP_VERSION}-cp38-cp38-linux_x86_64.whl && \ ls -l /build/dists/