File size: 685 Bytes
ec7e24f
0b5e429
 
 
e78e0f3
 
 
 
91bb037
ec7e24f
 
 
 
8b66575
 
 
 
84d11e6
8b66575
0335637
f663c35
ec7e24f
cd41c7b
0b5e429
 
 
 
 
 
1ad978f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
FROM ubuntu:22.04

WORKDIR /app

# ARG HOME= /app
# # RUN mkdir ${HOME}
# ENV SENTENCE_TRANSFORMERS_HOME=${HOME}
# ENV TORCH_HOME=${HOME}

RUN apt update \
    && apt install python3.10 pip -y \
    && rm -rf /var/lib/apt/lists/*

# # Install problematic flash attention from source
# RUN apt update \
#     && apt install git -y \
#     && rm -rf /var/lib/apt/lists/*

# RUN pip install git+https://github.com/Dao-AILab/flash-attention.git

COPY requirements.txt .
RUN python3.10 -m pip install --no-cache-dir -r requirements.txt

COPY . .

# Expose the port that the application listens on.
EXPOSE 8000

# Run the application.
CMD streamlit run inference_main.py  --server.port 7860