Spaces:
Sleeping
Sleeping
Add llama-cpp-python wheel via Git LFS
Browse files- .gitattributes +1 -0
- Dockerfile +6 -10
- app.py +5 -5
- llama_cpp_python-0.3.11-cp39-cp39-linux_x86_64.whl +3 -0
- requirements.txt +0 -1
.gitattributes
CHANGED
|
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
*.whl filter=lfs diff=lfs merge=lfs -text
|
Dockerfile
CHANGED
|
@@ -19,20 +19,16 @@ USER user
|
|
| 19 |
ENV HF_HOME /app/huggingface_cache
|
| 20 |
ENV PATH="/home/user/.local/bin:${PATH}"
|
| 21 |
|
| 22 |
-
# 7. requirements.txt ํ์ผ์ ๋ณต์ฌ.
|
| 23 |
COPY ./requirements.txt requirements.txt
|
|
|
|
| 24 |
|
| 25 |
-
# 8. llama-cpp-python์
|
| 26 |
-
|
| 27 |
-
ENV FORCE_CMAKE=1
|
| 28 |
-
ENV CC=gcc-11
|
| 29 |
-
ENV CXX=g++-11
|
| 30 |
-
|
| 31 |
-
# 9. ๋ผ์ด๋ธ๋ฌ๋ฆฌ ์ค์น. (๊ฒฝ๋ก ์์ ์๋ฃ!)
|
| 32 |
RUN pip install --no-cache-dir --upgrade -r requirements.txt
|
| 33 |
|
| 34 |
-
#
|
| 35 |
COPY . /app
|
| 36 |
|
| 37 |
-
#
|
| 38 |
CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"]
|
|
|
|
| 19 |
ENV HF_HOME /app/huggingface_cache
|
| 20 |
ENV PATH="/home/user/.local/bin:${PATH}"
|
| 21 |
|
| 22 |
+
# 7. requirements.txt ํ์ผ๊ณผ wheel ํ์ผ์ ๋ณต์ฌ.
|
| 23 |
COPY ./requirements.txt requirements.txt
|
| 24 |
+
COPY ./llama_cpp_python-0.3.11-cp39-cp39-linux_x86_64.whl .
|
| 25 |
|
| 26 |
+
# 8. llama-cpp-python์ wheel ํ์ผ๋ก ์ค์น ํ ๋๋จธ์ง ๋ผ์ด๋ธ๋ฌ๋ฆฌ ์ค์น.
|
| 27 |
+
RUN pip install --no-cache-dir ./llama_cpp_python-0.3.11-cp39-cp39-linux_x86_64.whl
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
RUN pip install --no-cache-dir --upgrade -r requirements.txt
|
| 29 |
|
| 30 |
+
# 9. ๋๋จธ์ง ๋ชจ๋ ์์ค ์ฝ๋๋ฅผ ์์
ํด๋์ ๋ณต์ฌ.
|
| 31 |
COPY . /app
|
| 32 |
|
| 33 |
+
# 10. ๋ชจ๋ ์ค๋น๊ฐ ๋๋๋ฉด, ์๋ฒ๋ฅผ ์คํ.
|
| 34 |
CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"]
|
app.py
CHANGED
|
@@ -13,12 +13,12 @@ app = FastAPI()
|
|
| 13 |
# 2. GGUF ๋ชจ๋ธ ๋ก๋ฉ ์ค๋น
|
| 14 |
# # TheBloke์ SOLAR ๋ชจ๋ธ์ ์์๋ก ์ฌ์ฉ.
|
| 15 |
# # 'repo_id'๋ ๋ชจ๋ธ์ด ์๋ ์ ์ฅ์, 'filename'์ ๊ทธ ์์ ํน์ GGUF ํ์ผ๋ช
.
|
| 16 |
-
|
| 17 |
-
|
| 18 |
|
| 19 |
-
# 7B ๋ชจ๋ธ์ธ Qwen 2.5๋ก ๋ณ๊ฒฝํ์ฌ ํ
์คํธ
|
| 20 |
-
model_repo_id = "Triangle104/Qwen2.5-7B-Instruct-Q4_K_S-GGUF"
|
| 21 |
-
model_filename = "qwen2.5-7b-instruct-q4_k_s.gguf"
|
| 22 |
|
| 23 |
# Hugging Face Hub์์ GGUF ํ์ผ์ ๋ค์ด๋ก๋ํ๊ณ , ๋ก์ปฌ ๊ฒฝ๋ก๋ฅผ ๊ฐ์ ธ์จ๋ค.
|
| 24 |
# ์ด ๊ณผ์ ์ ์๋ฒ ์์ ์ ํ๋ฒ๋ง ์คํ๋๋ค.
|
|
|
|
| 13 |
# 2. GGUF ๋ชจ๋ธ ๋ก๋ฉ ์ค๋น
|
| 14 |
# # TheBloke์ SOLAR ๋ชจ๋ธ์ ์์๋ก ์ฌ์ฉ.
|
| 15 |
# # 'repo_id'๋ ๋ชจ๋ธ์ด ์๋ ์ ์ฅ์, 'filename'์ ๊ทธ ์์ ํน์ GGUF ํ์ผ๋ช
.
|
| 16 |
+
model_repo_id = "TheBloke/SOLAR-10.7B-Instruct-v1.0-GGUF"
|
| 17 |
+
model_filename = "solar-10.7b-instruct-v1.0.Q4_K_S.gguf"
|
| 18 |
|
| 19 |
+
# # 7B ๋ชจ๋ธ์ธ Qwen 2.5๋ก ๋ณ๊ฒฝํ์ฌ ํ
์คํธ
|
| 20 |
+
# model_repo_id = "Triangle104/Qwen2.5-7B-Instruct-Q4_K_S-GGUF"
|
| 21 |
+
# model_filename = "qwen2.5-7b-instruct-q4_k_s.gguf"
|
| 22 |
|
| 23 |
# Hugging Face Hub์์ GGUF ํ์ผ์ ๋ค์ด๋ก๋ํ๊ณ , ๋ก์ปฌ ๊ฒฝ๋ก๋ฅผ ๊ฐ์ ธ์จ๋ค.
|
| 24 |
# ์ด ๊ณผ์ ์ ์๋ฒ ์์ ์ ํ๋ฒ๋ง ์คํ๋๋ค.
|
llama_cpp_python-0.3.11-cp39-cp39-linux_x86_64.whl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b23acdcb9de8fd44ccb85b73ece0aa211f31717697d3b5c5f3d37acbee095a04
|
| 3 |
+
size 4176374
|
requirements.txt
CHANGED
|
@@ -4,5 +4,4 @@ pydantic
|
|
| 4 |
python-multipart
|
| 5 |
torch
|
| 6 |
sentencepiece
|
| 7 |
-
llama-cpp-python
|
| 8 |
huggingface-hub
|
|
|
|
| 4 |
python-multipart
|
| 5 |
torch
|
| 6 |
sentencepiece
|
|
|
|
| 7 |
huggingface-hub
|