koh1018 commited on
Commit
dd11498
ยท
1 Parent(s): 8b098af

Add llama-cpp-python wheel via Git LFS

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ *.whl filter=lfs diff=lfs merge=lfs -text
Dockerfile CHANGED
@@ -19,20 +19,16 @@ USER user
19
  ENV HF_HOME /app/huggingface_cache
20
  ENV PATH="/home/user/.local/bin:${PATH}"
21
 
22
- # 7. requirements.txt ํŒŒ์ผ์„ ๋ณต์‚ฌ.
23
  COPY ./requirements.txt requirements.txt
 
24
 
25
- # 8. llama-cpp-python์„ '๋˜‘๋˜‘ํ•˜๊ฒŒ' ์„ค์น˜ํ•˜๊ธฐ ์œ„ํ•œ ๋ชจ๋“  ์˜ต์…˜์„ ์„ค์ •.
26
- ENV CMAKE_ARGS="-DLLAMA_CUBLAS=OFF -DLLAMA_HIPBLAS=OFF -DLLAMA_OPENBLAS=OFF -DLLAMA_CMAKE_BUILD_PARALLEL=ON"
27
- ENV FORCE_CMAKE=1
28
- ENV CC=gcc-11
29
- ENV CXX=g++-11
30
-
31
- # 9. ๋ผ์ด๋ธŒ๋Ÿฌ๋ฆฌ ์„ค์น˜. (๊ฒฝ๋กœ ์ˆ˜์ • ์™„๋ฃŒ!)
32
  RUN pip install --no-cache-dir --upgrade -r requirements.txt
33
 
34
- # 10. ๋‚˜๋จธ์ง€ ๋ชจ๋“  ์†Œ์Šค ์ฝ”๋“œ๋ฅผ ์ž‘์—… ํด๋”์— ๋ณต์‚ฌ.
35
  COPY . /app
36
 
37
- # 11. ๋ชจ๋“  ์ค€๋น„๊ฐ€ ๋๋‚˜๋ฉด, ์„œ๋ฒ„๋ฅผ ์‹คํ–‰.
38
  CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"]
 
19
  ENV HF_HOME /app/huggingface_cache
20
  ENV PATH="/home/user/.local/bin:${PATH}"
21
 
22
+ # 7. requirements.txt ํŒŒ์ผ๊ณผ wheel ํŒŒ์ผ์„ ๋ณต์‚ฌ.
23
  COPY ./requirements.txt requirements.txt
24
+ COPY ./llama_cpp_python-0.3.11-cp39-cp39-linux_x86_64.whl .
25
 
26
+ # 8. llama-cpp-python์„ wheel ํŒŒ์ผ๋กœ ์„ค์น˜ ํ›„ ๋‚˜๋จธ์ง€ ๋ผ์ด๋ธŒ๋Ÿฌ๋ฆฌ ์„ค์น˜.
27
+ RUN pip install --no-cache-dir ./llama_cpp_python-0.3.11-cp39-cp39-linux_x86_64.whl
 
 
 
 
 
28
  RUN pip install --no-cache-dir --upgrade -r requirements.txt
29
 
30
+ # 9. ๋‚˜๋จธ์ง€ ๋ชจ๋“  ์†Œ์Šค ์ฝ”๋“œ๋ฅผ ์ž‘์—… ํด๋”์— ๋ณต์‚ฌ.
31
  COPY . /app
32
 
33
+ # 10. ๋ชจ๋“  ์ค€๋น„๊ฐ€ ๋๋‚˜๋ฉด, ์„œ๋ฒ„๋ฅผ ์‹คํ–‰.
34
  CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"]
app.py CHANGED
@@ -13,12 +13,12 @@ app = FastAPI()
13
  # 2. GGUF ๋ชจ๋ธ ๋กœ๋”ฉ ์ค€๋น„
14
  # # TheBloke์˜ SOLAR ๋ชจ๋ธ์„ ์˜ˆ์‹œ๋กœ ์‚ฌ์šฉ.
15
  # # 'repo_id'๋Š” ๋ชจ๋ธ์ด ์žˆ๋Š” ์ €์žฅ์†Œ, 'filename'์€ ๊ทธ ์•ˆ์˜ ํŠน์ • GGUF ํŒŒ์ผ๋ช….
16
- # model_repo_id = "TheBloke/SOLAR-10.7B-Instruct-v1.0-GGUF"
17
- # model_filename = "solar-10.7b-instruct-v1.0.Q4_K_S.gguf"
18
 
19
- # 7B ๋ชจ๋ธ์ธ Qwen 2.5๋กœ ๋ณ€๊ฒฝํ•˜์—ฌ ํ…Œ์ŠคํŠธ
20
- model_repo_id = "Triangle104/Qwen2.5-7B-Instruct-Q4_K_S-GGUF"
21
- model_filename = "qwen2.5-7b-instruct-q4_k_s.gguf"
22
 
23
  # Hugging Face Hub์—์„œ GGUF ํŒŒ์ผ์„ ๋‹ค์šด๋กœ๋“œํ•˜๊ณ , ๋กœ์ปฌ ๊ฒฝ๋กœ๋ฅผ ๊ฐ€์ ธ์˜จ๋‹ค.
24
  # ์ด ๊ณผ์ •์€ ์„œ๋ฒ„ ์‹œ์ž‘ ์‹œ ํ•œ๋ฒˆ๋งŒ ์‹คํ–‰๋œ๋‹ค.
 
13
  # 2. GGUF ๋ชจ๋ธ ๋กœ๋”ฉ ์ค€๋น„
14
  # # TheBloke์˜ SOLAR ๋ชจ๋ธ์„ ์˜ˆ์‹œ๋กœ ์‚ฌ์šฉ.
15
  # # 'repo_id'๋Š” ๋ชจ๋ธ์ด ์žˆ๋Š” ์ €์žฅ์†Œ, 'filename'์€ ๊ทธ ์•ˆ์˜ ํŠน์ • GGUF ํŒŒ์ผ๋ช….
16
+ model_repo_id = "TheBloke/SOLAR-10.7B-Instruct-v1.0-GGUF"
17
+ model_filename = "solar-10.7b-instruct-v1.0.Q4_K_S.gguf"
18
 
19
+ # # 7B ๋ชจ๋ธ์ธ Qwen 2.5๋กœ ๋ณ€๊ฒฝํ•˜์—ฌ ํ…Œ์ŠคํŠธ
20
+ # model_repo_id = "Triangle104/Qwen2.5-7B-Instruct-Q4_K_S-GGUF"
21
+ # model_filename = "qwen2.5-7b-instruct-q4_k_s.gguf"
22
 
23
  # Hugging Face Hub์—์„œ GGUF ํŒŒ์ผ์„ ๋‹ค์šด๋กœ๋“œํ•˜๊ณ , ๋กœ์ปฌ ๊ฒฝ๋กœ๋ฅผ ๊ฐ€์ ธ์˜จ๋‹ค.
24
  # ์ด ๊ณผ์ •์€ ์„œ๋ฒ„ ์‹œ์ž‘ ์‹œ ํ•œ๋ฒˆ๋งŒ ์‹คํ–‰๋œ๋‹ค.
llama_cpp_python-0.3.11-cp39-cp39-linux_x86_64.whl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b23acdcb9de8fd44ccb85b73ece0aa211f31717697d3b5c5f3d37acbee095a04
3
+ size 4176374
requirements.txt CHANGED
@@ -4,5 +4,4 @@ pydantic
4
  python-multipart
5
  torch
6
  sentencepiece
7
- llama-cpp-python
8
  huggingface-hub
 
4
  python-multipart
5
  torch
6
  sentencepiece
 
7
  huggingface-hub