ztime commited on
Commit
2a6826a
1 Parent(s): 0dc41c6

change port export

Browse files
Files changed (3) hide show
  1. Dockerfile +3 -2
  2. README.md +4 -4
  3. start_server.sh +1 -1
Dockerfile CHANGED
@@ -18,9 +18,10 @@ RUN git clone https://github.com/ggerganov/llama.cpp.git llamacpp --depth 1 && \
18
  cd llamacpp && \
19
  cmake -B build -DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS && \
20
  cmake --build build --config Release --target main server && \
21
- cp build/bin/* ~/ && \
22
  touch /llama.log && \
23
- chmod 777 /llama.log
 
24
 
25
  # Download model
26
  RUN mkdir model && \
 
18
  cd llamacpp && \
19
  cmake -B build -DLLAMA_BLAS=ON -DLLAMA_BLAS_VENDOR=OpenBLAS && \
20
  cmake --build build --config Release --target main server && \
21
+ cp build/bin/* ../ && \
22
  touch /llama.log && \
23
+ chmod 777 /llama.log && \
24
+ ls -lt
25
 
26
  # Download model
27
  RUN mkdir model && \
README.md CHANGED
@@ -1,20 +1,20 @@
1
  ---
2
- title: OpenHermes-2.5-Mistral-7B-GGUF (Q4_K_M)
3
  colorFrom: purple
4
  colorTo: blue
5
  sdk: docker
6
  models:
7
  - teknium/OpenHermes-2.5-Mistral-7B
8
- - TheBloke/OpenHermes-2.5-Mistral-7B-GGUF
9
  tags:
10
  - inference api
11
  - openai-api compatible
12
  - llama-cpp-python
13
- - OpenHermes-2.5-Mistral-7B-GGUF
14
  - gguf
15
  pinned: false
16
  ---
17
 
18
- # OpenHermes-2.5-Mistral-7B-GGUF (Q4_K_M)
19
 
20
  Please refer to the [index.html](index.html) for more information.
 
1
  ---
2
+ title: openchat-3.5-0106-GGUF (Q4_K_M)
3
  colorFrom: purple
4
  colorTo: blue
5
  sdk: docker
6
  models:
7
  - teknium/OpenHermes-2.5-Mistral-7B
8
+ - TheBloke/openchat-3.5-0106-GGUF
9
  tags:
10
  - inference api
11
  - openai-api compatible
12
  - llama-cpp-python
13
+ - openchat-3.5-0106-GGUF
14
  - gguf
15
  pinned: false
16
  ---
17
 
18
+ # openchat-3.5-0106-GGUF (Q4_K_M)
19
 
20
  Please refer to the [index.html](index.html) for more information.
start_server.sh CHANGED
@@ -3,7 +3,7 @@
3
  # For mlock support
4
  ulimit -l unlimited
5
 
6
-
7
 
8
  /server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
9
  /llamacpp/build/bin/server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
 
3
  # For mlock support
4
  ulimit -l unlimited
5
 
6
+ /main -m model/gguf-model.bin --color -p '你好,你是谁'
7
 
8
  /server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin
9
  /llamacpp/build/bin/server --host 0.0.0.0 --port 7860 -m model/gguf-model.bin