tool: add linux autorun (no CUDA, ubuntu)
Browse files
DeepSeek-V2-Chat.q2_k.gguf/download_and_run.sh
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
#!/bin/bash
|
2 |
+
|
3 |
+
# Create and change to the directory
|
4 |
+
mkdir -p DeepSeek-V2-Chat.Q2_K.gguf
|
5 |
+
cd DeepSeek-V2-Chat.Q2_K.gguf
|
6 |
+
|
7 |
+
# Download the GGUF files
|
8 |
+
for i in {1..5}; do
|
9 |
+
wget "https://huggingface.co/leafspark/DeepSeek-V2-Chat-GGUF/resolve/main/DeepSeek-V2-Chat.q2_k.gguf/DeepSeek-V2-Chat.Q2_K-0000$i-of-00005.gguf?download=true" -O DeepSeek-V2-Chat.Q2_K-0000$i-of-00005.gguf
|
10 |
+
done
|
11 |
+
|
12 |
+
# Download the llama.cpp binaries based on the OS
|
13 |
+
case "$(uname -s)" in
|
14 |
+
Linux)
|
15 |
+
wget https://github.com/ggerganov/llama.cpp/releases/download/b2961/llama-b2961-bin-ubuntu-x64.zip
|
16 |
+
unzip llama-b2961-bin-ubuntu-x64.zip -d .
|
17 |
+
;;
|
18 |
+
Darwin)
|
19 |
+
if [[ $(uname -m) == 'arm64' ]]; then
|
20 |
+
wget https://github.com/ggerganov/llama.cpp/releases/download/b2961/llama-b2961-bin-macos-arm64.zip
|
21 |
+
unzip llama-b2961-bin-macos-arm64.zip -d .
|
22 |
+
else
|
23 |
+
wget https://github.com/ggerganov/llama.cpp/releases/download/b2961/llama-b2961-bin-macos-x64.zip
|
24 |
+
unzip llama-b2961-bin-macos-x64.zip -d .
|
25 |
+
fi
|
26 |
+
;;
|
27 |
+
esac
|
28 |
+
|
29 |
+
# Execute the server command
|
30 |
+
./server \
|
31 |
+
-m DeepSeek-V2-Chat.q2_k.gguf \
|
32 |
+
-c 4096 \
|
33 |
+
-i \
|
34 |
+
--mlock
|
35 |
+
--override-kv deepseek2.attention.q_lora_rank=int:1536
|
36 |
+
--override-kv deepseek2.attention.kv_lora_rank=int:512
|
37 |
+
--override-kv deepseek2.expert_shared_count=int:2
|
38 |
+
--override-kv deepseek2.expert_feed_forward_length=int:1536
|
39 |
+
--override-kv deepseek2.leading_dense_block_count=int:1
|