leafspark commited on
Commit
47eaf19
1 Parent(s): 179c68f

tool: add linux autorun (no CUDA, ubuntu)

Browse files
DeepSeek-V2-Chat.q2_k.gguf/download_and_run.sh ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/bin/bash
2
+
3
+ # Create and change to the directory
4
+ mkdir -p DeepSeek-V2-Chat.Q2_K.gguf
5
+ cd DeepSeek-V2-Chat.Q2_K.gguf
6
+
7
+ # Download the GGUF files
8
+ for i in {1..5}; do
9
+ wget "https://huggingface.co/leafspark/DeepSeek-V2-Chat-GGUF/resolve/main/DeepSeek-V2-Chat.q2_k.gguf/DeepSeek-V2-Chat.Q2_K-0000$i-of-00005.gguf?download=true" -O DeepSeek-V2-Chat.Q2_K-0000$i-of-00005.gguf
10
+ done
11
+
12
+ # Download the llama.cpp binaries based on the OS
13
+ case "$(uname -s)" in
14
+ Linux)
15
+ wget https://github.com/ggerganov/llama.cpp/releases/download/b2961/llama-b2961-bin-ubuntu-x64.zip
16
+ unzip llama-b2961-bin-ubuntu-x64.zip -d .
17
+ ;;
18
+ Darwin)
19
+ if [[ $(uname -m) == 'arm64' ]]; then
20
+ wget https://github.com/ggerganov/llama.cpp/releases/download/b2961/llama-b2961-bin-macos-arm64.zip
21
+ unzip llama-b2961-bin-macos-arm64.zip -d .
22
+ else
23
+ wget https://github.com/ggerganov/llama.cpp/releases/download/b2961/llama-b2961-bin-macos-x64.zip
24
+ unzip llama-b2961-bin-macos-x64.zip -d .
25
+ fi
26
+ ;;
27
+ esac
28
+
29
+ # Execute the server command
30
+ ./server \
31
+ -m DeepSeek-V2-Chat.q2_k.gguf \
32
+ -c 4096 \
33
+ -i \
34
+ --mlock
35
+ --override-kv deepseek2.attention.q_lora_rank=int:1536
36
+ --override-kv deepseek2.attention.kv_lora_rank=int:512
37
+ --override-kv deepseek2.expert_shared_count=int:2
38
+ --override-kv deepseek2.expert_feed_forward_length=int:1536
39
+ --override-kv deepseek2.leading_dense_block_count=int:1