Fadikkop commited on
Commit
bde311a
1 Parent(s): 4edc3c4

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +27 -12
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  language:
3
  - zh
4
  - en
@@ -8,41 +9,55 @@ language:
8
  - ko
9
  - it
10
  - ru
11
- license: apache-2.0
12
  library_name: transformers
 
 
13
  tags:
14
  - llama-cpp
15
  - gguf-my-repo
16
- pipeline_tag: text-generation
17
  inference: false
18
  ---
19
 
20
  # Fadikkop/openbuddy-openllama-3b-v10-bf16-Q4_K_M-GGUF
21
  This model was converted to GGUF format from [`OpenBuddy/openbuddy-openllama-3b-v10-bf16`](https://huggingface.co/OpenBuddy/openbuddy-openllama-3b-v10-bf16) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
22
  Refer to the [original model card](https://huggingface.co/OpenBuddy/openbuddy-openllama-3b-v10-bf16) for more details on the model.
23
- ## Use with llama.cpp
24
 
25
- Install llama.cpp through brew.
 
26
 
27
  ```bash
28
- brew install ggerganov/ggerganov/llama.cpp
 
29
  ```
30
  Invoke the llama.cpp server or the CLI.
31
 
32
- CLI:
33
-
34
  ```bash
35
- llama-cli --hf-repo Fadikkop/openbuddy-openllama-3b-v10-bf16-Q4_K_M-GGUF --model openbuddy-openllama-3b-v10-bf16.Q4_K_M.gguf -p "The meaning to life and the universe is"
36
  ```
37
 
38
- Server:
39
-
40
  ```bash
41
- llama-server --hf-repo Fadikkop/openbuddy-openllama-3b-v10-bf16-Q4_K_M-GGUF --model openbuddy-openllama-3b-v10-bf16.Q4_K_M.gguf -c 2048
42
  ```
43
 
44
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
45
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46
  ```
47
- git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m openbuddy-openllama-3b-v10-bf16.Q4_K_M.gguf -n 128
48
  ```
 
1
  ---
2
+ base_model: OpenBuddy/openbuddy-openllama-3b-v10-bf16
3
  language:
4
  - zh
5
  - en
 
9
  - ko
10
  - it
11
  - ru
 
12
  library_name: transformers
13
+ license: apache-2.0
14
+ pipeline_tag: text-generation
15
  tags:
16
  - llama-cpp
17
  - gguf-my-repo
 
18
  inference: false
19
  ---
20
 
21
  # Fadikkop/openbuddy-openllama-3b-v10-bf16-Q4_K_M-GGUF
22
  This model was converted to GGUF format from [`OpenBuddy/openbuddy-openllama-3b-v10-bf16`](https://huggingface.co/OpenBuddy/openbuddy-openllama-3b-v10-bf16) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
23
  Refer to the [original model card](https://huggingface.co/OpenBuddy/openbuddy-openllama-3b-v10-bf16) for more details on the model.
 
24
 
25
+ ## Use with llama.cpp
26
+ Install llama.cpp through brew (works on Mac and Linux)
27
 
28
  ```bash
29
+ brew install llama.cpp
30
+
31
  ```
32
  Invoke the llama.cpp server or the CLI.
33
 
34
+ ### CLI:
 
35
  ```bash
36
+ llama-cli --hf-repo Fadikkop/openbuddy-openllama-3b-v10-bf16-Q4_K_M-GGUF --hf-file openbuddy-openllama-3b-v10-bf16-q4_k_m.gguf -p "The meaning to life and the universe is"
37
  ```
38
 
39
+ ### Server:
 
40
  ```bash
41
+ llama-server --hf-repo Fadikkop/openbuddy-openllama-3b-v10-bf16-Q4_K_M-GGUF --hf-file openbuddy-openllama-3b-v10-bf16-q4_k_m.gguf -c 2048
42
  ```
43
 
44
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
45
 
46
+ Step 1: Clone llama.cpp from GitHub.
47
+ ```
48
+ git clone https://github.com/ggerganov/llama.cpp
49
+ ```
50
+
51
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
52
+ ```
53
+ cd llama.cpp && LLAMA_CURL=1 make
54
+ ```
55
+
56
+ Step 3: Run inference through the main binary.
57
+ ```
58
+ ./llama-cli --hf-repo Fadikkop/openbuddy-openllama-3b-v10-bf16-Q4_K_M-GGUF --hf-file openbuddy-openllama-3b-v10-bf16-q4_k_m.gguf -p "The meaning to life and the universe is"
59
+ ```
60
+ or
61
  ```
62
+ ./llama-server --hf-repo Fadikkop/openbuddy-openllama-3b-v10-bf16-Q4_K_M-GGUF --hf-file openbuddy-openllama-3b-v10-bf16-q4_k_m.gguf -c 2048
63
  ```