tinybiggames commited on
Commit
b30dc0e
1 Parent(s): 82c692c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -47
README.md CHANGED
@@ -1,51 +1,45 @@
1
- ---
2
- license: other
3
- tags:
4
- - generated_from_trainer
5
- - axolotl
6
- - llama-cpp
7
- - gguf-my-repo
8
- base_model: meta-llama/Meta-Llama-3-8B
9
- datasets:
10
- - cognitivecomputations/Dolphin-2.9
11
- - teknium/OpenHermes-2.5
12
- - m-a-p/CodeFeedback-Filtered-Instruction
13
- - cognitivecomputations/dolphin-coder
14
- - cognitivecomputations/samantha-data
15
- - microsoft/orca-math-word-problems-200k
16
- - Locutusque/function-calling-chatml
17
- - internlm/Agent-FLAN
18
- model-index:
19
- - name: out
20
- results: []
21
- ---
 
 
22
 
23
  # tinybiggames/dolphin-2.9.1-llama-3-8b-Q4_K_M-GGUF
24
  This model was converted to GGUF format from [`cognitivecomputations/dolphin-2.9.1-llama-3-8b`](https://huggingface.co/cognitivecomputations/dolphin-2.9.1-llama-3-8b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
25
  Refer to the [original model card](https://huggingface.co/cognitivecomputations/dolphin-2.9.1-llama-3-8b) for more details on the model.
26
- ## Use with llama.cpp
27
-
28
- Install llama.cpp through brew.
29
-
30
- ```bash
31
- brew install ggerganov/ggerganov/llama.cpp
32
- ```
33
- Invoke the llama.cpp server or the CLI.
34
-
35
- CLI:
36
-
37
- ```bash
38
- llama-cli --hf-repo tinybiggames/dolphin-2.9.1-llama-3-8b-Q4_K_M-GGUF --model dolphin-2.9.1-llama-3-8b.Q4_K_M.gguf -p "The meaning to life and the universe is"
39
- ```
40
-
41
- Server:
42
-
43
- ```bash
44
- llama-server --hf-repo tinybiggames/dolphin-2.9.1-llama-3-8b-Q4_K_M-GGUF --model dolphin-2.9.1-llama-3-8b.Q4_K_M.gguf -c 2048
45
- ```
46
-
47
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
48
-
49
- ```
50
- git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m dolphin-2.9.1-llama-3-8b.Q4_K_M.gguf -n 128
51
- ```
 
1
+ ---
2
+ license: other
3
+ tags:
4
+ - generated_from_trainer
5
+ - axolotl
6
+ - llama-cpp
7
+ - gguf-my-repo
8
+ - Dllama
9
+ - Infero
10
+ base_model: meta-llama/Meta-Llama-3-8B
11
+ datasets:
12
+ - cognitivecomputations/Dolphin-2.9
13
+ - teknium/OpenHermes-2.5
14
+ - m-a-p/CodeFeedback-Filtered-Instruction
15
+ - cognitivecomputations/dolphin-coder
16
+ - cognitivecomputations/samantha-data
17
+ - microsoft/orca-math-word-problems-200k
18
+ - Locutusque/function-calling-chatml
19
+ - internlm/Agent-FLAN
20
+ model-index:
21
+ - name: out
22
+ results: []
23
+ ---
24
 
25
  # tinybiggames/dolphin-2.9.1-llama-3-8b-Q4_K_M-GGUF
26
  This model was converted to GGUF format from [`cognitivecomputations/dolphin-2.9.1-llama-3-8b`](https://huggingface.co/cognitivecomputations/dolphin-2.9.1-llama-3-8b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
27
  Refer to the [original model card](https://huggingface.co/cognitivecomputations/dolphin-2.9.1-llama-3-8b) for more details on the model.
28
+ ## Use with tinyBigGAMES's Local LLM Inference Libraries
29
+
30
+ Add to **config.json**
31
+
32
+ ```Json
33
+ {
34
+ "filename": "dolphin-2.9.1-llama-3-8b.Q4_K_M.gguf",
35
+ "name": "dolphin-llama3:8B:4_K_M",
36
+ "max_context": 8000,
37
+ "template": "<|im_start|>%s\\n%s<|im_end|>\\n",
38
+ "template_end": "<|im_start|>assistant",
39
+ "stop": [
40
+ "<|im_start|>",
41
+ "<|im_end|>",
42
+ "assistant"
43
+ ]
44
+ }
45
+ ```