Angel367 commited on
Commit
1f20914
1 Parent(s): 59def15

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +76 -0
README.md ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: 01-ai/Yi-6B-200K
3
+ license: apache-2.0
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - llama-cpp
7
+ - gguf-my-repo
8
+ widget:
9
+ - example_title: Yi-34B-Chat
10
+ text: hi
11
+ output:
12
+ text: ' Hello! How can I assist you today?'
13
+ - example_title: Yi-34B
14
+ text: There's a place where time stands still. A place of breath taking wonder,
15
+ but also
16
+ output:
17
+ text: ' an eerie sense that something is just not right…
18
+
19
+ Between the two worlds lies The Forgotten Kingdom - home to creatures long since
20
+ thought extinct and ancient magic so strong it defies belief! Only here can
21
+ you find what has been lost for centuries: An Elixir Of Life which will restore
22
+ youth and vitality if only those who seek its power are brave enough to face
23
+ up against all manner of dangers lurking in this mysterious land! But beware;
24
+ some say there may even exist powerful entities beyond our comprehension whose
25
+ intentions towards humanity remain unclear at best ---- they might want nothing
26
+ more than destruction itself rather then anything else from their quest after
27
+ immortality (and maybe someone should tell them about modern medicine)? In any
28
+ event though – one thing remains true regardless : whether or not success comes
29
+ easy depends entirely upon how much effort we put into conquering whatever challenges
30
+ lie ahead along with having faith deep down inside ourselves too ;) So let’s
31
+ get started now shall We?'
32
+ ---
33
+
34
+ # Angel367/Yi-6B-200K-Q4_K_M-GGUF
35
+ This model was converted to GGUF format from [`01-ai/Yi-6B-200K`](https://huggingface.co/01-ai/Yi-6B-200K) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
36
+ Refer to the [original model card](https://huggingface.co/01-ai/Yi-6B-200K) for more details on the model.
37
+
38
+ ## Use with llama.cpp
39
+ Install llama.cpp through brew (works on Mac and Linux)
40
+
41
+ ```bash
42
+ brew install llama.cpp
43
+
44
+ ```
45
+ Invoke the llama.cpp server or the CLI.
46
+
47
+ ### CLI:
48
+ ```bash
49
+ llama-cli --hf-repo Angel367/Yi-6B-200K-Q4_K_M-GGUF --hf-file yi-6b-200k-q4_k_m.gguf -p "The meaning to life and the universe is"
50
+ ```
51
+
52
+ ### Server:
53
+ ```bash
54
+ llama-server --hf-repo Angel367/Yi-6B-200K-Q4_K_M-GGUF --hf-file yi-6b-200k-q4_k_m.gguf -c 2048
55
+ ```
56
+
57
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
58
+
59
+ Step 1: Clone llama.cpp from GitHub.
60
+ ```
61
+ git clone https://github.com/ggerganov/llama.cpp
62
+ ```
63
+
64
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
65
+ ```
66
+ cd llama.cpp && LLAMA_CURL=1 make
67
+ ```
68
+
69
+ Step 3: Run inference through the main binary.
70
+ ```
71
+ ./llama-cli --hf-repo Angel367/Yi-6B-200K-Q4_K_M-GGUF --hf-file yi-6b-200k-q4_k_m.gguf -p "The meaning to life and the universe is"
72
+ ```
73
+ or
74
+ ```
75
+ ./llama-server --hf-repo Angel367/Yi-6B-200K-Q4_K_M-GGUF --hf-file yi-6b-200k-q4_k_m.gguf -c 2048
76
+ ```