nvhf commited on
Commit
1a64292
1 Parent(s): 0fd7cd0

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +75 -0
README.md ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: humarin/chatgpt_paraphraser_on_T5_base
3
+ datasets:
4
+ - humarin/chatgpt-paraphrases
5
+ language:
6
+ - en
7
+ library_name: transformers
8
+ license: openrail
9
+ pipeline_tag: text2text-generation
10
+ tags:
11
+ - llama-cpp
12
+ - gguf-my-repo
13
+ inference:
14
+ parameters:
15
+ num_beams: 5
16
+ num_beam_groups: 5
17
+ num_return_sequences: 5
18
+ repetition_penalty: 10.01
19
+ diversity_penalty: 3.01
20
+ no_repeat_ngram_size: 2
21
+ temperature: 0.7
22
+ max_length: 128
23
+ widget:
24
+ - text: What are the best places to see in New York?
25
+ example_title: New York tourist attractions
26
+ - text: When should I go to the doctor?
27
+ example_title: Doctor's time
28
+ - text: Rammstein's album Mutter was recorded in the south of France in May and June
29
+ 2000, and mixed in Stockholm in October of that year.
30
+ example_title: Rammstein's album Mutter
31
+ ---
32
+
33
+ # nvhf/chatgpt_paraphraser_on_T5_base-Q6_K-GGUF
34
+ This model was converted to GGUF format from [`humarin/chatgpt_paraphraser_on_T5_base`](https://huggingface.co/humarin/chatgpt_paraphraser_on_T5_base) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
35
+ Refer to the [original model card](https://huggingface.co/humarin/chatgpt_paraphraser_on_T5_base) for more details on the model.
36
+
37
+ ## Use with llama.cpp
38
+ Install llama.cpp through brew (works on Mac and Linux)
39
+
40
+ ```bash
41
+ brew install llama.cpp
42
+
43
+ ```
44
+ Invoke the llama.cpp server or the CLI.
45
+
46
+ ### CLI:
47
+ ```bash
48
+ llama-cli --hf-repo nvhf/chatgpt_paraphraser_on_T5_base-Q6_K-GGUF --hf-file chatgpt_paraphraser_on_t5_base-q6_k.gguf -p "The meaning to life and the universe is"
49
+ ```
50
+
51
+ ### Server:
52
+ ```bash
53
+ llama-server --hf-repo nvhf/chatgpt_paraphraser_on_T5_base-Q6_K-GGUF --hf-file chatgpt_paraphraser_on_t5_base-q6_k.gguf -c 2048
54
+ ```
55
+
56
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
57
+
58
+ Step 1: Clone llama.cpp from GitHub.
59
+ ```
60
+ git clone https://github.com/ggerganov/llama.cpp
61
+ ```
62
+
63
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
64
+ ```
65
+ cd llama.cpp && LLAMA_CURL=1 make
66
+ ```
67
+
68
+ Step 3: Run inference through the main binary.
69
+ ```
70
+ ./llama-cli --hf-repo nvhf/chatgpt_paraphraser_on_T5_base-Q6_K-GGUF --hf-file chatgpt_paraphraser_on_t5_base-q6_k.gguf -p "The meaning to life and the universe is"
71
+ ```
72
+ or
73
+ ```
74
+ ./llama-server --hf-repo nvhf/chatgpt_paraphraser_on_T5_base-Q6_K-GGUF --hf-file chatgpt_paraphraser_on_t5_base-q6_k.gguf -c 2048
75
+ ```