TheBloke commited on
Commit
09ff06c
1 Parent(s): 74e45aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -0
README.md CHANGED
@@ -21,6 +21,10 @@ license: other
21
 
22
  These files are GGML format model files for [Gorilla LLM's Gorilla 7B](https://huggingface.co/gorilla-llm/gorilla-7b-hf-delta-v0).
23
 
 
 
 
 
24
  GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as:
25
  * [text-generation-webui](https://github.com/oobabooga/text-generation-webui)
26
  * [KoboldCpp](https://github.com/LostRuins/koboldcpp)
@@ -34,6 +38,13 @@ GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/gger
34
  * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/gorilla-7B-GGML)
35
  * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/gorilla-7B-fp16)
36
 
 
 
 
 
 
 
 
37
  <!-- compatibility_ggml start -->
38
  ## Compatibility
39
 
 
21
 
22
  These files are GGML format model files for [Gorilla LLM's Gorilla 7B](https://huggingface.co/gorilla-llm/gorilla-7b-hf-delta-v0).
23
 
24
+ **NOTE**: This is not a regular LLM. It is designed to allow LLMs to use tools by invoking APIs.
25
+
26
+ "Gorilla enables LLMs to use tools by invoking APIs. Given a natural language query, Gorilla can write a semantically- and syntactically- correct API to invoke. With Gorilla, we are the first to demonstrate how to use LLMs to invoke 1,600+ (and growing) API calls accurately while reducing hallucination. "
27
+
28
  GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/ggerganov/llama.cpp) and libraries and UIs which support this format, such as:
29
  * [text-generation-webui](https://github.com/oobabooga/text-generation-webui)
30
  * [KoboldCpp](https://github.com/LostRuins/koboldcpp)
 
38
  * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/gorilla-7B-GGML)
39
  * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/TheBloke/gorilla-7B-fp16)
40
 
41
+ ## Prompt template
42
+
43
+ ```
44
+ ###USER: find me an API to generate cute cat images
45
+ ###ASSISTANT:
46
+ ```
47
+
48
  <!-- compatibility_ggml start -->
49
  ## Compatibility
50