Commit
•
f2e8793
1
Parent(s):
6a77dce
Update readme with links to rag/tool use prompting guide, copying over same changes from unquantized model (#2)
Browse files- Update readme with links to rag/tool use prompting guide, copying over same changes from unquantized model (39e50201e2e2d7d5c866648efedb25447bf850bd)
Co-authored-by: Patrick Lewis <patrick-s-h-lewis@users.noreply.huggingface.co>
README.md
CHANGED
@@ -77,9 +77,12 @@ Command-R has been specifically trained with conversational tool use capabilitie
|
|
77 |
|
78 |
Command-R’s tool use functionality takes a conversation as input (with an optional user-system preamble), along with a list of available tools. The model will then generate a json-formatted list of actions to execute on a subset of those tools. Command-R may use one of its supplied tools more than once.
|
79 |
|
80 |
-
The model has been trained to recognise a special `directly_answer` tool, which it uses to indicate that it doesn’t want to use any of its other tools.
|
|
|
81 |
|
82 |
-
Comprehensive documentation
|
|
|
|
|
83 |
|
84 |
<details>
|
85 |
<summary><b>Usage: Rendering Tool Use Prompts [CLICK TO EXPAND]</b> </summary>
|
@@ -201,14 +204,14 @@ Deviating from this prompt template may reduce performance, but we encourage exp
|
|
201 |
Command-R’s grounded generation behavior takes a conversation as input (with an optional user-supplied system preamble), along with a list of retrieved document snippets.
|
202 |
The document snippets should be chunks, rather than long documents, typically around 100-400 words per chunk. Document snippets consist of key-value pairs. The keys should be short descriptive strings, the values can be text or semi-structured.
|
203 |
|
204 |
-
|
205 |
Finally, it will then insert grounding spans into the answer. See below for an example. This is referred to as `accurate` grounded generation.
|
206 |
|
207 |
The model is trained with a number of other answering modes, which can be selected by prompt changes . A `fast` citation mode is supported in the tokenizer, which will directly generate an answer with grounding spans in it, without first writing the answer out in full. This sacrifices some grounding accuracy in favor of generating fewer tokens.
|
208 |
|
209 |
-
|
210 |
|
211 |
-
|
212 |
|
213 |
<details>
|
214 |
<summary> <b>Usage: Rendering Grounded Generation prompts [CLICK TO EXPAND]</b> </summary>
|
|
|
77 |
|
78 |
Command-R’s tool use functionality takes a conversation as input (with an optional user-system preamble), along with a list of available tools. The model will then generate a json-formatted list of actions to execute on a subset of those tools. Command-R may use one of its supplied tools more than once.
|
79 |
|
80 |
+
The model has been trained to recognise a special `directly_answer` tool, which it uses to indicate that it doesn’t want to use any of its other tools. The ability to abstain from calling a specific tool can be useful in a range of situations, such as greeting a user, or asking clarifying questions.
|
81 |
+
We recommend including the `directly_answer` tool, but it can be removed or renamed if required.
|
82 |
|
83 |
+
Comprehensive documentation for working with command-R's tool use prompt template can be found [here](https://docs.cohere.com/docs/prompting-command-r).
|
84 |
+
|
85 |
+
The code snippet below shows a minimal working example on how to render a prompt.
|
86 |
|
87 |
<details>
|
88 |
<summary><b>Usage: Rendering Tool Use Prompts [CLICK TO EXPAND]</b> </summary>
|
|
|
204 |
Command-R’s grounded generation behavior takes a conversation as input (with an optional user-supplied system preamble), along with a list of retrieved document snippets.
|
205 |
The document snippets should be chunks, rather than long documents, typically around 100-400 words per chunk. Document snippets consist of key-value pairs. The keys should be short descriptive strings, the values can be text or semi-structured.
|
206 |
|
207 |
+
Command-R’s grounded generation behavior takes a conversation as input (with an optional user-supplied system preamble, indicating task, context and desired output style), along with a list of retrieved document snippets.
|
208 |
Finally, it will then insert grounding spans into the answer. See below for an example. This is referred to as `accurate` grounded generation.
|
209 |
|
210 |
The model is trained with a number of other answering modes, which can be selected by prompt changes . A `fast` citation mode is supported in the tokenizer, which will directly generate an answer with grounding spans in it, without first writing the answer out in full. This sacrifices some grounding accuracy in favor of generating fewer tokens.
|
211 |
|
212 |
+
Comprehensive documentation for working with command-R's grounded generation prompt template can be found [here](https://docs.cohere.com/docs/prompting-command-r).
|
213 |
|
214 |
+
The code snippet below shows a minimal working example on how to render a prompt.
|
215 |
|
216 |
<details>
|
217 |
<summary> <b>Usage: Rendering Grounded Generation prompts [CLICK TO EXPAND]</b> </summary>
|