Commit
f2e8793
1 Parent(s): 6a77dce

Update readme with links to rag/tool use prompting guide, copying over same changes from unquantized model (#2)

Browse files

- Update readme with links to rag/tool use prompting guide, copying over same changes from unquantized model (39e50201e2e2d7d5c866648efedb25447bf850bd)


Co-authored-by: Patrick Lewis <patrick-s-h-lewis@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +8 -5
README.md CHANGED
@@ -77,9 +77,12 @@ Command-R has been specifically trained with conversational tool use capabilitie
77
 
78
  Command-R’s tool use functionality takes a conversation as input (with an optional user-system preamble), along with a list of available tools. The model will then generate a json-formatted list of actions to execute on a subset of those tools. Command-R may use one of its supplied tools more than once.
79
 
80
- The model has been trained to recognise a special `directly_answer` tool, which it uses to indicate that it doesn’t want to use any of its other tools. We recommend including the `directly_answer` tool, but encourage experimentation.
 
81
 
82
- Comprehensive documentation and guides on prompting strategies for tool use will be provided shortly.
 
 
83
 
84
  <details>
85
  <summary><b>Usage: Rendering Tool Use Prompts [CLICK TO EXPAND]</b> </summary>
@@ -201,14 +204,14 @@ Deviating from this prompt template may reduce performance, but we encourage exp
201
  Command-R’s grounded generation behavior takes a conversation as input (with an optional user-supplied system preamble), along with a list of retrieved document snippets.
202
  The document snippets should be chunks, rather than long documents, typically around 100-400 words per chunk. Document snippets consist of key-value pairs. The keys should be short descriptive strings, the values can be text or semi-structured.
203
 
204
- By default, Command-R will generate grounded responses by first predicting which documents are relevant, then predicting which ones it will cite, then generating an answer.
205
  Finally, it will then insert grounding spans into the answer. See below for an example. This is referred to as `accurate` grounded generation.
206
 
207
  The model is trained with a number of other answering modes, which can be selected by prompt changes . A `fast` citation mode is supported in the tokenizer, which will directly generate an answer with grounding spans in it, without first writing the answer out in full. This sacrifices some grounding accuracy in favor of generating fewer tokens.
208
 
209
- The code snippet below shows a minimal working example on how to render a prompt, generate and parse a completion.
210
 
211
- Comprehensive documentation and guides on prompting strategies on grounded generation will be provided in follow-ups at a later stage.
212
 
213
  <details>
214
  <summary> <b>Usage: Rendering Grounded Generation prompts [CLICK TO EXPAND]</b> </summary>
 
77
 
78
  Command-R’s tool use functionality takes a conversation as input (with an optional user-system preamble), along with a list of available tools. The model will then generate a json-formatted list of actions to execute on a subset of those tools. Command-R may use one of its supplied tools more than once.
79
 
80
+ The model has been trained to recognise a special `directly_answer` tool, which it uses to indicate that it doesn’t want to use any of its other tools. The ability to abstain from calling a specific tool can be useful in a range of situations, such as greeting a user, or asking clarifying questions.
81
+ We recommend including the `directly_answer` tool, but it can be removed or renamed if required.
82
 
83
+ Comprehensive documentation for working with command-R's tool use prompt template can be found [here](https://docs.cohere.com/docs/prompting-command-r).
84
+
85
+ The code snippet below shows a minimal working example on how to render a prompt.
86
 
87
  <details>
88
  <summary><b>Usage: Rendering Tool Use Prompts [CLICK TO EXPAND]</b> </summary>
 
204
  Command-R’s grounded generation behavior takes a conversation as input (with an optional user-supplied system preamble), along with a list of retrieved document snippets.
205
  The document snippets should be chunks, rather than long documents, typically around 100-400 words per chunk. Document snippets consist of key-value pairs. The keys should be short descriptive strings, the values can be text or semi-structured.
206
 
207
+ Command-R’s grounded generation behavior takes a conversation as input (with an optional user-supplied system preamble, indicating task, context and desired output style), along with a list of retrieved document snippets.
208
  Finally, it will then insert grounding spans into the answer. See below for an example. This is referred to as `accurate` grounded generation.
209
 
210
  The model is trained with a number of other answering modes, which can be selected by prompt changes . A `fast` citation mode is supported in the tokenizer, which will directly generate an answer with grounding spans in it, without first writing the answer out in full. This sacrifices some grounding accuracy in favor of generating fewer tokens.
211
 
212
+ Comprehensive documentation for working with command-R's grounded generation prompt template can be found [here](https://docs.cohere.com/docs/prompting-command-r).
213
 
214
+ The code snippet below shows a minimal working example on how to render a prompt.
215
 
216
  <details>
217
  <summary> <b>Usage: Rendering Grounded Generation prompts [CLICK TO EXPAND]</b> </summary>