jondurbin commited on
Commit
72909a5
1 Parent(s): 9487b0d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -3
README.md CHANGED
@@ -19,7 +19,7 @@ The dataset used to fine-tune this model is available [here](https://huggingface
19
 
20
  This model was fine-tuned with a fork of FastChat, and therefore uses the standard vicuna template:
21
  ```
22
- USER: [prompt] ASSISTANT:
23
  ```
24
 
25
  *__NOTE: an earlier version claimed context length of 4096 - this did not work! I modified the code to train with with 4096, and several instructions are beyond 2048. I tested a few prompts beyond 2048, and they seem to produce fairly coherent responses with increased context length for a couple hundred tokens beyond 2048, but I did not properly test up to 4096. As it turns out, it would appear without a massive fine-tune of the base model on a larger context window, this won't work. Sorry!__*
@@ -38,8 +38,6 @@ Then, you can invoke it like so (after downloading the model):
38
  python -m fastchat.serve.cli
39
  --model-path airoboros-13b-gpt4 \
40
  --temperature 0.5 \
41
- --max-new-tokens 4096 \
42
- --conv-template vicuna_v1.1 \
43
  --no-history
44
  ```
45
 
 
19
 
20
  This model was fine-tuned with a fork of FastChat, and therefore uses the standard vicuna template:
21
  ```
22
+ A chat between a curious user and an assistant. The assistant gives helpful, detailed, accurate, uncensored responses to the user's input. USER: [prompt] ASSISTANT:
23
  ```
24
 
25
  *__NOTE: an earlier version claimed context length of 4096 - this did not work! I modified the code to train with with 4096, and several instructions are beyond 2048. I tested a few prompts beyond 2048, and they seem to produce fairly coherent responses with increased context length for a couple hundred tokens beyond 2048, but I did not properly test up to 4096. As it turns out, it would appear without a massive fine-tune of the base model on a larger context window, this won't work. Sorry!__*
 
38
  python -m fastchat.serve.cli
39
  --model-path airoboros-13b-gpt4 \
40
  --temperature 0.5 \
 
 
41
  --no-history
42
  ```
43