Update README.md
Browse files
README.md
CHANGED
@@ -22,6 +22,8 @@ This model was fine-tuned with a fork of FastChat, and therefore uses the standa
|
|
22 |
USER: [prompt] ASSISTANT:
|
23 |
```
|
24 |
|
|
|
|
|
25 |
The most important bit, to me, is the context obedient question answering support, without extensive prompt engineering.
|
26 |
|
27 |
### Usage
|
|
|
22 |
USER: [prompt] ASSISTANT:
|
23 |
```
|
24 |
|
25 |
+
*__NOTE: an earlier version claimed context length of 4096 - this did not work! I modified the code to train with with 4096, and several instructions are beyond 2048. I tested a few prompts beyond 2048, and they seem to produce fairly coherent responses with increased context length for a couple hundred tokens beyond 2048, but I did not properly test up to 4096. As it turns out, it would appear without a massive fine-tune of the base model on a larger context window, this won't work. Sorry!__*
|
26 |
+
|
27 |
The most important bit, to me, is the context obedient question answering support, without extensive prompt engineering.
|
28 |
|
29 |
### Usage
|