jianguozhang
commited on
Commit
•
fb42602
1
Parent(s):
19d2bf0
Update README.md
Browse files
README.md
CHANGED
@@ -110,7 +110,7 @@ alt="drawing" width="700"/>
|
|
110 |
- For sampled FLAN data:
|
111 |
- We follow their original data format, i.e., we did not set special tokens to separate in-context learning examples.
|
112 |
- In summary:
|
113 |
-
- We recommend you use our format and add our special tokens (such as `<USER>` and `<SYSTEM>` ) to get better performance. However, you may not necessary need to exactly follow our format if you do observe random behavios.
|
114 |
- We found that T5 model series such as Flan-t5 and DialogStudio-T5 may generate repetitive tokens during inference. If you find such repetition issues, you can set the `repetition_penalty` in model.generate(), such as 1.5, to mitigate them. Note that `repetition_penalty=1.0` by default.
|
115 |
# Usage
|
116 |
|
|
|
110 |
- For sampled FLAN data:
|
111 |
- We follow their original data format, i.e., we did not set special tokens to separate in-context learning examples.
|
112 |
- In summary:
|
113 |
+
- We recommend you use our format and add our special tokens (such as `<USER>` and `<SYSTEM>` ) to get better performance. However, you may not necessary need to exactly follow our format if you do not observe random behavios.
|
114 |
- We found that T5 model series such as Flan-t5 and DialogStudio-T5 may generate repetitive tokens during inference. If you find such repetition issues, you can set the `repetition_penalty` in model.generate(), such as 1.5, to mitigate them. Note that `repetition_penalty=1.0` by default.
|
115 |
# Usage
|
116 |
|