Text2Text Generation
Transformers
PyTorch
Safetensors
t5
dialog
text-generation-inference
Inference Endpoints
jianguozhang commited on
Commit
fb42602
1 Parent(s): 19d2bf0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -110,7 +110,7 @@ alt="drawing" width="700"/>
110
  - For sampled FLAN data:
111
  - We follow their original data format, i.e., we did not set special tokens to separate in-context learning examples.
112
  - In summary:
113
- - We recommend you use our format and add our special tokens (such as `<USER>` and `<SYSTEM>` ) to get better performance. However, you may not necessary need to exactly follow our format if you do observe random behavios.
114
  - We found that T5 model series such as Flan-t5 and DialogStudio-T5 may generate repetitive tokens during inference. If you find such repetition issues, you can set the `repetition_penalty` in model.generate(), such as 1.5, to mitigate them. Note that `repetition_penalty=1.0` by default.
115
  # Usage
116
 
 
110
  - For sampled FLAN data:
111
  - We follow their original data format, i.e., we did not set special tokens to separate in-context learning examples.
112
  - In summary:
113
+ - We recommend you use our format and add our special tokens (such as `<USER>` and `<SYSTEM>` ) to get better performance. However, you may not necessary need to exactly follow our format if you do not observe random behavios.
114
  - We found that T5 model series such as Flan-t5 and DialogStudio-T5 may generate repetitive tokens during inference. If you find such repetition issues, you can set the `repetition_penalty` in model.generate(), such as 1.5, to mitigate them. Note that `repetition_penalty=1.0` by default.
115
  # Usage
116