Sengxian commited on
Commit
45db462
1 Parent(s): b44f8e5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -1
app.py CHANGED
@@ -73,7 +73,8 @@ if __name__ == "__main__":
73
  """
74
  An Open Bilingual Pre-Trained Model. [Visit our github repo](https://github.com/THUDM/GLM-130B)
75
  GLM-130B uses two different mask tokens: `[MASK]` for short blank filling and `[gMASK]` for left-to-right long text generation. When the input does not contain any MASK token, `[gMASK]` will be automatically appended to the end of the text. We recommend that you use `[MASK]` to try text fill-in-the-blank to reduce wait time (ideally within seconds without queuing).
76
- \nNote: We suspect that there is a bug in the current FasterTransformer INT4 implementation that leads to gaps in the output compared to the FP16 model (e.g. more repititions), which we are troubleshooting, and the current model output is **for reference only**
 
77
  """)
78
 
79
  with gr.Row():
 
73
  """
74
  An Open Bilingual Pre-Trained Model. [Visit our github repo](https://github.com/THUDM/GLM-130B)
75
  GLM-130B uses two different mask tokens: `[MASK]` for short blank filling and `[gMASK]` for left-to-right long text generation. When the input does not contain any MASK token, `[gMASK]` will be automatically appended to the end of the text. We recommend that you use `[MASK]` to try text fill-in-the-blank to reduce wait time (ideally within seconds without queuing).
76
+
77
+ Note: We suspect that there is a bug in the current FasterTransformer INT4 implementation that leads to gaps in the output compared to the FP16 model (e.g. more repititions), which we are troubleshooting, and the current model output is **for reference only**
78
  """)
79
 
80
  with gr.Row():