请问chatglm3-6b支持Flash Attention 2.0吗
#51
by
chentao111
- opened
ValueError: ChatGLMForConditionalGeneration does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co/model/chatglm3-6b/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new
我们已经在最新的开源模型 glm-4-9b-chat中实现,欢迎使用
zRzRzRzRzRzRzR
changed discussion status to
closed