请问chatglm3-6b支持Flash Attention 2.0吗

#51
by chentao111 - opened

ValueError: ChatGLMForConditionalGeneration does not support Flash Attention 2.0 yet. Please request to add support where the model is hosted, on its model hub page: https://huggingface.co/model/chatglm3-6b/discussions/new or in the Transformers GitHub repo: https://github.com/huggingface/transformers/issues/new

Knowledge Engineering Group (KEG) & Data Mining at Tsinghua University org

我们已经在最新的开源模型 glm-4-9b-chat中实现,欢迎使用

zRzRzRzRzRzRzR changed discussion status to closed

Sign up or log in to comment