Bug Report
#10 opened 8 months ago
by
kxzxvbk
How to solve Model inference speed is too slow?
#9 opened over 1 year ago
by
lanshan
求助AttributeError: 'GLMModel' object has no attribute 'stream_chat'
#8 opened over 1 year ago
by
rainstorm130
int4/8 quantization so that we can deploy on consumer-grid gpu card
1
#7 opened over 1 year ago
by
Yhyu13
Is this trained the same way as 6B model, and the only difference is the model size?
1
#6 opened over 1 year ago
by
huashiyiqike
Cannot find mask token in the input 的错误
#4 opened over 1 year ago
by
LZM2bobo
Support accelerate for GLM
#2 opened over 1 year ago
by
larrylawl
RuntimeError: expected self and mask to be on the same device, but got mask on cpu and self on cuda:0
4
#1 opened over 1 year ago
by
bupalinyu