cicdatopea
cicdatopea
AI & ML interests
None yet
Recent Activity
new activity
about 18 hours ago
OPEA/Qwen2-VL-72B-Instruct-int2-sym-inc:Qwen2.5-VL-72B-Instruct quant?
new activity
8 days ago
OPEA/DeepSeek-R1-int2-gptq-sym-inc:Could we get a int3 version of gptq please?
Organizations
cicdatopea's activity
Qwen2.5-VL-72B-Instruct quant?
1
#1 opened 4 days ago
by
samgreen

VLLM 0.7.2 can start the model normally, but there is no output when simulating a request using Curl, it blocks!
3
#2 opened about 2 months ago
by
JZMALi
Could we get a int3 version of gptq please?
1
#1 opened 8 days ago
by
davidsyoung
Qwen-32B overflow issue
8
#1 opened 18 days ago
by
cicdatopea

how to run this model
4
#1 opened 11 days ago
by
cicdatopea

without licence
1
#2 opened 15 days ago
by
Futureli
how to inference this model?
1
#1 opened 18 days ago
by
xiximayou
so consider build a model for GPU?
1
#1 opened 19 days ago
by
kq

Your quants are not listed in the base model
2
#2 opened about 1 month ago
by
dazipe
sglang inference issue
7
#1 opened about 2 months ago
by
su400
Start on cpu with vllm.
1
#1 opened 2 months ago
by
kuliev-vitaly
“a larger accuracy drop in Chinese tasks"? how much exectaly?
1
#1 opened 3 months ago
by
chuangzhidian
A bug when running the demo inference on GPU
1
#5 opened 3 months ago
by
HuggingLianWang
vllm
23
#4 opened 3 months ago
by
NikolaSigmoid
Base model please!
2
#2 opened 3 months ago
by
deltanym

alternative serving framework
2
#1 opened 3 months ago
by
erichartford

Update README.md
#1 opened 3 months ago
by
n1ck-guo