qps100以上 推荐下显卡
#18 opened 11 days ago
by
duzhihua
Missing Finetuning instruction for bge-reranker-v2-m3 ?
1
#17 opened 12 days ago
by
jackkwok
Multi-GPU at FP16? Examples. Large memory allocations.
1
#16 opened 14 days ago
by
flash9001
How to make it run on GPU?
1
#15 opened 15 days ago
by
HarshalPa
Add Sentence Transformers config
#14 opened about 1 month ago
by
peakji
ONNX version
#13 opened about 2 months ago
by
Malithius
Anyway to 'drop' model to save GPU ram?
1
#12 opened about 2 months ago
by
rag-perplexity
cutoff score to consider for LLM call
4
#11 opened about 2 months ago
by
karthikfds
bf16 vs fp16
1
#10 opened about 2 months ago
by
Totole
Document length for v2-m3?
3
#9 opened about 2 months ago
by
rag-perplexity
请问该v2-m3最大支持多少token
1
#8 opened about 2 months ago
by
devillaws
有什么加速的方案吗?
1
#7 opened about 2 months ago
by
hanswang1973
corss-lingual reranking
2
#6 opened 2 months ago
by
victorkeke
支持在langchain框架下使用吗
1
#4 opened 2 months ago
by
Nicole828
Missing pytorch_model.bin file?
1
#3 opened 2 months ago
by
baobo5625
need onnx model
#1 opened 2 months ago
by
LowPower