Help! Hope to get an inference configuration that can run on multiple GPUs.
#25 opened about 1 month ago
by
Lokis
Praise and Criticism
8
#23 opened 2 months ago
by
ChuckMcSneed
License for tokenizer
#22 opened 2 months ago
by
marksverdhei
Does the "Average Generation Length" in the press release mean the average number of output tokens?
#20 opened 2 months ago
by
yumemio
Miqu / Mistral Medium f16 / bf16 weights
#19 opened 2 months ago
by
Nexesenex
不知道下载哪些内容
1
#18 opened 2 months ago
by
qcnace
Change rope scaling to match max embedding size
#16 opened 3 months ago
by
Blackroot
[AUTOMATED] Model Memory Requirements
#15 opened 3 months ago
by
model-sizer-bot
Model load error
2
#14 opened 3 months ago
by
caisarl76
Old Mistral Large Not Released + No Base Model present.
#13 opened 3 months ago
by
User8213
No chat template
6
#12 opened 3 months ago
by
zyddnys
consolidated vs model safetensors - what's the difference?
12
#9 opened 3 months ago
by
jukofyork
Are we gonna get the base model for finetuning?
1
#8 opened 3 months ago
by
rombodawg
GGUF quants pl0x
1
#5 opened 3 months ago
by
AIGUYCONTENT
Is this "large" or "large2"?
6
#4 opened 3 months ago
by
ZeroWw
Le baiser du chef
#2 opened 3 months ago
by
nanowell