dm
prudant
AI & ML interests
Tech enthusiast, avid AI learner, and perpetual seeker of new knowledge.
Organizations
prudant's activity
serving with open ai protocol
6
#7 opened 2 months ago
by
prudant
just for curiosity
9
#1 opened about 2 months ago
by
prudant
License
3
#4 opened 2 months ago
by
mrfakename
Quantization request
5
#1 opened about 2 months ago
by
dillfrescott
AWQ quants
#1 opened about 2 months ago
by
prudant
How can I finetune this for domain especific applications?
7
#1 opened about 2 months ago
by
KaiKapioka
bigger Hermes 2 Pro God
2
#6 opened about 2 months ago
by
prudant
awq quants
18
#2 opened 2 months ago
by
prudant
Template
2
#1 opened 2 months ago
by
cyrilzakka
Do you have a quantized version of the model that works with sentence_transformers?
2
#5 opened 3 months ago
by
sungkim
Problem with Text Embeddings Inference
4
#4 opened 3 months ago
by
prudant
may be little bugs
3
#14 opened 3 months ago
by
prudant
serving the model
2
#13 opened 3 months ago
by
prudant
Optimize inference speed
5
#9 opened 3 months ago
by
CoolWP
Any plans to make model for coding tasks?
4
#5 opened 3 months ago
by
juskazan
OOMS on 8 GB GPU, is it normal?
3
#2 opened 3 months ago
by
tanimazsin130
Something is wrong with chatml tokens
12
#2 opened 4 months ago
by
kurnevsky
Context lenght?
#10 opened 3 months ago
by
prudant
template format
4
#3 opened 4 months ago
by
prudant
quant versions
4
#2 opened 5 months ago
by
prudant
YiTokenizer does not exist or is not currently imported.
9
#1 opened 6 months ago
by
iChrist
Context length
#1 opened 5 months ago
by
prudant
spanish model
4
#16 opened 6 months ago
by
prudant
Context Length and Max New Tokens
3
#1 opened 6 months ago
by
Shivam098
Quantizations for llama.cpp
4
#23 opened 9 months ago
by
rozek
Quantized version of Mistral 7B (4bit or 8 bit)
3
#18 opened 8 months ago
by
ianuvrat
What hardware do I need for reasonable performance?
4
#3 opened 11 months ago
by
TS0001