tp
Onix22
ยท
AI & ML interests
None yet
Recent Activity
new activity
18 days ago
nbeerbower/Dumpling-Qwen2.5-VL-7B:What it is made for?
new activity
27 days ago
huihui-ai/Qwen2-VL-7B-Instruct-abliterated:Request gguf
Organizations
None yet
Onix22's activity
you should make 4 or 4.25 bit quantization to make it usable on 24GB cards
#1 opened about 16 hours ago
by
Onix22
What it is made for?
3
#1 opened 19 days ago
by
Onix22
Request gguf
2
#1 opened 5 months ago
by
ayishfeng

So where is the model I would like to try it?
2
#1 opened about 1 month ago
by
Onix22
This model breaks laws of physic.
2
#4 opened about 1 year ago
by
Onix22
New activity in
TheBloke/WizardLM-Uncensored-SuperCOT-StoryTelling-30B-SuperHOT-8K-GPTQ
over 1 year ago
Does it work for anyone?
3
#1 opened over 1 year ago
by
Onix22
Running out of memory with 12GB of VRAM on 3080TI
3
#39 opened almost 2 years ago
by
faaaaaaaaaaaa
If 4-bit = 8GB model would 8-bit = 16GB? And how much better would it be?
2
#40 opened almost 2 years ago
by
Wildstar50
30B 3bit seems pretty sweet by the official evaluation
2
#1 opened almost 2 years ago
by
Yhyu13
update is needed
2
#5 opened almost 2 years ago
by
Onix22
3bit version
1
#1 opened almost 2 years ago
by
Onix22