GGUF quants?

#2
by 6346y9uey - opened

As above

Owner

oh ok, I'll do IQ4_XS

llm_load_print_meta: model ftype      = IQ4_XS - 4.25 bpw
llm_load_print_meta: model params     = 121.89 B
llm_load_print_meta: model size       = 61.35 GiB (4.32 BPW) 
llm_load_print_meta: general.name     = New Volume
llm_load_print_meta: BOS token        = 128000 '<|begin_of_text|>'
llm_load_print_meta: EOS token        = 128001 '<|end_of_text|>'
llm_load_print_meta: LF token         = 128 'Ä'
llm_load_print_meta: EOT token        = 128009 '<|eot_id|>'
llm_load_tensors: ggml ctx size =    0.64 MiB

Sign up or log in to comment