Edit model card

gguf versions of OpenLLaMa 3B

Newer quantizations

There are now more quantization types in llama.cpp, some lower than 4 bits. Currently these are not supported, maybe because some weights have shapes that don't divide by 256.

Perplexity on wiki.test.406

Coming soon...

Downloads last month
249
GGUF
Model size
3.43B params
Architecture
llama
Unable to determine this model's library. Check the docs .