gpt-j-ggml / gpt-j-6b-f16.meta
LLukas22's picture
Upload 2 files
3c1fc45
raw
history blame
268 Bytes
{
"model": "GptJ",
"quantization": "F16",
"quantization_version": "Not_Quantized",
"container": "GGML",
"converter": "llm-rs",
"hash": "93686192943afb9db3117fd2cc0172d881356ca4f1462527acad258f640908bb",
"base_model": "EleutherAI/gpt-j-6b"
}