gpt4all-j-ggml / gpt4all-j-f16.meta
LLukas22's picture
Upload new model file: 'gpt4all-j-f16.bin'
7386b97
raw
history blame
267 Bytes
{
"model": "GptJ",
"quantization": "F16",
"quantization_version": "Not_Quantized",
"container": "GGML",
"converter": "llm-rs",
"hash": "a507b3585a15317f8e1724d0d9aa8c1f0df437db32d0acc0b7d2330e9afcf71c",
"base_model": "nomic-ai/gpt4all-j"
}