build: AutoGPTQ for meta-llama/Llama-2-7b-chat-hf: 4bits, gr128, desc_act=False
Browse files- gptq_model-4bit-128g.bin +3 -0
gptq_model-4bit-128g.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:981e70a261a93caeb561ab1ece8050b367a3dd07cfe65e052cf47fdb77439bd3
|
3 |
+
size 3896904828
|