Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
astronomer
/
Llama-3-8B-GPTQ-4-Bit
like
6
Follow
Astronomer
5
Text Generation
Transformers
Safetensors
wikitext
llama
llama-3
facebook
meta
astronomer
gptq
pretrained
quantized
finetuned
Inference Endpoints
text-generation-inference
4-bit precision
arxiv:
2210.17323
License:
llama-3
Model card
Files
Files and versions
Community
1
Train
Deploy
Use this model
main
Llama-3-8B-GPTQ-4-Bit
1 contributor
History:
9 commits
davidxmle
Update README.md
26ed101
verified
7 months ago
.gitattributes
Safe
1.52 kB
initial commit
7 months ago
LICENSE.txt
Safe
7.8 kB
Upload Llama 3 license, use policy and modified generation_config
7 months ago
README.md
Safe
6.32 kB
Update README.md
7 months ago
USE_POLICY.md
Safe
4.7 kB
Upload Llama 3 license, use policy and modified generation_config
7 months ago
config.json
Safe
1.01 kB
Upload folder using huggingface_hub
7 months ago
generation_config.json
Safe
136 Bytes
Upload Llama 3 license, use policy and modified generation_config
7 months ago
model.safetensors
Safe
5.74 GB
LFS
Rename gptq_model-4bit-128g.safetensors to model.safetensors
7 months ago
quantize_config.json
Safe
264 Bytes
Upload folder using huggingface_hub
7 months ago
special_tokens_map.json
Safe
301 Bytes
Upload folder using huggingface_hub
7 months ago
tokenizer.json
Safe
9.08 MB
Upload folder using huggingface_hub
7 months ago
tokenizer_config.json
Safe
50.6 kB
Upload folder using huggingface_hub
7 months ago