Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
TheBloke
/
StableBeluga2-70B-GPTQ
like
90
Text Generation
Transformers
Safetensors
conceptofmind/cot_submix_original
conceptofmind/flan2021_submix_original
conceptofmind/t0_submix_original
conceptofmind/niv2_submix_original
English
llama
text-generation-inference
4-bit precision
gptq
arxiv:
2307.09288
arxiv:
2306.02707
License:
llama2
Model card
Files
Files and versions
Community
14
Train
Deploy
Use this model
refs/pr/4
StableBeluga2-70B-GPTQ
2 contributors
History:
47 commits
mber
fix documentation for loading the model, since the fused attention module doesnt work here either.
af510cc
11 months ago
.gitattributes
1.52 kB
initial commit
11 months ago
LICENSE.txt
7.02 kB
Initial GPTQ model commit
11 months ago
Notice
112 Bytes
Initial GPTQ model commit
11 months ago
README.md
15.3 kB
fix documentation for loading the model, since the fused attention module doesnt work here either.
11 months ago
USE_POLICY.md
4.77 kB
Initial GPTQ model commit
11 months ago
config.json
679 Bytes
Initial GPTQ model commit
11 months ago
generation_config.json
137 Bytes
Initial GPTQ model commit
11 months ago
gptq_model-4bit--1g.safetensors
35.3 GB
LFS
Initial GPTQ model commit
11 months ago
quantize_config.json
183 Bytes
Initial GPTQ model commit
11 months ago
special_tokens_map.json
411 Bytes
Initial GPTQ model commit
11 months ago
tokenizer.json
1.84 MB
Initial GPTQ model commit
11 months ago
tokenizer.model
500 kB
LFS
Initial GPTQ model commit
11 months ago
tokenizer_config.json
649 Bytes
Initial GPTQ model commit
11 months ago