Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
neuralmagic
's Collections
FP8 LLMs for vLLM
Sparse Foundational Llama 2 Models
Compressed LLMs for nm-vllm
Compression Papers
DeepSparse Sparse LLMs
Sparse Finetuning MPT
FP8 LLMs for vLLM
updated
about 24 hours ago
Accurate FP8 quantized models by Neural Magic, ready for use with vLLM!
Upvote
14
+4
neuralmagic/Meta-Llama-3-8B-Instruct-FP8
Text Generation
•
Updated
10 days ago
•
742
•
3
neuralmagic/Meta-Llama-3-8B-Instruct-FP8-KV
Text Generation
•
Updated
about 24 hours ago
•
3.59k
•
1
neuralmagic/Meta-Llama-3-70B-Instruct-FP8
Text Generation
•
Updated
10 days ago
•
129
neuralmagic/Mixtral-8x7B-Instruct-v0.1-FP8
Text Generation
•
Updated
8 days ago
•
88
•
1
neuralmagic/Mixtral-8x22B-Instruct-v0.1-FP8
Text Generation
•
Updated
1 day ago
•
16
neuralmagic/Qwen2-72B-Instruct-FP8
Text Generation
•
Updated
9 days ago
•
26
•
3
neuralmagic/Qwen2-7B-Instruct-FP8
Text Generation
•
Updated
6 days ago
•
37
neuralmagic/Qwen2-1.5B-Instruct-FP8
Text Generation
•
Updated
6 days ago
neuralmagic/Qwen2-0.5B-Instruct-FP8
Text Generation
•
Updated
6 days ago
•
14
neuralmagic/Mixtral-8x22B-Instruct-v0.1-FP8-dynamic
Text Generation
•
Updated
8 days ago
•
11
Upvote
14
+10
Share collection
View history
Collection guide
Browse collections