Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
aws-neuron
/
optimum-neuron-cache
like
13
Follow
AWS Inferentia and Trainium
64
License:
apache-2.0
Model card
Files
Files and versions
Community
255
6b8c4b2
optimum-neuron-cache
/
inference-cache-config
8 contributors
History:
46 commits
dacorvo
HF staff
Update inference-cache-config/Llama3.1-70b.json
7b0370b
verified
2 months ago
Llama3.1-70b.json
289 Bytes
Update inference-cache-config/Llama3.1-70b.json
2 months ago
gpt2.json
398 Bytes
Add more gpt2 configurations
8 months ago
llama-variants.json
2.63 kB
Update inference-cache-config/llama-variants.json
5 months ago
llama.json
1.67 kB
Update inference-cache-config/llama.json
2 months ago
llama2-70b.json
287 Bytes
Create llama2-70b.json
5 months ago
llama3-70b.json
283 Bytes
Update inference-cache-config/llama3-70b.json
2 months ago
mistral-variants.json
3.29 kB
Remove SalesForce embedding model
8 months ago
mistral.json
1.8 kB
Update inference-cache-config/mistral.json
2 months ago
mixtral.json
583 Bytes
Update inference-cache-config/mixtral.json
2 months ago
stable-diffusion.json
1.91 kB
Update inference-cache-config/stable-diffusion.json
2 months ago