Inference Endpoints Container Types
When you create an Endpoint, you have a variety of options when selecting a container type.
Default
The default container type is the easiest way to deploy endpoints and is flexible thanks to custom Inference Handlers. The Hugging Face Inference Toolkit is now public at https://github.com/huggingface/huggingface-inference-toolkit.
Custom
Select a custom container type if you’d like to customize the image and include a custom container.
Text Embeddings Inference
Select the Text Embeddings Inference container type to gain all the benefits of TEI for your Endpoint. You’ll see this option in the UI if supported for that model.
Text Generation Inference
Select the Text Generation Inference container type to gain all the benefits of TGI for your Endpoint. You’ll see this option in the UI if supported for that model.
Text Generation Inference (INF2)
Select the Text Generation Inference Inferentia2 Neuron container type for models you’d like to deploy with TGI on an AWS Inferentia2 instance. You’ll see this option in the UI if supported for that model.
Text Generation Inference (TPU)
Select the Text Generation Inference TPU container type for models you’d like to deploy with TGI on a Google Cloud TPU instance. You’ll see this option in the UI if supported for that model.
NVIDIA NIM (no longer available in UI)
The NIM container type will no longer be officially supported for already existing Endpoints in Inference Endpoints beginning October 1st, 2024. Select the NIM container type for models supported by NVIDIA. You’ll see this option in the UI if supported for that model.
< > Update on GitHub