Optimum for Intel® Gaudi® AI Accelerator
Optimum for Intel Gaudi AI accelerator is the interface between Hugging Face libraries (Transformers, Diffusers, Accelerate,…) and Intel Gaudi AI Accelerators (HPUs). It provides a set of tools that enable easy model loading, training and inference on single- and multi-HPU settings for various downstream tasks as shown in the table below.
Learn the basics and become familiar with training transformers on HPUs with 🤗 Optimum. Start here if you are using 🤗 Optimum for Intel Gaudi for the first time!
Practical guides to help you achieve a specific goal. Take a look at these guides to learn how to use 🤗 Optimum for Intel Gaudi to solve real-world problems.
The Intel Gaudi AI accelerator family currently includes three product generations: Intel Gaudi 1, Intel Gaudi 2, and Intel Gaudi 3. Each server is equipped with 8 devices, known as Habana Processing Units (HPUs), providing 128GB of memory on Gaudi 3, 96GB on Gaudi 2, and 32GB on the first-gen Gaudi. For more details on the underlying hardware architecture, check out the Gaudi Architecture Overview. Optimum for Intel Gaudi library is fully compatible with all three generations of Gaudi accelerators.
For in-depth examples of running workloads on Gaudi, explore the following blog posts:
- Benchmarking Intel Gaudi 2 with NVIDIA A100 GPUs
- Accelerating Vision-Language Models: BridgeTower on Habana Gaudi2
The following model architectures, tasks and device distributions have been validated for Optimum for Intel Gaudi:
In the tables below, ✅ means single-card, multi-card and DeepSpeed have all been validated.
- Transformers:
- Diffusers
Architecture | Training | Inference | Tasks |
---|---|---|---|
Stable Diffusion | |||
Stable Diffusion XL | |||
Stable Diffusion Depth2img | |||
LDM3D | |||
FLUX.1 | |||
Text to Video |
- PyTorch Image Models/TIMM:
Architecture | Training | Inference | Tasks |
---|---|---|---|
FastViT |
- TRL:
Architecture | Training | Inference | Tasks |
---|---|---|---|
Llama 2 | ✅ | ||
Llama 2 | ✅ | ||
Stable Diffusion | ✅ |
Other models and tasks supported by the 🤗 Transformers and 🤗 Diffusers library may also work. You can refer to this section for using them with 🤗 Optimum for Intel Gaudi. In addition, this page explains how to modify any example from the 🤗 Transformers library to make it work with 🤗 Optimum for Intel Gaudi.
< > Update on GitHub