Inference Providers documentation

🤗 Use Hugging Face Inference Providers with GitHub Copilot Chat in VS Code

Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

🤗 Use Hugging Face Inference Providers with GitHub Copilot Chat in VS Code

Demo

Use frontier open LLMs like Kimi K2, DeepSeek V3.1, GLM 4.5 and more in VS Code with GitHub Copilot Chat powered by Hugging Face Inference Providers 🔥

⚡ Quick start

  1. Install the HF Copilot Chat extension here.
  2. Open VS Code’s chat interface.
  3. Click the model picker and click “Manage Models…“.
  4. Select “Hugging Face” provider.
  5. Enter your Hugging Face Token. You can get one from your settings page.
  6. Choose the models you want to add to the model picker. 🥳

VS Code 1.104.0+ is required to install the HF Copilot Chat extension. If “Hugging Face” doesn’t appear in the Copilot provider list, update VS Code, then reload.

✨ Why use the Hugging Face provider in Copilot

  • Access SoTA open‑source LLMs with tool calling capabilities.
  • Single API to switch between multiple providers like Groq, Cerebras, Together AI, SambaNova, and more.
  • Built for high availability (across providers) and low latency.
  • Transparent pricing: what the provider charges is what you pay.

💡 The free Hugging Face user tier gives you a small amount of monthly inference credits to experiment. Upgrade to Hugging Face PRO or Team or Enterprise for $2 in monthly credits plus pay‑as‑you‑go access across all providers!

Check out the whole workflow in action in the video below:

< > Update on GitHub