Inference Providers documentation

Hub Integration

Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

Hub Integration

Inference Providers is tightly integrated with the Hugging Face Hub. No matter which provider you use, the usage and billing will be centralized in your Hugging Face account.

Model search

When listing models on the Hub, you can filter to select models deployed on the inference provider of your choice. For example, to list all models deployed on Fireworks AI infra: https://huggingface.co/models?inference_provider=fireworks-ai.

It is also possible to select all or multiple providers and filter their available models: https://huggingface.co/models?inference_provider=all.

Features using Inference Providers

Several Hugging Face features utilize Inference Providers and count towards your monthly credits. The included monthly credits for PRO and Enterprise should cover moderate usage of these features for most users.

Inference Widgets

Interactive widgets available on model pages (e.g. deepseek-ai/DeepSeek-V3-0324). This is the entry point to quickly test a model on the Hub.

Inference Playground

A comprehensive chat interface supporting various models and providers available at https://huggingface.co/playground.

Data Studio AI

Converts text to SQL queries on dataset pages (e.g. open-r1/codeforces-cots).

User Settings

In your user account settings, you are able to:

  • set your own API keys for the providers you’ve signed up with. If you don’t, your requests will be billed on your HF account. More details in the billing section.
  • order providers by preference. This applies to the widget and code snippets in the model pages.
< > Update on GitHub