Inference Providers documentation

Featherless AI

Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

Featherless AI

All supported Featherless AI models can be found here

Featherless AI is a serverless AI inference platform that offers access to thousands of open-source models.

Our goal is to make all AI models available for serverless inference. We provide inference via API to a continually expanding library of open-weight models.

Supported tasks

Chat Completion (LLM)

Find out more about Chat Completion (LLM) here.

import os
from huggingface_hub import InferenceClient

client = InferenceClient(
    provider="featherless-ai",
    api_key=os.environ["HF_TOKEN"],
)

completion = client.chat.completions.create(
    model="mistralai/Magistral-Small-2506",
    messages=[
        {
            "role": "user",
            "content": "What is the capital of France?"
        }
    ],
)

print(completion.choices[0].message)

Chat Completion (VLM)

Find out more about Chat Completion (VLM) here.

import os
from huggingface_hub import InferenceClient

client = InferenceClient(
    provider="featherless-ai",
    api_key=os.environ["HF_TOKEN"],
)

completion = client.chat.completions.create(
    model="CEIA-UFG/Gemma-3-Gaia-PT-BR-4b-it",
    messages=[
        {
            "role": "user",
            "content": [
                {
                    "type": "text",
                    "text": "Describe this image in one sentence."
                },
                {
                    "type": "image_url",
                    "image_url": {
                        "url": "https://cdn.britannica.com/61/93061-050-99147DCE/Statue-of-Liberty-Island-New-York-Bay.jpg"
                    }
                }
            ]
        }
    ],
)

print(completion.choices[0].message)

Text Generation

Find out more about Text Generation here.

import os
from huggingface_hub import InferenceClient

client = InferenceClient(
    provider="featherless-ai",
    api_key=os.environ["HF_TOKEN"],
)

completion = client.chat.completions.create(
    model="mistralai/Magistral-Small-2506",
    messages="\"Can you please let us know more details about your \"",
)

print(completion.choices[0].message)
< > Update on GitHub