Skip to main content

AI providers & models supported by GPT for Work

GPT for Work supports models from Anthropic, Azure, DeepSeek, Google, Mistral, OpenAI, OpenRouter, Perplexity, and xAI. GPT for Work also supports open-source models through Ollama and any OpenAI-compatible API endpoint. The tables below show which models you can use with and without an API key, in which GPT for Work add-ons, and at what price.

Models you can use without an API key

You can use models from OpenAI, Google, Anthropic, and Perplexity.

info

Reasoning models, vision models, and web search models typically cost much more than regular, text-only models. Learn more.

Models that support prompt caching get a 75% discount on cached input tokens.

Models you can use with an API key

You can use models from OpenAI, Perplexity, Google, Anthropic, OpenRouter, DeepSeek, Mistral, Azure, xAI, and open-source models through Ollama and any OpenAI-compatible API endpoint.

info

Reasoning models, vision models, and web search models typically cost much more than regular, text-only models. Learn more.

You pay the API cost directly to the AI provider.

Models available through dedicated endpoints (Azure, Ollama, other local servers and cloud-based platforms) are currently free for personal use. Contact us to use GPT for Work with dedicated endpoints for professional use.

Notes

Reasoning models

Reasoning models are trained to think before they answer, producing an internal chain of thought before responding to a prompt. Reasoning models generate two types of tokens:

  • Completion tokens make up the model's response.

  • Reasoning tokens make up the model's internal chain of thought.

You are billed for both types of tokens.

Vision models

Vision models can process images as input. The following features support vision models:

  • Custom prompt bulk AI tool

  • Prompt images (Vision) bulk AI tool

  • GPT_VISION function

Image inputs are measured and charged in tokens, just like text inputs. How images are converted to text tokens depends on the model. You can find more information about the conversion in the OpenAI documentation and Anthropic documentation.

Web search models

Web search models can gather the latest information from the web and use it as context when generating responses. The larger the context, the more information a model can retrieve from each web source, producing richer and more detailed responses.

The total cost of using a web search model is calculated differently depending on whether you use an API key:

  • With an API key: You only pay the token cost (per 1M tokens).

  • Without an API key: You pay the token cost (per 1M tokens) and a separate search cost (per 1,000 searches). The search cost varies based on the context size as shown below.

    ProviderModelLowMediumHigh

    Google

    gemini-2.5-flash

    $35.00 (fixed context size)

    Perplexity

    sonar

    $5.00

    $8.00

    $12.00

    Perplexity

    sonar-pro

    $6.00

    $10.00

    $14.00

    Perplexity

    sonar-reasoning

    $5.00

    $8.00

    $12.00

Custom endpoints

You can use any OpenAI-compatible API endpoint with GPT for Work. You can connect to two main types of services:

  • Cloud-based LLM platforms provide access to models over the internet with no software installation or setup required on your part. Popular examples include Anyscale, Fireworks AI, and Together AI. The available models vary from platform to platform.

  • Local LLM servers run on a local machine, such as your own computer or another computer on a local network. Popular examples include LM Studio, LocalAI, and Open WebUI. The available models depend on what's installed on the server you're using.

What's next

tip

The tables on this page were created with Awesome Table Apps.