AI providers & models supported by GPT for Work
GPT for Work supports models from Anthropic, Azure, DeepSeek, Google, Mistral, OpenAI, OpenRouter, Perplexity, and xAI. GPT for Work also supports open-source models through Ollama and any OpenAI-compatible API endpoint. The tables below show which models you can use with and without an API key, in which GPT for Work add-ons, and at what price.
Use the table filters to see which models support reasoning, vision, and web search.
Models you can use without an API key
You can use models from OpenAI, Google, Anthropic, and Perplexity.
Models that support prompt caching get a 75% discount on cached input tokens.
Models you can use with an API key
You can use models from OpenAI, Perplexity, Google, Anthropic, OpenRouter, DeepSeek, Mistral, Azure, xAI, and open-source models through Ollama and any OpenAI-compatible API endpoint.
You pay the API cost directly to the AI provider.
Models available through dedicated endpoints (Azure, Ollama, other local servers and cloud-based platforms) are currently free for personal use. Contact us to use GPT for Work with dedicated endpoints for professional use.
Notes
Legal considerations
Ensure the AI models you use comply with your local laws and suit your needs.
DeepSeek should not be used in regions where its use is prohibited. When its use is allowed, avoid sharing personal, confidential, or sensitive information.
Reasoning models
Reasoning models are trained to think before they answer, producing an internal chain of thought before responding to a prompt. Reasoning models generate two types of tokens:
Completion tokens make up the model's response.
Reasoning tokens make up the model's internal chain of thought.
You are billed for both types of tokens.
Vision models
Vision models can process images as inputs. The Analyze images (Vision) bulk AI tool and the GPT_VISION function use such models.
Image inputs are measured and charged in tokens, just like text inputs. How images are converted to text tokens depends on the model. You can find more information about the conversion in the OpenAI documentation and Anthropic documentation.
Web search models
Web search models can gather the latest information from the web and use it as context when generating responses. The larger the context, the more information a model can retrieve from each web source, producing richer and more detailed responses.
The total cost of using a web search model is calculated differently depending on whether you use an API key:
With an API key: You only pay the token cost (per 1M tokens).
Without an API key: You pay the token cost (per 1M tokens) and a separate search cost (per 1,000 searches). The search cost varies based on the context size as shown below.
Provider Model Low Medium High Google
gemini-2.5-flash
$35.00 (fixed context size)
Perplexity
sonar
$5.00
$8.00
$12.00
Perplexity
sonar-pro
$6.00
$10.00
$14.00
Perplexity
sonar-reasoning
$5.00
$8.00
$12.00
Custom endpoints
You can use any OpenAI-compatible API endpoint with GPT for Work. You can connect to two main types of services:
Cloud-based LLM platforms provide access to models over the internet with no software installation or setup required on your part. Popular examples include Anyscale, Fireworks AI, and Together AI. The available models vary from platform to platform.
Local LLM servers run on a local machine, such as your own computer or another computer on a local network. Popular examples include LM Studio, LocalAI, and Open WebUI. The available models depend on what's installed on the server you're using.
What's next
The tables on this page were created with Awesome Table Apps.