Skip to main content

How your balance is consumed

Your balance in GPT for Excel and GPT for Sheets decreases as you consume tokens (except for the GPT_WEB function, which is priced per 1000 executions).

Token consumption includes:

  • Input = Text or images sent to the model providers, including custom instructions, prompt templates that we use under the hood (which are very small)
  • Output = Results returned by the model providers

Input token cost is generally lower that output token cost and varies according to the model used. Use the cost estimator to estimate what your use case will cost.

Example of token consumption

Consider a GPT formula, translating some source text (1) from English (2) to Chinese (3). We provide some Custom instructions in GPT for Sheets sidebar (4). The output (5) consists of the translated text.

Here is how the various elements from the formula and sidebar break down into input tokens:

InputValue# tokens
TextIn the Pride Lands of Africa, a pride of lions rule over the kingdom from Pride Rock. King Mufasa and Queen Sarabi's newborn son, Simba, is presented to the gathering animals by Rafiki the mandrill, the kingdom's shaman and advisor. Mufasa's younger brother, Scar, covets the throne.70 tokens
Source languageEnglish1 token
Target languageChinese1 token
Custom instructionsTranslate the user text with utmost fidelity. Ensure direct and literal translation without adding, interpreting, or modifying content. Any additional comments, opinions, or deviations are explicitly disallowed.35 tokens
Prompt templatePrompt templates are sent along with your own input to describe the task to be performed by the AI and the expected output format. 19 tokens

The number of output tokens is only based on the translated text, displayed as the result of your formula:

OutputValue# tokens
Text在非洲的狮子王国中,一群狮子统治着自豪岩上的王国。国王姆法萨和王后萨拉比的新生儿儿子辛巴由国王国的神秘顾问和萨满——拉菲基(Rafiki)猴子猿(mandrill)呈现给了聚集在一起的动物们。姆法萨的弟弟斯卡恩觊觎王位。146 tokens
info

Unlike other functions, the GPT_MATCH function's output doesn't account as tokens.

Token count and languages

Token counts vary by language because the way text is split into tokens differs from one language to another.

On average (figures from this study):

  • English: 1 word ≈ 1.3 tokens
  • French: 1 word ≈ 2 tokens
  • German: 1 word ≈ 2.1 tokens
  • Spanish: 1 word ≈ 2.1 tokens
  • Chinese: 1 word ≈ 2.5 tokens
  • Russian: 1 word ≈ 3.3 tokens
  • Vietnamese: 1 word ≈ 3.3 tokens
  • Arabic: 1 word ≈ 4 tokens
  • Hindi: 1 word ≈ 6.4 tokens

Refer to the study for other languages.

These figures are presented for estimation purposes only and are not guaranteed.

Prompt template tokens

Prompt templates are sent along with your own input to describe the task to be performed by the AI and the expected output format.

  • For GPT for Word and GPT for Docs, prompt templates use approximately 100 tokens

  • For GPT for Sheets and GPT for Excel functions, prompt templates vary according to the function executed.

    To find the token count for each function prompt template, visit the cost estimator and select a function. The number of input tokens shown when the input fields are empty is the token count for the prompt template.

    The prompt template for GPT_TRANSLATE corresponds to 12 tokens