Documentation

AI estimates and costs

Documentation

AI estimates and costs

Understand how AI estimates and costs are determined.

Our AI agent provides you with an estimate for using any AI provider/model on a task before spending any money.

To understand how estimates and costs can vary by AI provider/model, it's essential to talk about some technical basics about AI models.

Tokens, models and Mini Agent AI cost estimatorIllustration explaining token basics, input vs output token pricing and how different models tokenize content differently, all feeding into the Mini Agent AI cost estimator.Tokens, models and cost estimationDifferent providers and models count and price tokens differentlyTokens: the basic unitModels measure work in tokens, not words."To""ken""s"• Tokens can be shorter than a word.• Sometimes they are just characters or syllables.• Different models split the same text differently.Input vs output tokensModels charge differently for prompts and responses.Input tokens (prompt)e.g. $0.50 / 1M input tokensOutput tokens (response)e.g. $1.50 / 1M output tokensNewer models can be 5–10× more expensivethan older ones, even at the same provider.Same text, different countsEach model slices tokens its own way.Model A3000Model B3500Model C2700Counts can shift by provider and model,even on identical prompts and responses.Mini Agent · AI cost estimatorUses public pricing + token info to approximate real usage costs.Estimates can’t be perfect, but they stay as close as possible to actual provider billing.
  • AI models use tokens to measure their work. Tokens are generally smaller than a word, often the size of a syllable, but can also be a single character. Yes, it can get complicated.
  • AI models charge tokens for both input (i.e. the prompt you give it) and output (i.e. the response they give you). In addition, input and output tokens have different costs, not to mention tokens for recently released AI models can be 5 or 10 times more expensive than older AI models.
  • All AI models tokenize content differently, even if you use the same provider. So while one model could quantify a prompt and response as 3000 tokens, another model could just easily quantify an identical prompt and response as 3500 tokens or 2700 tokens.

Given all the potential variations between tokens associated with providers/models, the Mini Agent AI estimator attempts to use all publicly available information, on costs and tokens, to give you an estimation as close as possible to the actual cost of running an AI provider/model on your task.

AI agent settings you control that influence cost

Before triggering the AI agent to work on your task, you can adjust AI provider/model parameters in the 'AI Resolution Settings' window by clicking on the gear icon besides the 'Generate AI Resolution' button.

You can see the 'AI Resolution Settings' window pictured here. The most critical factors that influence cost by order of importance are:

  • AI provider/model: Newer and specialized models always cost more per token. If your task only requires basic knowledge, old and inexpensive models can produce results as good as newer and more expensive models.
  • AI model max tokens: You can tell the AI agent to limit a model's response to a set amount of tokens. Setting max tokens doesn't necessarily mean all these tokens will be consumed/charged for a response, since a model might very well respond with half or less tokens. However, considering some models can be very chatty, keeping this maxiumum amount of tokens in check, ensures shorter responses that translate into lower costs.
  • AI prompt size: The AI agent always uses a task's title and description to generate a prompt, as well as any file attachments associated with the task. You can reduce token consumption by optimizing both a task's title and description, as well as disabling file attachments in the bottom checkbox, so they're ignored for prompt generation.
AI task resolution button
aicosts
Explore by topic

Other docs that touch the same themes and keywords.