Overview
Cost estimates in Freeplay give you a clear view of how you are spending your API tokens. Freeplay breaks down costs by project, prompt template, and evaluation so you can trace your spend and identify where costs are highest. Freeplay also provides rough estimates of evaluation costs before you run them, helping you make informed decisions about your testing and evaluation strategy.Where to find cost estimates
You can view cost information in two places within Freeplay.Account settings —> Usage
The Usage tab in Account Settings is the primary source of truth for costs and cost estimates. Navigate to Account Settings and select the Usage tab. This page provides two key metrics:- App spend - estimated spend based on logged token counts to Freeplay. This reflects the cost of running your application’s LLM calls, not money spent in or through Freeplay itself.
- Spend via Freeplay - the total spend from activity within the Freeplay platform. This includes:
- Online evaluations and auto-categories
- Test runs
- Prompt playground usage
- AI features
Estimated evaluation costs
On the evaluations page, each evaluation displays an estimated cost. This estimate gives you a rough idea of how much it costs to run that evaluation over your completions. The estimate is calculated as follows:- Takes the average input and output tokens for the specific prompt template or agent (including input variables) from the last week
- Calculates the average cost based on those token counts
This is a rough estimate and is subject to change based on volume, sampling rate, and other factors.
How costs are calculated
Freeplay bases all cost estimates on token usage. Token costs are calculated using up-to-date pricing information from each provider. Cost estimates may vary if you:- Use different LLM providers
- Have custom deployment configurations (e.g., Azure OpenAI, AWS Bedrock)
- Have negotiated pricing agreements with providers
Custom per-token costs
For certain providers such as LiteLLM, you can provide custom input and output per-token cost information. This allows Freeplay to reflect your actual costs more accurately when you have special pricing or use self-hosted models. To configure custom token costs, see LiteLLM Proxy.Frequently asked questions
How can I reduce my costs?
How can I reduce my costs?
- Model selection — switching to a smaller or less expensive model for tasks that do not require the most capable model
- Prompt optimization — reducing token counts by writing more concise prompts
- Sampling rate — adjusting the sampling rate for online evaluations to run them on a subset of completions rather than all of them
- Evaluation frequency — running test evaluations less frequently or on smaller datasets during development
How is the per-token cost calculated?
How is the per-token cost calculated?
Freeplay uses publicly available pricing from each LLM provider to estimate the cost per input and output token. These rates are updated regularly to reflect the latest published pricing.
What if I have my own models or custom provider agreements?
What if I have my own models or custom provider agreements?
If you use self-hosted models, custom deployments, or have negotiated pricing with providers, the default per-token costs may not reflect your actual spend. For providers that support it (such as LiteLLM), you can configure custom per-token costs in Freeplay to get more accurate estimates. For other providers, treat the estimates as a relative benchmark for comparing costs across prompts and models.

