About

A practical LLM cost simulator for builders.

llmcalc.app exists because LLM pricing has gotten genuinely complicated. The headline "$X per million tokens" rarely matches what you actually pay. Cache reads, cache writes, batch discounts, long-context tiers, reasoning tokens, retries, multi-turn loops — these compound, and most calculators ignore them.

What this is

  • A multi-model token counter that runs entirely in your browser.
  • A cost calculator with input/output/cache/batch/long-context fields modeled correctly.
  • Three differentiated calculators: prompt-caching ROI, batch decision tool, and a multi-turn agent-loop simulator with workload templates.
  • Model and comparison reference pages, all driven from a single models.json.

What it isn't

  • A replacement for your provider's invoice. Always reconcile against the real bill.
  • A live monitor. Pricing data is verified periodically, not in real time.
  • Affiliated with OpenAI, Anthropic, Google, or Meta. We're a third party.

How the data is kept honest

Every model in our index carries two things you should look at: a last_verified date and a confidence flag (official or inferred). The changelog tracks pricing changes and verification dates publicly.

Spot a wrong number? Use the report incorrect price link in the footer. Includes auto-filled context for the model and the official docs URL.

Tracked models

Currently 10 frontier models across Anthropic, OpenAI, Google, and Meta. See the full list.

Tech

Static site (Astro + Tailwind). Tokenization runs client-side via tiktoken-class libraries. No analytics cookies. Privacy-respecting traffic counts via Cloudflare Web Analytics.