Pricing & Cost

Input Tokens

Quick Answer

Tokens in the prompt, billed separately (usually cheaper than output tokens).

Input tokens are tokens in your prompt/context. APIs charge per input token (usually less than output tokens). Understanding input token count is crucial for cost optimization. Techniques reducing input tokens: prompt compression, semantic caching, instruction templating. Cached input tokens cost less (prompt caching feature). Input token optimization is often more valuable than output optimization. Different APIs have different per-token pricing.

Last verified: 2026-04-08

Compare models

See how different LLMs compare on benchmarks, pricing, and speed.

Browse all models →