Architecture

Perplexity

Quick Answer

A metric measuring how well a model predicts text, calculated as the exponential of cross-entropy loss.

Perplexity is the exponential of average cross-entropy loss. It measures how 'surprised' the model is by test data. Lower perplexity indicates better predictions. Perplexity of 1 means perfect predictions (P=1 for correct tokens). Perplexity is intuitive: a perplexity of 50 means the model is effectively equally uncertain among 50 equally likely tokens. It's widely used as an evaluation metric for language models. Perplexity is comparable across datasets only if they use the same tokenizer. Recent models have dramatically lower perplexities than earlier ones.

Last verified: 2026-04-08

Compare models

See how different LLMs compare on benchmarks, pricing, and speed.

Browse all models →