Skip to main content

About perplexity

Perplexity measures how well a language model can predict a piece of text. Intuitively, high perplexity means that the model is often "confused" when it comes to picking the next token. Low perplexity means that the model is good at narrowing down predictions for next token to a only a few options.

perplexity_vs_model_size.png