Table with LLMs Token Output Limit

Summary: The table provides a comparison of Context Window and Maximum Output Tokens of various large language models (LLMs) including Claude 3, GPT-4 Turbo, Gemini, Mixtral, and Llama.

Every model has a limit to the number of tokens in the input prompt, normally called "Context Window." Additionally, each model has a limit on the number of tokens that it can generate in the output. This limit is sometimes called "maximum new tokens" or "maximum output tokens."

Let me know if you want me to add more information to the table. Other columns that might be useful, or other LLMs.

0
Subscribe to my newsletter

Read articles from Juan Pablo Mesa Lopez directly inside your inbox. Subscribe to the newsletter, and don't miss out.

Written by

Juan Pablo Mesa Lopez
Juan Pablo Mesa Lopez