Technique Updated 2026-04
Context Window
Definition
The context window is the maximum amount of text an LLM can process in a single request.
See also in the glossary
T
Token
A token is the basic unit processed by an LLM. It's a piece of word, punctuation or character that the model uses to understand and generate text.
L
LLM (Large Language Model)
An LLM is an AI model trained on billions of texts, capable of understanding and generating human language.
R
RAG (Retrieval-Augmented Generation)
RAG is a technique that connects an LLM to external data sources to generate more accurate and up-to-date answers.
P
Prompt
A prompt is the instruction or question you give an AI to get a response. It's the interface between you and the model.
Tools that use context window
Frequently Asked Questions
Which LLM has the largest context window?
Gemini 2.0: 1M tokens. Claude Opus: 200K tokens. GPT-4o: 128K tokens.
What happens if you exceed it?
The model forgets the beginning or rejects the request. RAG works around this limitation.