Definition
The context window (or context length) is the maximum number of tokens a model can consider when generating a response.
Why It Matters: - Determines how much text you can include in a prompt - Limits conversation history in chat applications - Affects ability to process long documents
Evolution of Context Windows: - GPT-3: 4,096 tokens - GPT-4: 8K-128K tokens - Claude 3: 200K tokens - Gemini 1.5: 1M+ tokens
Working with Limits: - Summarize long documents - Use RAG for relevant retrieval - Chunk large inputs - Prioritize recent context
Examples
With a 100K context window, you can paste an entire book into a single prompt.
Related Terms
AI models trained on massive text datasets that can understand and generate human-like text.
A technique combining information retrieval with text generation to improve accuracy.
The basic unit of text that language models process, roughly 3/4 of a word.
Want more AI knowledge?
Get bite-sized AI concepts delivered to your inbox.
Free intelligence briefs. No spam, unsubscribe anytime.