veda.ng
Back to Glossary

Context Window

Context Window infographic

A context window is the maximum amount of text a language model can process in a single interaction, measured in tokens. Everything the model can 'see' at once (your prompt, the conversation history, any documents you've pasted in) must fit inside this window. Early GPT models had context windows of 4,000 tokens, roughly 3,000 words.

Modern models like Claude and GPT-4 have windows of 128,000 to 200,000 tokens or more, enough to hold an entire novel. When content exceeds the context window, the model either truncates it or cannot process it at all. The model has no memory of what fell outside the window. So, very long conversations can cause models to 'forget' earlier messages.

Context window size directly determines what tasks a model can perform. A small window can answer questions and write short documents. A large window can analyze entire codebases, summarize lengthy reports, or maintain coherent long-form conversations. The race to extend context windows is one of the central engineering challenges in LLM development.

Interactive Visualizer

Context Window Visualizer

Explore how different AI models handle varying amounts of text within their context windows. Adjust the content to see when you exceed the model's capacity.

2,000 tokens

GPT-3 Context Window

04,000 tokens
Used Tokens
2,000
50.0% of window
Remaining
2,000
tokens available
💡 Tip: When content exceeds the window, older parts of the conversation are typically removed first.