Context Window
The amount of text an AI model can remember and process at once, measured in tokens. Bigger windows mean longer conversations and documents.
What is Context Window?
A context window is the working memory of an AI model - how much text it can keep track of while generating a response.
It's measured in tokens (roughly 3/4 of a word each). Claude 3.5 Sonnet has a 200K token window (about 150,000 words). GPT-4 ranges from 8K to 128K depending on the version.
When you hit the limit, the model starts forgetting earlier parts of your conversation or cuts off mid-response. Builders use larger windows for analyzing full codebases, processing long documents, or maintaining context across complex multi-step tasks.
Longer contexts cost more because the model does exponentially more computation. A 100K token input costs roughly 100x more to process than a 1K token input.
Good to Know
How Vibe Coders Use Context Window
Frequently Asked Questions
Your Idea to AI Business In Days
Join Dan, Zehra and 0 others building AI businesses in days with video tutorials and 1 on 1 support.
Related Terms
A specialized database that stores data as mathematical vectors (embeddings) to enable fast semantic search and AI-powered similarity matching.
The text instruction you give an AI model to tell it what you want it to do, like asking ChatGPT to write code or explain a concept.
The practice of crafting specific instructions to get better outputs from AI models like ChatGPT, Claude, or Gemini.
A technique that lets AI models search your documents or databases before answering, combining real-time data retrieval with text generation.
Google's multimodal AI that can understand and generate text, images, audio, video, and code in a single conversation.
Join 0 others building with AI