What is a Context Window?
The maximum amount of text (measured in tokens) that a language model can process in a single inference call.
Every LLM has a finite context window. Fitting all relevant information into this window is a core challenge in AI development, especially for long documents or large knowledge bases. Techniques like RAG and structured context containers help work around this limitation by retrieving only what's needed.
Related concepts
A technique that retrieves relevant documents or data at inference time and injects them into the model's context window before generating a response.
The practice of deliberately designing, structuring, and managing the information provided to AI models to improve output quality and relevance.
Further reading
Put context into practice
Create your first context container and connect it to your AI tools in minutes.
Create Your First Container