
Answer-first summary for fast verification
Answer: The model truncates or ignores excess tokens
When a model has a fixed context window of 8,000 tokens and receives an input of 10,000 tokens, the model will **truncate or ignore the excess tokens** beyond its capacity. This is a standard behavior in language models where inputs exceeding the context limit are typically truncated from the beginning, end, or both to fit within the available window. The model does not automatically expand its context window (A), compress inputs into embeddings (C), or fail permanently (D).
Author: Ritesh Yadav
Ultimate access to all questions.
If a model has a context window of 8,000 tokens, what happens when the user inputs 10,000 tokens?
A
The model automatically expands its context window
B
The model truncates or ignores excess tokens
C
The model compresses input into embeddings to fit
D
The model fails permanently
No comments yet.