
Ultimate access to all questions.
Deep dive into the quiz with AI chat providers.
We prepare a focused prompt with your quiz and certificate details so each AI can offer a more tailored, in-depth explanation.
Q1 – If a model has a context window of 8,000 tokens, what happens when the user inputs 10,000 tokens?
A
The model automatically expands its context window
B
The model truncates or ignores excess tokens
C
The model compresses input into embeddings to fit
D
The model fails permanently
Explanation:
When a language model has a fixed context window of 8,000 tokens and receives an input of 10,000 tokens, it cannot process all the tokens at once. Here's what happens:
Context Window Limitation: The context window represents the maximum number of tokens the model can process in a single forward pass.
Truncation Mechanism: The model will typically truncate the input by:
Removing tokens from the beginning or end of the input
Keeping only the most recent or relevant tokens
Ignoring the excess tokens beyond the 8,000 limit
Why Not Other Options:
A (Automatic expansion): Models have fixed context windows that cannot be dynamically expanded without architectural changes.
C (Compression into embeddings): While embeddings represent tokens, they don't compress token count - each token still requires embedding space.
D (Permanent failure): Models don't fail permanently; they simply process what fits within their context window.
Practical Implications: This truncation can lead to loss of important context, especially for long documents or conversations. Some advanced models use techniques like sliding windows or hierarchical processing for longer contexts, but standard models with fixed context windows simply truncate.
Key Takeaway: Always check the context window limitations of a model before providing input, as excess tokens will be lost through truncation.