
Ultimate access to all questions.
Deep dive into the quiz with AI chat providers.
We prepare a focused prompt with your quiz and certificate details so each AI can offer a more tailored, in-depth explanation.
A summarization LLM produces outputs that are too short and miss important details. What's the best adjustment?
A
Increase top-p from 0.7 → 0.9
B
Reduce temperature to 0.2
C
Set max_tokens higher
D
Increase repetition penalty
Explanation:
When a summarization LLM produces outputs that are too short and miss important details, the most direct solution is to increase the max_tokens parameter.
max_tokens controls output length: This parameter sets the maximum number of tokens the model can generate in its response. If outputs are too short, increasing this value allows the model to generate more detailed summaries.min_tokens parameter (if available) could also be set to ensure a minimum output length