
Answer-first summary for fast verification
Answer: Decrease temperature
## Explanation **Temperature** controls the randomness/creativity of model outputs: - **Lower temperature**: Makes the model more deterministic, focused, and factual - **Higher temperature**: Increases creativity and randomness, which can lead to imaginative but potentially off-topic responses Since the HR assistant is giving "imaginative but off-topic answers," the team needs to make it more factual by **decreasing the temperature**. Option A is correct because: 1. Lower temperature reduces randomness in token selection 2. The model becomes more focused on the most probable next tokens 3. This results in more predictable, factual, and on-topic responses **Why other options are incorrect:** - **B**: Increasing top-p (nucleus sampling) would actually increase diversity by considering a broader set of tokens - **C**: Raising max-tokens only increases response length, not factual accuracy - **D**: Disabling stop-sequences would allow responses to continue indefinitely, not improve factual accuracy
Author: Jin H
Ultimate access to all questions.
No comments yet.
An HR assistant built on Bedrock starts giving imaginative but off-topic answers. The team wants more factual responses. What should they change?
A
Decrease temperature
B
Increase top-p
C
Raise max-tokens
D
Disable stop-sequences