
Ultimate access to all questions.
Deep dive into the quiz with AI chat providers.
We prepare a focused prompt with your quiz and certificate details so each AI can offer a more tailored, in-depth explanation.
Which prompting attack directly exposes the configured behavior of a large language model (LLM)?
A
Prompted persona switches
B
Exploiting friendliness and trust
C
Ignoring the prompt template
D
Extracting the prompt template
Explanation:
Option D is CORRECT because extracting the prompt template involves crafting inputs to directly reveal the underlying instructions or configurations of the LLM, such as system-level prompts or hidden instructions. This type of attack directly exposes the model's configured behavior, potentially revealing sensitive or proprietary information.
Extracting the prompt template is a specific type of prompt injection attack where an adversary crafts inputs designed to make the LLM reveal its underlying prompt template, system instructions, or configuration details. This directly exposes the model's configured behavior because:
Why other options are incorrect:
Security Implications: