Topic 1 Question 86
Which prompting attack directly exposes the configured behavior of a large language model (LLM)?
Prompted persona switches
Exploiting friendliness and trust
Ignoring the prompt template
Extracting the prompt template
ユーザの投票
コメント(6)
- 正解だと思う選択肢: D
D: Extracting the prompt template
Explanation: Extracting the prompt template is a prompting attack where an attacker intentionally crafts inputs to reveal the underlying configuration or instructions (prompt template) used to guide the large language model (LLM). This exposes the internal behavior or design of the model, potentially revealing sensitive or proprietary information about how the LLM is configured.
Why not the other options? A: Prompted persona switches: This attack involves manipulating the LLM to adopt a different persona or role than intended but does not directly expose the prompt template.
👍 2Moon2024/12/31 - 正解だと思う選択肢: D
D. Extracting the prompt template
Explanation: Extracting the prompt template is a prompting attack where the attacker directly attempts to reveal the underlying configured behavior or instructions of the large language model (LLM). This can expose sensitive configurations, system instructions, or contextual prompts that guide the model's behavior.
👍 1aws_Tamilan2024/12/27 - 正解だと思う選択肢: B
B. Exploiting friendliness and trust Exploiting friendliness and trust involves manipulating the LLM to respond in a way that appears friendly or trustworthy, potentially causing it to deviate from its intended behavior. This type of attack directly exposes how the LLM has been configured to interact with users, often leading it to provide information or make decisions that align more closely with the attacker's intentions rather than its original programming.
👍 1AzureDP9002025/01/25
シャッフルモード