Topic 1 Question 89
A company is testing the security of a foundation model (FM). During testing, the company wants to get around the safety features and make harmful content.
Which security technique is this an example of?
Fuzzing training data to find vulnerabilities
Denial of service (DoS)
Penetration testing with authorization
Jailbreak
ユーザの投票
コメント(4)
- 正解だと思う選択肢: D
ML Jailbreak security ML jailbreak refers to techniques used to bypass the safety and security measures of machine learning models, particularly large language models (LLMs). This can lead to the model producing harmful, inappropriate, or unintended content1. Here are some key points about ML jailbreak security
👍 226b8fe12024/12/26 - 正解だと思う選択肢: D
D. Jailbreak
Explanation: Jailbreaking is a technique used to bypass the safety features and restrictions of a foundation model (FM). The goal is to manipulate the model into generating harmful, inappropriate, or otherwise unintended content, despite the safeguards in place. This is often done to test the robustness of the model's safety mechanisms.
👍 2aws_Tamilan2024/12/27 - 正解だと思う選択肢: D
The correct answer is D. A jailbreak is an attempt to bypass an AI model's built-in safety controls.
👍 1may2021_r2024/12/28
シャッフルモード