Which of the following types of testing can MOST effectively mitigate prompt hacking?
A.
Load
B.
Input
C.
Regression
D.
Adversarial
The Answer Is:
D
This question includes an explanation.
Explanation:
Prompt hacking manipulates large language models by injecting adversarial instructions into inputs to bypass or override safeguards. The AAISM framework identifies adversarial testing as the most effective way to simulate such manipulative attempts, expose vulnerabilities, and improve the resilience of controls. Load testing evaluates performance, input testing checks format validation, and regression testing validates functionality after changes. None of these directly address the manipulation of natural language inputs. Adversarial testing is therefore the correct approach to mitigate prompt hacking risks.
[References:, AAISM Exam Content Outline – AI Risk Management (Testing and Assurance Practices), AI Security Management Study Guide – Adversarial Testing Against Prompt Manipulation, ]
AAISM PDF/Engine
Printable Format
Value of Money
100% Pass Assurance
Verified Answers
Researched by Industry Experts
Based on Real Exams Scenarios
100% Real Questions
Get 65% Discount on All Products,
Use Coupon: "ac4s65"