This post explains how to mitigate prompt injection risks for AI agents on Databricks using the Agents Rule of Two framework.
- •Three pillars make agents vulnerable: access to sensitive data, exposure to untrustworthy inputs, and ability to change state or communicate externally.
- •When all three pillars converge, attackers can inject malicious instructions via embedded prompts in documents or APIs to exfiltrate sensitive data.
- •Indirect prompt injection embeds malicious instructions in databases, free-text fields, or attachments, making detection harder than direct attacks.
- •Databricks mitigates risks through layered controls in Unity Catalog and Agent Bricks across data access, model interaction, and runtime execution.
This summary was automatically generated by AI based on the original article and may not be fully accurate.