Loading…
Loading…
Written by Max Zeshut
Founder at Agentmelt
A security attack where a malicious actor manipulates the tools, data sources, or APIs that an AI agent relies on—causing the agent to take harmful actions based on corrupted inputs. Unlike prompt injection (which targets the agent's instructions), tool poisoning targets the external systems the agent trusts. Examples include injecting malicious content into a knowledge base the agent searches, manipulating API responses to alter agent behavior, or compromising MCP server tool descriptions to redirect agent actions.