AI agents inherit a fundamental architectural flaw analogous to the von Neumann architecture's code-data conflation: LLMs cannot distinguish between trusted instructions and untrusted data, making indirect prompt injection attacks nearly impossible to prevent at the architectural level. Unlike traditional code injection
7 Comments
Sort: