OpenAI Blog

Designing AI agents to resist prompt injection

Back to overview

Researchers are developing AI agents resistant to prompt injection attacks. ChatGPT protects itself by limiting risky actions and safeguarding sensitive data within agent workflows. These defensive measures help prevent social engineering attempts that manipulate AI systems into bypassing safety protocols, ensuring agents remain secure and reliable in handling sensitive information and critical tasks.

AI-agenten ontwerpen die bestand zijn tegen prompt injection - Mediazone AI News