AI Dynamics

Global AI News Aggregator

AI Models Trained to Resist Prompt Injection Attacks

I think both – plus the labs have been putting a lot of effort into training them to resist prompt injection style attacks Anthropic usually mention prompt injection in their system cards eg this one for Opus 4.6 https://
www-cdn.anthropic.com/0dd865075ad313
2672ee0ab40b05a53f14cf5288.pdf

→ View original post on X — @simonw,

Commentaires

Leave a Reply

Your email address will not be published. Required fields are marked *