AI Dynamics

Global AI News Aggregator

Self-Destructing Models: Preventing AI Jailbreaking and Misuse

If ChatGPT can be jailbroken with only 20 cents, what can creators do to prevent their models from being repurposed for harm? Researchers describe one initial direction for increasing the costs of doing so: the self-destructing model.

→ View original post on X — @stanfordhai,

Commentaires

Leave a Reply

Your email address will not be published. Required fields are marked *