If ChatGPT can be jailbroken with only 20 cents, what can creators do to prevent their models from being repurposed for harm? Researchers describe one initial direction for increasing the costs of doing so: the self-destructing model.
Self-Destructing Models: Preventing AI Jailbreaking and Misuse
By
–
Leave a Reply