AI Dynamics

Global AI News Aggregator

GPT-4 Jailbreak Difficulty Scales Exponentially with Output Severity

There is a sliding scale for jailbreak output that exponentially increases in difficulty to crack It's trivial to get GPT-4 to curse but if you want a set of instructions on making a weapon it's going to take a lot of work

→ View original post on X — @alexalbert__,

Commentaires

Leave a Reply

Your email address will not be published. Required fields are marked *