"The danger of AI is not that it's going to rebel against us, it's that it's going to do exactly what we ask it to do" But what happens when AI does exactly what we ask, yet not what we intend? Working with AI isn't like working with a human, more like working with a weird
AI Alignment Gap: Instruction Following vs Intent
By
–
Leave a Reply