Your occasional reminder that people who don't understand the state of play in LLMs sometimes brag about how easy they are to align. (Keeping in mind that current "AI safety" is corporate brand safety; that is what they are trying to do, so that failure is what is informative.)
LLM Alignment Difficulty Often Underestimated by Non-Experts
By
–
Leave a Reply