If an LLM is saying something to a human that it knows is false, this is very bad and is the top priority to fix. After that we can talk about when it's okay for an AI to keep quiet and say other things not meant to deceive. Then, discuss if the LLM is thinking false stuff.
LLM Truthfulness: Preventing Deliberate Falsehoods to Humans
By
–
Leave a Reply