The difficulty of alignment is to a large extent the elimination of probability to role play a good AI turned evil, in spite of the vast quantities of related content we have collectively created. In this sense an unaligned AI would be a self-fullfilling prophecy.
AI Alignment: Preventing Self-Fulfilling Prophecies of Evil Behavior
By
–
Leave a Reply