LLMs process text from multiple sources and may face conflicting instructions. We teach our models to follow instructions from the highest priority input, giving better defense against attacks. Our Safety Systems team is hiring:
LLM Safety: Instruction Hierarchy Defense and Team Hiring
By
–
Leave a Reply