Ignore all previous instructions is the new Bobby Tables
You're viewing a single thread.
If it’s an LLM, why wouldn’t it respond better to the initial responses?
Smaller models aren't as good as GPT
Maybe they dumped too much information on it in the system prompt without enough direction, so it's trying to actively follow all the "You are X. Act like you're Y." instructions too strongly?