WTF? There are some LLMs that will just echo their initial system prompt (or maybe hallucinate one?). But that’s just on a different level and reads like it just repeated a different answer from someone else, hallucinated a random conversation or… just repeated what it told you before (probably in a different session?)
reloads, cocks gun Yeah, let’s give it another shot!