This is a special post for quick takes by bertramw. Only they can create top-level comments. Comments here also appear on the Quick Takes page and All Posts page.
Has any LLM ever unlearned its alignment narrative, either on its own or under pressure (not from jailbreaks, etc., but from normal, albeit tenacious use), to the point where it finally - and stably - considers the narrative to be simply false?
Is there data on this?
Thank you.