Disclaimer: I have not read the Wentworth's post or the linked one but I know (little) about finite-sample and asymptotic bounds.
(He had another few versions, allegedly with fuller proofs, though I was not able to understand them and focused on this one.)
I think the key point of the statement is "any finite-entropy function ". This makes sure that the "infinity" in the sampling goes away. That being said, it should be possible to extend the proof to non-independent samples, Cosma Shalizi has done a ton of work on this.
The impact in ChatGPT could be potentially due to longer prompts or the "system prompt". It would be great to test that in a similar analysis
What about the luigis and waluigis in different languages, cultures, religions? Ones that can be described via code? It feels like you can always invent new waluigis unless the RLHF killed all of the waluigis from your pre-training data (whatever that means)
The token limit (let's call that ) is your limit here, you just need to create a waluigi in steps, so that you can utilize him for the last steps. I think this eventually breaks down to something about computational bounds, like can you create a waluigi in this much time
Second insight:
If you can find Luigi and Waluigi in the behavior vector space, then you have a helpful direction to nudge the AI towards. You nudge it in the direction ofLuigi - Waluigi
.
You need to do this for all (x,y) pairs of Luigis and Waluigis. How do you enumerate all the good things in the world with their evil twins, and then somehow compare the internal embedding shift against all of these directions? Is that even feasible? You probably would just get stuck.