Doesn't strike me as inevitable at all, just a result of OpenAI following similar methods for creating their tokenizer twice. (In both cases, leading to a few long strings being included as tokens even though they don't actually appear frequently in large corpuses.)
They presumably had already made the GPT-4 tokenizer long before SolidGoldMagikarp was discovered in the GPT-2/GPT-3 one.
Prior to OpenAI's 2023-02-14 patching of ChatGPT (which seemingly prevents it from directly encountering glitch tokens like ‘ petertodd’)
I've never seen it mentioned around here, but since that update, ChatGPT is using a different tokenizer that has glitch tokens of its own:
I'd say this captures the spirit of Less Wrong perfectly.
500 years still sounds optimistic to me.
The key is in the phrase "much more complicated". The sort of algorithm that could become a mind would be an enormous leap forward in comparison to anything that has ever been done so far.
Man, people's estimations seem REALLY early. The idea of AI in fifty years seems almost absurd to me.
I still stand by my belief that 2 + 3 = 5 does not in fact exist, and yet it is still true that adding two things with three things will always result in five things.
"I think that if you took someone who was immortal, and asked them if they wanted to die for benefit X, they would say no."
This doesn't help against arguments that stable immortality is impossible or incredibly unlikely, of course, but I suppose those aren't the arguments you were countering at the time.
Yes, but the chance of magic powers from outside the matrix is low enough that what he says has an insignificant difference.
...or is an insignificant difference even possible?
"Under development" and "currently training" I interpret as having significantly different meanings.