All of zchuang's Comments + Replies

zchuang10

Can you share the difference between the times it is aware (I know 33% is the floor) and the times where Claude is not aware/sandbagging? 

zchuang7-2

The Claude plays pokemon discourse makes me feel out of touch because I don't understand how this is not an update people had with chess playing LLMs. If anything, I think the data contamination is worse for pokemon because of gamefaq and other guides online. 

I think people are mistaking benchmarks and trailheads. Maths benchmarks are important because they theoretically speed up AI research and benchmark to useful intelligence. Claude playing pokemon doesn't tell you much because it's not a map on intelligence nor of generalisation. 

It's a good unhobbling eval, because it's a task that should be easy for current frontier LLMs at System 1 level, and they only fail because some basic memory/adaptation faculties that humans have are outright missing for AIs right now. No longer failing will be a milestone in no longer obviously missing such features (assuming the improvements are to general management of very long context, and not something overly eval-specific).

4Oliver Daniels
Map on "coherence over long time horizon" / agency. I suspect contamination is much worse for chess (so many direct examples of board configurations and moves) 
zchuang90

Sorry, but why did you connect apathy and lying-flat [sic] with fast-follower culture. Lying flat or tang ping is about youth apathy and nihilism about the job market and existential angst about life in the 21st century. It's not about corporate or industrial culture from the top end or specific technological political economy strategies.

zchuang190

I don't know if this is helpful but as someone who was quite good at competitive Pokemon during their teenage years and also still keeps up with nuzlocking type things for fun, I would note that Pokemon's game design is made to be a low context intensity RPG especially in early generations where the linearity is pushed to allow kids to do it. 

If your point holds true on agency, I think the more important pinch points will be Lavender Town and Sabrina because those require backtracking through the storyline to get things.

I think mid-late game GSC would also be important to try because there are huge level gaps and transitions in the storyline that would make it hard to progress.

Sorry but aren't we in a fast takeoff world at the point of WBE.  What's the disjunctive world of no recursive self-improvement and WBE? 

1RussellThor
I guess a world with a high chance of happening is where we develop AGI with HW not that much different from what we currently have, i.e. AGI in <5 years. The Von Neumann Bottleneck is a fundamental limit, so we may have many fast IQ 160 AGI, or a slower than human IQ 200 one that thinks for 6 months and concludes with high confidence that we need to build better hardware for it to improved more. There is large room for improvement with a new chip design it has come up with. Then we have a choice - instead of building such HW to run an AGI we do WBE instead - inefficiently with the VNB HW with the understanding that with more advanced HW we will run WBE rather than AGI.

He posted on a twitter a request to talk to people who feel strongly here.

Yeah, re-reading I realise I was unclear. Given your claim: "by the time we get to 2000 in that, such AGIs will be automating huge portions of AI R&D,". I'm asking the following:

  1. Is the 2000 mark predicated on automation of things we can't envision now (finding secret sauce to singularity) or is it predicated off pushing existing things like AI R&D finds better compute or is it a combination of both?
  2. What's the empirical on the ground representative modal action you're seeing at 2025 from either your vignette (e.g. I found the diplomacy AI super impo
... (read more)
4Daniel Kokotajlo
1. A bit of both I guess? Insofar as there are things we can't imagine now, but that a smart hardworking human expert could do in 2000 seconds if and when the situation arose / if and when they put their mind to it, then those things will be doable by AI in 2026 in my median projection. 2. Unfortunately, my view and Richard's views don't really diverge strongly until it's pretty much too late, as you can see from the chart. :( I think both of us will be very interested in trying to track/measure growth in t-AGI, but he'd be trying to measure the growth rate and then assume it continues going at that rate for years and years, whereas I'd be trying to measure the growth rate so I can guess at how long we have left before it 'goes critical' so to speak. For 'going critical' we can look at how fast AI R&D is going and try to measure and track that and notice when it starts to speed up due to automation. It's already probably sped up a little bit, e.g. Copilot helping with coding...

Sorry for a slightly dumb question but in your part of the table you set 2000 as the year before singularity and your explanation is that 2000-second tasks jump to singularity. Is your model of fast take-off then contingent on there being more special sauce for intelligence being somewhat redundant as a crux because recursive self-improvement is just much more effective. I'm having trouble envisioning a 2000-second task + more scaling and tuning --> singularity. 

Additional question is what your model of falsification is for let's say 25-second task... (read more)

3Daniel Kokotajlo
Oooh, thanks for pointing out the typo! Fixed. I'm not sure I understand your questions, can you elaborate/explain them more? Note that the numbers in this chart don't represent "There is at least one task of length X that AI can do" but rather "There exist AGIs which can do pretty much all relevant intellectual tasks of length X or less." So my claim is, by the time we get to 2000 in that, such AGIs will be automating huge portions of AI R&D, which will speed it up, which will get us to 10,000 in less than a year, which will speed things up even more, etc.