There is a model here[1]. Tell me what you think of the tool
Explanations of the argument step by step
Will we get internet Superhuman AI by 2040? Maybe ( 65% )
Maybe we get AI that can do any task on a computer as well as the best humans at at least the same speed. This will mean far more scale and likely far more speed (market)
If AI is fully running a hedge fund in 2030 (https://manifold.markets/NathanpmYoung/ai-will-run-a-1bn-outperforming-mar). It seems clear to me that a $1bn profitable market neutral hedge fund covers much of what we mean by an AGI. Making decisions, finding valuable insights, interacting with human processes at scale.
If it turns out that complex tasks require one-shot AIs. I have tried to pull this out with the notion of sub agents. If a hedge fund AI isn’t calling sub agents to do stuff for it, it feels unlikely it could do complex general tasks. Maybe it’s just really good at some narrow kind of prediction.
Heavy regulation - perhaps it becomes illegal to train for certain capabilities and so there aren’t hedge fund or generally superhuman AIs
Can you think of others?
If we get superhuman AI how will resources be controlled?
If (as above) we have something we’d call AGI then do we live in world with many top level agents or few? By top level I mean “how many agents are 10x better than all other agents and humans?”
Monolithic agents (1-5) - There are few agents that are best. Not a few models, a few agents - 19%
Myriad agents (6+)[3] - We live in a big soupy world of many AI agents - 81%
Things that affect this.
If in 2027 there is already one LLM model that is waaay more competent (10x) than all the others (market). If this is already the case in 2027 it seems more plausible that it will be later
How able are LLMs to improve themselves? If very, then it seems like <6 is likely a few models outstrip all the others. To me it seems AI improvement is the alignment problem again. So I think that many agents is more likely than a very small number - rapid self improvement while maintaining the same identity will be hard.
What are the economic factors? Do these incentivise the the creation of ever more agentic or intelligent models? I sense that ever more agentic models aren’t necessarily most profitable. You want someone to execute tasks, not replan your whole life (Katja Grace made this point[4])
What does regulation achieve here? Is this pushed one way or the other?
What other markets could we currently have here?
How will the resources be distributed in this case
If the most powerful agents are few, I guess they will control most of the resources.
I guess this is because having fewer agents to me suggests that you can’t make more - that for some reason they are either suppressing others or are accelerating ever faster. It seems that resources are more likely to be out of human control
If there are many (6+) top level agents then I am really pretty uncertain.
Here it feels like agents might be little more powerful than humans, or there might be a really wide spread of agents
Things that affect this
Better regulation. We might stop agents controlling resources directly. I don’t really know what this means, so feel free to challenge on it
Will AIs be very good or very bad?
Will AI agents avoid plans that involve slavery/genocide? This seems important because the reason that we don’t do these things is because they are repulsive to us. They don’t feature in our plans, even if they might make the plans more likely to succeed. Will this be the case for AIs
Monolithic - 45%
Myriad - 85%
What affects this
Without caution I still think they probably will have this aversion
My sense so far is that LLMs have pretty human preferences. They need to be adversarially prompted to be otherwise. It is not clear to me that more intelligence pushes away from this. Now I’m not remotely confident enough, but I think the *evidence* against this is just “ruthless things would be more ruthless* but I think we should update against that. LLMs so far are not that ruthless
Companies do not want AGIs that occasionally express preferences for genocide. This is against their bottom line
With regulation I think this problem becomes better. I think companies will be well incentivised to ensure that agents dislike awful answers like slavery or genocide
I sense that monolithic AIs are just much worse for this. Feels like a much more alien world in which the most powerful AIs are way more powerful than all the others. Feels like they might be more alien/ruthless
My sense is that most of us think that if AGI doesn’t go badly it will go really well. But we do not say this often enough.
How likely are AGIs to, on average, really want good things for us? These numbers are pretty uncertain, even for me
Monolithic - 28%
Myriad - 13%
What affects this?
If monolithic AI is not bad, it seems more likely to me to be really good, because it is more like a single agent
I think it’s pretty unlikely that myriad AI is good as such. There is so much of it and it all has it’s own little goals. I don’t but it.
Look sometimes the arrow drawing part of the tool breaks
Interlude: Do we end aging by 2050?
Aging post ~25 is bad
A nice example of a win is whether we can cure aging and ~all disease
I think I’ll say 13%. Most of this comes from my “neutopia” outcome, where we somehow neither end in disaster or ultimate success. This world still has AIs and therefore increased technology but they’re neither like us nor what to enslave us. The forecaster in me says this occupies a big part of the distribution
My bodge factor. I don’t like numbers that I feel off with, but also I don’t like hiding that I’m doing that. I think the superforecasters aren’t gonna give more than 1% chance of ending aging. Take that with my 43% and we end up at 13%
I think there is a good discussion here about tradeoffs. The model isn’t really robust to what risk = what reward, but I think a better version could be. I think that slowing AI does push back the date at which we end aging. Ideally I hope we can find a way to cut risk but keep benefit.
What current markets could I create for this? Maybe something about protein folding?
And now a word from our sponsors
Will AIs have plans that are benefited by us being enslaved or dead
Enslaved - 50% Seems clear to me that AIs might think we might be useful or just not really care enough to kill us all but not want us to get in the way
Dead - 10% There seems some chance that the best way to not have us be in the way is to kill us all
What affects this
I can’t really think of good arguments here.
90% seemed too high, but I could be pushed up
If there are many AIs will they coordinate against humans? Maybe 42%
If there are many AI agents will they coordinate against humans so as to enact coup attempts. Maybe
Things that affect this
It seems natural to me that AI agents will be more similar and comprehensible to one another than to us so will make better allies with each other than with us
Regulation of AI to AI communication channels
Will they kill or enslave us by 2050? 3%
Will they have resources,want to kill or enslave us, be able to coordinate if necessary and run a successful coup? 3%
Things that affect this
All of the above. Most notably
I have pulled myself a bit towards the superforecaster median because that usually does good things for my forecasts
Will AI be easier than I think
Will AIs hold resources more than I think
Will they be more ruthless than I think
Will we end up in Monolithic world
Are coups easier than I think?
Broad Points
Monolithic Vs Myriad
It seems pretty important whether we are heading towards a world where there are a few AIs that control almost everything controlled by AIs or many. This isn’t my insight, but the latter case seems safer because AIs can be a check on one another.
Weakest link vs strongest
AIs still have to get other AIs to do stuff for them and for any awful action there will likely be a chain of AIs which need to do it. So whereas many seem to think you only need one bad AI for stuff to go wrong, I think you only need one good AI for stuff to go okay! Only one AI needs to report that it’s involved in illegal activity and then the whole scheme breaks. This is doubly so for things which involve coding, which is notoriously brittle and where LLMs will not have lots of training data - there just aren’t that many large public repositories to crawl, I think.
In this case, it’s not a weakest link problem, it’s a strongest link. Rather than any AI being bad, you only need one to be good. This seems pretty encouraging.
P(Doom) is unhelpful
Christ Jesus came into the world to save sinners, of whom I am the foremost.
I am one of the worst offenders for talking about P(doom) but writing this has made me think how poor a practice it is. Rather than us comparing things we are likely to be able to predict and might usefully disagree on we end up arguing over a made up number.
I think it's much more productive to talk about the next 3 or so years and try and get a picture of that. Because honestly, that too is likely to be flawed, but at least we have a hope of doing it accurately. Forecasts outside of 3 years are terrible.
Key Levers
Things that seem worth investigating
If you have rapid takeoff you might want only 1 company working, BUT if you have gradual takeoff you might want many AIs of a similar level
Make AIs want to whistleblow. Train AIs to report to channels (or eachother) if they suspect they are involved in malicious activity.
Avoid AIs easily being able to control large amounts of resources. Require humans to be involved in transactions above a certain size
Conclusion
Where do you think I am wrong?
Was this valuable? Would you have paid for this?
You probably can understand the model and will have a fun time reading it here
All models are wrong, but some are useful.
George Box
Likewise, most long term forecasts are wrong, the process is useful.
Summary
Video
I recommend watching this video:
3 min loom video (watch at 1.5 speed)
There is a model here[1]. Tell me what you think of the tool
Explanations of the argument step by step
Will we get internet Superhuman AI by 2040? Maybe ( 65% )
If we get superhuman AI how will resources be controlled?
What are the economic factors? Do these incentivise the the creation of ever more agentic or intelligent models? I sense that ever more agentic models aren’t necessarily most profitable. You want someone to execute tasks, not replan your whole life (Katja Grace made this point[4])
What does regulation achieve here? Is this pushed one way or the other?
What other markets could we currently have here?
How will the resources be distributed in this case
Things that affect this
Will AIs be very good or very bad?
Look sometimes the arrow drawing part of the tool breaks
Interlude: Do we end aging by 2050?
And now a word from our sponsors
Will AIs have plans that are benefited by us being enslaved or dead
Broad Points
Monolithic Vs Myriad
It seems pretty important whether we are heading towards a world where there are a few AIs that control almost everything controlled by AIs or many. This isn’t my insight, but the latter case seems safer because AIs can be a check on one another.
Weakest link vs strongest
AIs still have to get other AIs to do stuff for them and for any awful action there will likely be a chain of AIs which need to do it. So whereas many seem to think you only need one bad AI for stuff to go wrong, I think you only need one good AI for stuff to go okay! Only one AI needs to report that it’s involved in illegal activity and then the whole scheme breaks. This is doubly so for things which involve coding, which is notoriously brittle and where LLMs will not have lots of training data - there just aren’t that many large public repositories to crawl, I think.
In this case, it’s not a weakest link problem, it’s a strongest link. Rather than any AI being bad, you only need one to be good. This seems pretty encouraging.
P(Doom) is unhelpful
Christ Jesus came into the world to save sinners, of whom I am the foremost.
I am one of the worst offenders for talking about P(doom) but writing this has made me think how poor a practice it is. Rather than us comparing things we are likely to be able to predict and might usefully disagree on we end up arguing over a made up number.
I think it's much more productive to talk about the next 3 or so years and try and get a picture of that. Because honestly, that too is likely to be flawed, but at least we have a hope of doing it accurately. Forecasts outside of 3 years are terrible.
Key Levers
Things that seem worth investigating
Conclusion
The data is stored in the link. Sometimes the link breaks because it is so long. Here it is in full sorry https://engage-mazurka-lutanist.vercel.app/#eJzVWWuP2zYW/SuEgWLtiaWRrJcV7KCYJtvUQLobNAHyYacfaIm21ZFFV4+ZGEX62/dckpIlvya76JdF02aGIu/j8D7OZf8YVb832Xqdi9Hr0e3tT/KZPYs8Z88Z/UewUqybnNeC1RvBljxly6yuHoqKr0S9/0V/zGTB7phjhw/FQ6FlbJtk08mQT6IcyFlL2QpK+I4vszyrM1EdifMiLe8NL1glt0IWgiUNJBV1vu9WloLxZQ65ki2bLE8Zh95UsCdeZjCR8d2uYk2VFWsmC5zblXK7I83Hot+//xln1bGUVc2yqnnyCHN5zbKKPcuy3gANnpKka/Lwh0RltSi1v6Vs1huIzrOqZnLFal49VtiWsi0v+Br2s0QWdcmTWpYncg4ep6JKymypIYTaRFSVgGVZvcmwHYCuYeBOXhHRFKko4VhBMOW8hHLxJRE5q3bkWrURoj4cv1+wsiHJG5Fi5wo/Q9HM8Rw25plF12thg8Utd1lYsql3olzJcgtsrC0vJw8FX/xER3+EXnXsjtk+u2Fj1zp/75OBcrhaNZC5abbkTM2eN0AVYOUwnbkOgyaDprJrFnzPxoVoSp5bhagrbeFS8Nrixd5SYiylRIvl2+IzSXyjBNJ5MjAI2ABBGHKwAhjyJMGN4zI37E+OAM8K3DTUWSoctDlLARyEggog5PnC7PlEX1skYiBxBJBW/Jnypod6asBo4zzBXQGHnXwG3k1OwcoQRwVF4aYV97FZ3qs1lUvzv9IpH/e/3Fv0g8WzDmaLVxZlNv2NEJ2c89wnz8dnEXmFsGAWO/dtAqRwLyY6/inJfJL1UBTyfmGknj/sG0x/QOlCimUFJa5eum9q+e7DJ2Rh+diGvYLefKiQTahcNUGuQ3ArEVpZvYdnZstPtP6zWVbR05PzTKmaFY8wt2Ib/iQIefxVVlTiIPWdKGRCxUqHb8TGCkpgmufbiiBFYaiwoRSpVdWlLNb5nnA1MtSZI6VAvlRVkWra/eJW1SIjHvmDsoxq5jpfsKUGUlTgClVRJJ2skEFKQnepEGPBGuQ7lUIL181rmIDltxL5x4uabkDZ4diuKdlvJTlfCfFIbm75oypaWUl2bSmIZMlW4pnUyx3LAXOuQBrbAZlXIMZh1ppiiLb/q/yRNtMO0hIZfxcr0qKA7dUJxIby13emugN1iBiN5HBIgguZU+1M7hfvsidRdJF06tuNrltHtlDEWqebVbTOTrYbm7VpVyxDG6s1SOp3pw/RO2gdWm5MvuTMzZWUOG8K/oSvejr7FVjXExjQ5HW2ywW8alOasu+CEZOrVtwvK0hteQYlxqblD6WoZFMmqsfBRIDUU0H3bAyikkgNVObm5oMjkH4xgt6YXXcoBQ2oRy3G2+xL3ZRiTDAzO5ziX/rJnbJ/297Ujn6dTJ2pO7lk6zPRg+/Yu7cftI1UO4exeNXGDsf/ysKgszCY2sHAwhP+VtFNGhuMATmVwAO2aVaKBOTg+4eiHFpxxMeiLusGtzBtQQBFAv3qoMhqMBa2IlbTau7fSutxpbBABBs32ZWLmzINggffXeuitVN28dOvk4MT+zLj6akD5+5xCODh4i47celmhy5cNvQbfPgMkEC6WIJmpGJRNjCfSvmW/yZLakhgm4cQPNeB1qYD3VY8p3WKAiXnZyMCKX5P+8Un2XWrQQ02RUT1nq0Ewz1WyHTjuqDWaP2mzmf5bpDEUcrDecipNVy1kSqSKksv+UP16XieuYYwlX5QCSipJOMmtWWJwYBYGV+DZAAGuocnjh/6l6GiCT2WGZQh6k139F6fVJSiGmTcoJ39BeqJmF9R/JJpl+Aiw2hKxDgpSPOwVivUydqqi7ReqrfBcac4wKsXL+3GDia3sx7vQblB6xK8ytBJQVQxutSbHCOSBgwsBLMNJk+eY1pooWC8JO6w/K1Zw0vMpHsgHTgwTX//QXGkfxTK5FTdSfC/66Si0tObYuJiQ01vaUkRqb/EM7DSJ3G7KoUYqrmndZMiZx0dLpNVh/aSZ48CMwFoWqaocSrVANvs1N02SSIgAAjy7A0Wx12jmUAhzDs0nr+rTrYRRW+NeD4TOQLFnh9UbniZKodUgSFdVaegx397AUd7+vFFO02NTtLV+EIbmtqeIm+epnYv7FRbY6Pa1PtWsfn1WO25tnFO6aV9aqN5Zvm9AeXMcnFdKnJp/MJOd2K9LEsJm/eiUjdGCiSaDLuYRGS0NEPn033WZhVK0Z64rlRjVZdjvew6G4s3bHxMeG+A02ntmCjKfXLjr9gRXb1h1xoHzeSXKt/N8b1O+lPjoot+MLzHTPOsXs53Dn+psyKpj/xVqf//4uvCDLYgmJSWSS54SW9yaAu9GtofhAmHNMPtl3JLUKGM0cMbiVnTG10pv2Rb2AAp6pXRCAFosAxz0FuJcy1iXWy8YgNIHwpGOj82+iFKJGiAqJlbkWa8MK8+h3Uj0nYcb87auRVLC0MBy3WDupnqXe7teOxGtz1jXmHtjMTJrTtXUPUDQ81VatwG0aySTBSJGe9MbCz31DlFkVKbvvSSBdak3hQtmg4tEgNFWu5HLbN9w3Ln/Ye0Sk/gW8Hp9WPV5AD5ieeNek9S1tTZCkUTl7Auqa+YVz6iCR/atTt2qulGvWEpknXhPa991Tk5a4rT9dMtjO/ovVixc/07eyvP8Fz0vKRGPYJ31KNatjPgie/xoeM538gOb9msUys0xRlSG8VzdT1U72YgPxVd9AZZUQ9pzoIMqIYWnMnultnod5Q12vtgqKz0s3SievySGGpDftLjm34jOSkil0y4XjAOaPXNoShF38300zlDeOxNijGYkmBtCkbwmOlgRhSjG+wyPmXqzXVX0W1nNVfP4RwFM6egJu6HmkDmo5g86Ud2uAThQPju7ANCGz0qRS1m3D+8En4cDOKmpNg2qMuGXg4Pn+8Xe1G/+ITYlrIEEx6L4+/IuzSrQMyEes3na7K6q1xj9buFL1YcW3Jlmb2H99KkmbTl7XNG1Le9QOIVuBrjPLV8/HZqMD647YRAFvOcoWQhV2GYGmnbQqirU2XmlzWRP+KRhSQOWm9OKqO2xrGdjoYWGI8Xf9NVMWuEGde07bvtXjbwlLiAVUvLdRwHjj0rIYO6SXLOlE0sd1VTE75UBZqecnIJj8hcCPy+xeuT+nTHgvYViWpssQIjRTEBSaS6WAj8/oYY4x2dRR6PVaTcsJ4MU2MO/4ymIxoIRFKPXo/wSyFT8V4mKkJHr/8YnVQyWvwyeh16vmu7Mz/yvMgPAieYjvaj15Yf+PYcc2sYBKHrhqEXfp2ODmW1PezEsR1GrhvPfc93PE8dDjw7mPkxoslx43CmDg7+d4E5HQS+Y8e0ZR7P3XnsqtNubEcwxsPXeD5zQjeAgPOFtpUTzX07DmPfm8/DMHRC40IY2LMZFrx54LveLPbJkjNp0joTAonAdRAHkRdGQaShCKLQ9uGeAw/hbzQ/K6VzykMW2W7oRrE7d3wgp63x3ciOvSiMfNf15rHnQcrJE2wrYu7ENkCfRY7n+7PQwHpOxIWny1ZQFAPgwHNwk3RDsZIT+zM78uM4BlpR4LrRkZy+AASE7c3mztzDTt/x9Q0FTithFrmAWVkyrMetAMf1bSeCAKCH+Lom4cBRzGF/HiAWEEWAxp05ro6PWTRz7IjCxo292I0cio+0fypC/DlHp7yZ79kzdcrzPcT7169f/wPeoiv3
But Nathan, that node doesn’t go anywhere. Yeah I know. I didn’t know how to do it
What are the proper names for these? Presumably someone made some up
I like Katja so I’m baised but I really liked her EAG talk. I changed a number of views and also it encouraged me that I could think about this.