All of Andy_McKenzie's Comments + Replies

 identity is irretrievably lost when the brain activity stops 

My point here is that this is a very strong claim about neuroscience -- that molecular structure doesn’t encode identity/memories. 

2Cameron Berg
I think you might be taking the quotation a bit too literally—we are of course not literally advocating for the death of scientists, but rather highlighting that many of the largest historical scientific innovations have been systematically rejected by one's contemporaries in their field.  Agree that scientists change their minds and can be convinced by sufficient evidence, especially within specific paradigms. I think the thornier problem that Kuhn and others have pointed out is that the introduction of new paradigms into a field are very challenging to evaluate for those who are already steeped in an existing paradigm, which tends to cause these people to reject, ridicule, etc those with strong intuitions for new paradigms, even when they demonstrate themselves in hindsight to be more powerful or explanatory than existing ones. 

Thanks for the clarification and your thoughts. In my view, the question is to what extent the polymer gel embedding is helpful from the perspective of maintaining morphomolecular structure, so that it is worth the trade-off of removing the lipids, which could potentially also have information content. https://brainpreservation.github.io/Biomolecules#how-lipid-biomolecules-in-cell-membranes-could-affect-ion-flow

You are in good company in thinking that clearing and embedding the tissue in a hydrogel is the best approach. Others with expertise in the area ha... (read more)

4Nathan Helm-Burger
Yes, more research definitely seems like the best answer to me too.  I'm hopeful that some of these open questions about the relative efficacy of various techniques, and about how to image and then process the data, will be resolved before I'm on my deathbed and need to make a final will / living will. 

Thanks for your interest! 

Does OBP plan to eventually expand their services outside the USA?

In terms of our staff traveling to other locations to do the preservation procedure, unfortunately not in the immediate future. We don't have the funding for this right now. 

And how much would it cost if you didn’t subsidize it?

There are so many factors. It depends a lot on where in the world we are talking about. If we are talking about someone who legally dies locally in Salem, perhaps a minimal estimated budget would be (off the top of my head, unoffici... (read more)

We discuss the possibility of fluid preservation after tissue clearing in our article: 

An alternative option is to perform tissue clearing prior to long-term preservation (118). This would remove the lipids in the brain, but offer several advantages, including repeated non-invasive imaging, and potentially reduced oxidative damage over time (119).

And also in our fluid preservation article we have a whole section on it. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11058410/#S7

I'm not sure why this option is much more robust that formaldehyde fixation a... (read more)

5Nathan Helm-Burger
Ah, I mean, robust as in physically robust. The plastic gel that results is quite sturdy. The embedded clarified tissue is relatively easily handled without damaging it. Just my impression from having worked with fresh human brain tissue (delicate), cryopreserved-only (also delicate, in some ways even more so since the cold causes it to stick to tools and surfaces when handling it. In other ways, such as the glassy ice/preservative mixture being firm, not so delicate), formaldehyde or paraformaldehyde fixed (sturdier, relatively firm instead of mooshy, relatively easy to slice cleanly and deliberately), gel embedded (quite sturdy), hard resin embedded (extremely sturdy, appropriate for super thin slicing for electron microscopy). I realize that reconstruction is a different problem from preservation, but if your goal (like mine is) is to preserve in such a way that you facilitate reconstruction, this seems like a big win. Imagine, for instance, you must evaluate whether the reconstruction technology is sufficiently advanced to attempt to do a reconstruction for a particular brain. The threshold for 'good enough tech to make an attempt' is much lower if the brain is already prepared in a way (e.g. 1 cm thick slices of clarified stabilized tissue) such that it can easily be non-destructively imaged now, and then imaged again later if need be. A cryopreserved brain, on the other hand, needs to wait until you are very sure that the tech is good enough to handle it. Subjectively, I would feel quite reassured by my chosen method of brain preservation being: 1) stable at room temperature, and relatively robust to less-than-exquisitely-careful handling 2) ready to be non-destructively and relatively cheaply imaged, such that it would be an easy call to make a first attempt 3) preserved in a form that would facilitate reconstruction and multi-protein labeling

I can't speak for Adele, but here is one somewhat recent article by neuroscientists discussing memory storage mechanisms: https://bmcbiol.biomedcentral.com/articles/10.1186/s12915-016-0261-6

DNA is discussed as one possible storage mechanism in the context of epigenetic alterations to neurons. See the section by Andrii Rudenko and Li-Huei Tsai.

2jmh
Thanks. Just took a quick glance as the abstract but looks interesting. Will have something to read while waiting at the airport for a flight tomorrow.

This is an important question. While I don't have a full answer, my impression is that yes, it seems to preserve the important information present in DNA. More information here: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11058410/#S4.4

Thanks for the comment. I'm definitely not assuming that p(success) would be a monocausal explanation. I'm mostly presenting this data to give evidence against that assumption, because people frequently make statements such as "of course almost nobody wants cryonics, they don't expect it will work". 

I also agree that "is being revived good in expectation / good with what probability" is another common concern. Personally, I think niplav has some good analysis of net-negative revival scenarios: https://niplav.site/considerations_on_cryonics.html

Btw, ac... (read more)

Very high-effort, comprehensive post. Any interest in making some of your predictions into markets on Manifold or some other prediction market website? Might help get some quantifications. 

A simple solution is to just make doctors/hospitals liable for harm which occurs under their watch, period. Do not give them an out involving performative tests which don’t actually reduce harm, or the like. If doctors/hospitals are just generally liable for harm, then they’re incentivized to actually reduce it.

Can you explain more what you actually mean by this? Do you mean if someone comes into the hospital and dies, the doctors are responsible, regardless of why they died? If you mean that we figure out whether the doctors are responsible for whether th... (read more)

4johnswentworth
If someone comes into the hospital and dies, the doctors are responsible, regardless of why they died. Same for injuries, sickness, etc. That would be the simplest and purest version, though it would probably be expensive. One could maybe adjust in some ways, e.g. the doctors' responsibility is lessened if the person had some very legible problem from the start (before they showed up to the doctor/hospital), or the doctor's responsibility is always lessened by some baseline amount corresponding to the (age-adjusted) background rate of death/injury/sickness. But the key part is that a court typically does not ask whether the death/injury/sickness is the doctor's fault. They just ask whether it occurred under the doctor/hospital's watch at all.

Out of curiosity, what makes you think that the initial freezing process causes too much information loss? 

I agree with most of this post, but it doesn’t seem to address the possibility of whole brain emulation. However, many/(?most) would argue this is unlikely to play a major role because AGI will come first.

Thanks so much for putting this together Mati! If people are interested in cryonics/brain preservation and would like to learn about (my perspective on) the field from a research perspective, please feel free to reach out to me: https://andrewtmckenzie.com/

I also have some external links/essays available here: https://brainpreservation.github.io/

It seems to me like your model is not necessarily taking into account technical debt sufficiently enough. https://neurobiology.substack.com/p/technical-debt-probably-the-main-roadblack-in-applying-machine-learning-to-medicine

It seems to me like this is the main thing that will slow down the extent to which foundation models can consistently beat newly trained specialized models.

Anecdotally, I know several people who don’t like to use chatgpt because its training cuts off in 2021. This seems like a form of technical debt.

I guess it depends on how easily ada... (read more)

Sounds good, can't find your email address, DM'd you. 

Those sound good to me! I donated to your charity (the Animal Welfare Fund) to finalize it. Lmk if you want me to email you the receipt. Here's the manifold market: 

Bet

Andy will donate $50 to a charity of Daniel's choice now.

If, by January 2027, there is not a report from a reputable source confirming that at least three companies, that would previously have relied upon programmers, and meet a defined level of success, are being run without the need for human programmers, due to the independent capabilities of an AI developed by Op... (read more)

2Daniel Kokotajlo
Sounds good, thank you! Emailing the receipt would be nice.

Sounds good, I'm happy with that arrangement once we get these details figured out. 

Regarding the human programmer formality, it seems like business owners would have to be really incompetent for this to be a factor. Plenty of managers have coding experience. If the programmers aren't doing anything useful then they will be let go or new companies will start that don't have them. They are a huge expense. I'm inclined to not include this since it's an ambiguity that seems implausible to me. 

Regarding the potential ban by the government, I wasn't r... (read more)

5Daniel Kokotajlo
How about this: --Re the first grey area: We rule in your favor here. --Re the second grey area: You decide, in 2027, based on your own best judgment, whether or not it would have happened absent regulation. I can disagree with your judgment, but I still have to agree that you won the bet (if you rule in your favor).

Understandable. How about this? 

Bet

Andy will donate $50 to a charity of Daniel's choice now.

If, by January 2027, there is not a report from a reputable source confirming that at least three companies, that would previously have relied upon programmers, and meet a defined level of success, are being run without the need for human programmers, due to the independent capabilities of an AI developed by OpenAI or another AI organization, then Daniel will donate $100, adjusted for inflation as of June 2023, to a charity of Andy's choice.

Terms

Reputable Sourc... (read more)

2Daniel Kokotajlo
Given your lack of disposable money I think this would be a bad deal for you, and as for me, it is sorta borderline (my credence that the bet will resolve in your favor is something like 40%?) but sure, let's do it. As for what charity to donate to, how about Animal Welfare Fund | Effective Altruism Funds. Thanks for working out all these details! Here are some grey area cases we should work out: --What if there is a human programmer managing the whole setup, but they are basically a formality? Like, the company does technically have programmers on staff but the programmers basically just form an interface between the company and ChatGPT and theoretically if the managers of the company were willing to spend a month learning how to talk to ChatGPT effectively they could fire the human programmers? --What if it's clear that the reason you are winning the bet is that the government has stepped in to ban the relevant sorts of AI?

I’m wondering if we could make this into a bet. If by remote workers we include programmers, then I’d be willing to bet that GPT-5/6, depending upon what that means (might be easier to say the top LLMs or other models trained by anyone by 2026?) will not be able to replace them.

8Daniel Kokotajlo
I've made several bets like this in the past, but it's a bit frustrating since I don't stand to gain anything by winning -- by the time I win the bet, we are well into the singularity & there isn't much for me to do with the money anymore. What are the terms you have in mind? We could do the thing where you give me money now, and I give it back with interest later.  

These curves are due to temporary plateaus, not permanent ones. Moore's law is an example of a constraint that seems likely to plateau. I'm talking about takeoff speeds, not eventual capabilities with no resource limitations, which I agree would be quite high and I have little idea of how to estimate (there will probably still be some constraints, like within-system communication constraints). 

1Decaeneus
Understood, and agreed, but I'm still left wondering about my question as it pertains to the first sigmoidal curve that shows STEM-capable AGI. Not trying to be nitpicky, just wondering how we should reason about the likelihood that the plateau of that first curve is not already far above the current limit of human capability. A reason to think so may be something to do with irreducible complexity making things very hard for us at around the same level that it would make them hard for a (first-gen) AGI. But a reason to think the opposite would be that we have line of sight to a bunch of amazing tech already, it's just a question of allocating the resources to support sufficiently many smart people working out the details. Another reason to think the opposite is that having a system that's (in some sense) directly optimized to be intelligent might just have a plateau drawn from a higher-meaned distribution than one that's optimized for fitness, and develops intelligence as a useful tool in that direction, since the pressure-on-intelligence for that sort of caps out at whatever it takes to dominate your immediate environment.

Does anyone know of any AI-related predictions by Hinton? 

Here's the only one I know of - "People should stop training radiologists now. It's just completely obvious within five years deep learning is going to do better than radiologists because it can get a lot more experience. And it might be ten years but we got plenty of radiologists already." - 2016, slightly paraphrased 

This seems like still a testable prediction - by November 2026, radiologists should be completely replaceable by deep learning methods, at least other than regulatory requirements for trained physicians. 

4Ilio
This is indeed an interesting losing* bet. He was mostly right on the technical side (yes deep learning now do better than the average radiologist on many tasks). He was completely wrong on the societal impact (no we still need to train radiologists). This was the same story with ophthalmologists when deep learning significantly shorten the time needed to perform part of their job: they just spent the saved time on doing more. *16+5=21, not 26 😉
6bvbvbvbvbvbvbvbvbvbvbv
Fyi actually radiology is not mostly looking at pictures but doing imagery-guided surgery (for example embolisation) which is significantly harder to automate. Same for family octors : it's not just following guidelines and renewing scripts but a good part is physical examination. I agree that AI can do a lot of what happens in medicine though.

Thanks! I agree with you about all sorts of AI alignment essays being interesting and seemingly useful. My question was more about how to measure the net rate of AI safety research progress. But I agree with you that an/your expert inside view of how insights are accumulating is a reasonable metric. I also agree with you that the acceptance of TAI x-risk in the ML community as a real thing is useful and that - while I am slightly worried about the risk of overshooting, like Scott Alexander describes - this situation seems to be generally improving. 

Re... (read more)

4Steven Byrnes
Oh, I somehow missed that your original question was about takeoff speeds. When you wrote “algorithmic insights…will lead to dramatically faster AI development”, I misread it as “algorithmic insights…will lead to dramatically more powerful AIs”. Oops. Anyway, takeoff speeds are off-topic for this post, so I won’t comment on them, sorry. :)
1sanxiyn
I would not describe development of deep learning as discontinuous, but I would describe it as fast. As far as I can tell, development of deep learning happened by accumulation of many small improvements over time, sometimes humorously described as graduate student descent (better initialization, better activation function, better optimizer, better architecture, better regularization, etc.). It seems possible or even probable that brain-inspired RL could follow the similar trajectory once it took off, absent interventions like changes to open publishing norm.

Good essay! Two questions if you have a moment: 

1. Can you flesh out your view of how the community is making "slow but steady progress right now on getting ready"? In my view, much of the AI safety community seems to be doing things that have unclear safety value to me, like (a) coordinating a pause in model training that seems likely to me to make things less safe if implemented (because of leading to algorithmic and hardware overhangs) or (b) converting to capabilities work (quite common, seems like an occupational hazard for someone with initially... (read more)

Steven Byrnes*Ω5146

Can you flesh out your view of how the community is making "slow but steady progress right now on getting ready"?

  • I finished writing this less than a year ago, and it seems to be meaningfully impacting a number of people’s thinking, hopefully for the better. I personally feel strongly like I’m making progress on a worthwhile project and would like lots more time to carry it through, and if it doesn’t work out I have others in the pipeline. I continue to have ideas at a regular clip that I think are both important and obvious-in-hindsight, and to notice new
... (read more)

I didn't realize you had put so much time into estimating take-off speeds. I think this is a really good idea. 

This seems substantially slower than the implicit take-off speed estimates of Eliezer, but maybe I'm missing something. 

I think the amount of time you described is probably shorter than I would guess. But I haven't put nearly as much time into it as you have. In the future, I'd like to. 

Still, my guess is that this amount of time is enough that there are multiple competing groups, rather than only one. So it seems to me like there w... (read more)

6Daniel Kokotajlo
It is substantially slower than the takeoff speed estimates of Eliezer, yes. I'm definitely disagreeing with Eliezer on this point. But as far as I can tell my view is closer to Eliezer's than to Hanson's, at least in upshot. (I'm a bit confused about this--IIRC Hanson also said somewhere that takeoff would last only a couple of years? Then why is he so confident it'll be so broadly distributed, why does he think property rights will be respected throughout, why does he think humans will be able to retire peacefully, etc.?) I also think it's plausible that there will be multiple competing groups rather than one singleton AI, though not more than 80% plausible; I can easily imagine it just being one singleton. I think that even if there are multiple competing groups, however, they are very likely to coordinate to disempower humans. From the perspective of the humans it'll be as if they are an AI singleton, even though from the perspective of the AIs it'll be some interesting multipolar conflict (that eventually ends with some negotiated peaceful settlement, I imagine) After all, this is what happened historically with colonialism. Colonial powers (and individuals within conquistador expeditions) were constantly fighting each other. 

Thanks for writing this up as a shorter summary Rob. Thanks also for engaging with people who disagree with you over the years. 

Here's my main area of disagreement: 

General intelligence is very powerful, and once we can build it at all, STEM-capable artificial general intelligence (AGI) is likely to vastly outperform human intelligence immediately (or very quickly).

I don't think this is likely to be true. Perhaps it is true of some cognitive architectures, but not for the connectionist architectures that are the only known examples of human-like ... (read more)

1Decaeneus
Hi Andy - how are you gauging the likely relative proportions of AI capability sigmoidal curves relative to the current ceiling of human capability? Unless I'm misreading your position, it seems like you are presuming that the sigmoidal curves will (at least initially) top out at a level that is on the same order as human capabilities. What informs this prior? Due to the very different nature of our structural limitations (i.e. a brain that's not too big for a mother's hips to safely carry and deliver, specific energetic constraints, the not-very-precisely-directed nature of the evolutionary process) vs an AGI's system's limitations (which are simply different) it's totally unclear to me why we should expect the AGI's plateaus to be found at close-to-human levels.

Agreed. A common failure mode in these discussions is to treat intelligence as equivalent to technological progress, instead of as an input to technological progress. 

Yes, in five years we will likely have AIs that will be able to tell us exactly where it would be optimal to allocate our scientific research budget. Notably, that does not mean that all current systemic obstacles to efficient allocation of scarce resources will vanish. There will still be the same perverse incentive structure for funding allocated to scientific progress as there is toda... (read more)

I can see how both Yudkowsky's and Hanson's arguments can be problematic because they either assume fast or slow takeoff scenarios, respectively, and then nearly everything follows from that. So I can imagine why you'd disagree with every one of Hanson's paragraphs based on that. If you think there's something he said that is uncorrelated with the takeoff speed disagreement, I might be interested, but I don't agree with Hanson about everything either, so I'm mainly only interested if it's also central to AI x-risk. I don't want you to waste your time. ... (read more)

4Daniel Kokotajlo
I think there are probably disagreements I have with Hanson that don't boil down to takeoff speeds disagreements, but I'm not sure. I'd have to reread the article again to find out. To be clear, I definitely don't expect takeoff to take hours or days. Quantitatively I expect something like what takeoffspeeds.com says when you input the values of the variables I mentioned above. So, eyeballing it, it looks like it takes slightly more than 3 years to go from 20% R&D automation to 100% R&D automation, and then to go from 100% R&D automation to "starting to approach the fundamental physical limits of how smart minds running on ordinary human supercomputers can be" in about 6 months, during which period about 8 OOMs of algorithmic efficiency is crossed. To be clear I don't take that second bit very seriously at all, I think this takeoffspeeds.com model is much better as a model of pre-AGI takeoff than of post-AGI takeoff. But I do think that we'll probably go from AGI to superintelligent AGI in less than six months. How long it takes to get to nanotech or (name your favorite cool sci-fi technology) is less clear to me, but I expect it to be closer to one year than ten, and possibly more like one month. I would love to discuss this more & read attempts to estimate these quantities.  

To clarify, when I mentioned growth curves, I wasn't talking about timelines, but rather takeoff speeds. 

In my view, rather than indefinite exponential growth based on exploiting a single resource, real-world growth follows sigmoidal curves, eventually plateauing. In the case of a hypothetical AI at a human intelligence level, it would face constraints on its resources allowing it to improve, such as bandwidth, capital, skills, private knowledge, energy, space, robotic manipulation capabilities, material inputs, cooling requirements, legal and regulat... (read more)

I too was talking about takeoff speeds. The website I linked to is takeoffspeeds.com.

Me & the other LWers you criticize do not expect indefinite exponential growth based on exploiting a single resource; we are well aware that real-world growth follows sigmoidal curves. We are well aware of those constraints and considerations and are attempting to model them with things like the model underlying takeoffspeeds.com + various other arguments, scenario exercises, etc.

I agree that much of LW has moved past the foom argument and is solidly on Eliezers side r... (read more)

Here's a nice recent summary by Mitchell Porter, in a comment on Robin Hanson's recent article (can't directly link to the actual comment unfortunately): 

Robin considers many scenarios. But his bottom line is that, even as various transhuman and posthuman transformations occur, societies of intelligent beings will almost always outweigh individual intelligent beings in power; and so the best ways to reduce risks associated with new intelligences, are socially mediated methods like rule of law, the free market (in which one is free to compete, but also

... (read more)
6Daniel Kokotajlo
Wait, how is it not how growth curves have worked historically? I think my position, which is roughly what you get when you go to this website and set the training requirements parameter to 1e30 and software returns to 2.5, is quite consistent with how growth has been historically, as depicted e.g. How Roodman's GWP model translates to TAI timelines - LessWrong (Also I resent the implication that SIAI/MIRI hasn't tended to directly engage with those arguments. The FOOM debate + lots of LW ink has been spilled over it + the arguments were pretty weak anyway & got more attention than they deserved)

AIs can potentially trade with humans too though, that's the whole point of the post. 

Especially if the AI's have architectures/values that are human brain-like and/or if humans have access to AI tools, intelligence augmentation, and/or whole brain emulation. 

Also, it's not clear why AIs will find it easier to coordinate with one another than humans and humans or humans and AIs. Coordination is hard for game theoretic reasons. 

These are all standard points, I'm not saying anything new here. 

When you write "the AI" throughout this essay, it seems like there is an implicit assumption that there is a singleton AI in charge of the world. Given that assumption, I agree with you. But if that assumption is wrong, then I would disagree with you.  And I think the assumption is pretty unlikely. 

No need to relitigate this core issue everywhere, just thought this might be useful to point out. 

8quetzal_rainbow
What's the difference? Multiple AIs can agree to split the universe and gains from disassembling biosphere/building Dyson sphere/whatever and forget to include humanity in negotiations. Unless preferences of AIs are diametrically opposed, they can trade.
7trevor
Why is the assumption of a unilateral AI unlikely?  That's a very important crux, big if true, and it would be worth figuring out to explain it to people in fewer words so that more people will collide with it. In this post, So8res explicity states: This is well in line with the principle of instrumental convergence,  and instrumental convergence seems to be a prerequisite for creating substantial amounts of intelligence. What we have right now is not-very-substantial amounts of intelligence, and hopefully we will only have not-very-substantial amounts of intelligence for a very long time, until we can figure out some difficult problems. But the problem is that a firm might develop substantial amounts of intelligence sooner instead of later.
Answer by Andy_McKenzie104

I agree this is a very important point and line of research. This is how humans deal with sociopaths, after all.

Here’s me asking a similar question and Rob Bensinger’s response: https://www.lesswrong.com/posts/LLRtjkvh9AackwuNB/on-a-list-of-lethalities?commentId=J42Fh7Sc53zNzDWCd

One potential wrinkle is that in a very fast take off world AI’s could potentially coordinate very well because they would basically be the same, or close branches of the same AI.

"Science advances one funeral at a time" -> this seems to be both generally not true as well as being a harmful meme (because it is a common argument used to argue against life extension research).

https://www.lesswrong.com/posts/fsSoAMsntpsmrEC6a/does-blind-review-slow-down-science 

Interesting, thanks. All makes sense and no need to apologize. I just like it when people write/think about schizophrenia and want to encourage it, even as a side project. IMO, it's a very important thing for our society to think about. 

A lot of the quotes do find decreased connectivity, but some of them find increased connectivity between certain regions. It makes me think that there's a probability there might be something more complicated than just "increased or decreased", but rather specific types of connections. But that's just a guess, and I think an explanation across all cortical connections is more parsimonious and therefore more likely a priori. 

Of your criteria of "things to explain", here are some thoughts: 

4.1 The onset of schizophrenia is typically in the late-tee... (read more)

4Steven Byrnes
Update: there’s some discussion of antipsychotics in my follow-up post: Model of psychosis, take 2  :)
4Steven Byrnes
LOL I’m not focused on this at all. I think I’ve spent a whopping four days of my life thinking hard about schizophrenia—one day in 2021 that didn’t go anywhere, one day last summer where I read a bunch of papers and thought of this hypothesis and felt pretty good about it and then moved on to other things, then one more day like a week later to research and write the blindness + schizophrenia post, and yesterday to write this post. Schizophrenia not a significant personal or professional interest of mine. I am very impressed with myself for fooling you. Or maybe you’re just being polite. :) (Understanding schizophrenia is plausibly indirectly helpful for my professional interests, for various reasons. Also, I have a rule-of-thumb that if I can write a decent blog post in four hours, I should just do it, often it leads to unexpected good things!) Yeah the “things to explain” could have been more accurately titled “aspects of schizophrenia that I can easily think of right now, from either off the top of my head or skimming the wikipedia article”. :-P I think the cognitive deficits are very straightforwardly and naturally predicted by my hypothesis. I wrote something about nicotine but a different commenter said that what I wrote was flagrantly wrong. (I put a warning in the OP.) Guess I need to think about that more. Honestly, I don’t have a great understanding of what nicotine does to the brain in the first place. Something something acetylcholine :-P I haven’t looked into antipsychotics / neuroleptics, and agree that doing so would be an obvious next step, and indeed maybe I should have done it before posting this. Sorry. I’ll put it on my to-do list.

Interesting theory and very important topic. 

I think the best data source here is probably neuroimaging. Here's a recent review: https://www.frontiersin.org/articles/10.3389/fnins.2022.1042814/full. Here are some quotes from that: 

For functional studies, be they fluorodeoxyglucose positron emission tomography (FDG PET), rs-fMRI, task-based fMRI, diffusion tensor imaging (DTI) or MEG there generally is hypoactivation and disconnection between brain regions. ...

Histologically this gray matter reduction is accompanied by dendritic and synaptic densi

... (read more)
2Steven Byrnes
Thanks! Sorta. I kinda feel like my hypothesis (or something very close to it) really is an elegant explanation for everything about schizophrenia. Of course, I don’t know that—among other things, I don’t know everything about schizophrenia (obviously). I was writing this partly in the hopes that you or other commenters would tell me about aspects of schizophrenia that my hypothesis can’t explain, or contradicts, if such aspects exist. And then I can drop that hypothesis and find something better to believe. :) Do you think that anything in your excerpt contradicts my hypothesis? Seems to be almost entirely decreases in connectivity, right? That said, I don’t put much stock in functional connectivity comparisons anyway—e.g. you don’t really know if those regions are talking directly to each other vs correlated for some other reason, and even leaving that aside, you can’t disentangle what control-vs-SCZ difference is caused by direct physical connectivity differences versus “when schizophrenics are hanging out in the fMRI machine their wandering minds tend to be thinking about different things than when the control group people are hanging out in the fMRI machine”, or whatever. [I very generally find it quite hard to learn anything useful from neuroimaging data, compared to most other types of neuroscience data / evidence. But maybe that’s just me. You do you. :-) ]

A quote I find relevant: 

“A happy life is impossible, the highest thing that man can aspire to is a heroic life; such as a man lives, who is always fighting against unequal odds for the good of others; and wins in the end without any thanks. After the battle is over, he stands like the Prince in the re corvo of Gozzi, with dignity and nobility in his eyes, but turned to stone. His memory remains, and will be reverenced as a hero's; his will, that has been mortified all his life by toiling and struggling, by evil payment and ingratitude, is absorbed into Nirvana.” - Arthur Schopenhauer

Good point. 

I know your question was probably just rhetorical, but to answer it regardless -- I was confused in part because it would have made sense to me if he had said it would "better" if AGI timelines were short. 

Lots of people want short AGI timelines because they think the alignment problem will be easy or otherwise aren't concerned about it and they want the perceived benefits of AGI for themselves/their family and friends/humanity (eg eliminating disease, eliminating involuntary death, abundance, etc). And he could have just said "better... (read more)

One of the main counterarguments here is that the existence of multiple AGIs allows them to compete with one another in ways that could benefit humanity. E.g. policing one another to ensure alignment of the AGI community with human interests. Of course, whether this actually would outweigh your concern in practice is highly uncertain and depends on a lot of implementation details. 

You're right that the operative word in "seems more likely" is "seems"! I used the word "seems" because I find this whole topic really confusing and I have a lot of uncertainty. 

It sounds like there may be a concern that I am using the absurdity heuristic or something similar against the idea of fast take-off and associated AI apocalypse. Just to be clear, I most certainly do not buy absurdity heuristic arguments in this space, would not use them, and find them extremely annoying. We've never seen anything like AI before, so our intuition (which might suggest that the situation seems absurd) is liable to be very wrong. 

4Thane Ruthenis
Oh, I think I should've made clearer that I wasn't aiming that rant at you specifically. Just outlining my general impression of how the two views feel socially.

A few comments: 

  1. A lot of slow takeoff, gradual capabilities ramp-up, multipolar AGI world type of thinking. Personally, I agree with him this sort of scenario seems both more desirable and more likely. But this seems to be his biggest area of disagreement with many others here. 
  2. The biggest surprise to me was when he said that he thought short timelines were safer than long timelines. The reason for that is not obvious to me. Maybe something to do with contingent geopolitics. 
  3. Doesn't seem great to dismiss people's views based on psychologizing about them. But, these are off-the-cuff remarks, held to a lower standard than writing. 
lc*1411

The biggest surprise to me was when he said that he thought short timelines were safer than long timelines. The reason for that is not obvious to me. Maybe something to do with contingent geopolitics.

What do you expect him to say? "Yeah, longer timelines and consolidated AGI development efforts are great, I'm shorting your life expectancies as we speak"? The only way you can be a Sam Altman is by convincing yourself that nuclear proliferation makes the world safer.

6Thane Ruthenis
I think the operative word in "seems more likely" here is "seems". It seems like a more sophisticated, more realistic, more modern and satisfyingly nuanced view, compared to "the very first AGI we train explodes like a nuclear bomb and unilaterally sets the atmosphere on fire, killing everyone instantly". The latter seems like an old view, a boringly simplistic retrofuturistic plot. It feels like there's a relationship between these two scenarios, and that the latter one is a rough first-order approximation someone lifted out of e. g. The Terminator to get people interested in the whole "AI apocalypse" idea at the onset of it all. Then we gained a better understanding, sketched out detailed possibilities that take into account how AI and AI research actually work in practice, and refined that rough scenario. As the result, we got that picture of a slower multipolar catastrophe. A pleasingly complicated view! One that respectfully takes into account all of these complicated systems of society and stuff. It sure feels like how these things work in real life! "It's not like the AI wakes up and decides to be evil," perish the thought. That seeming has very little to do with reality. The unilateral-explosion isn't the old, outdated scenario — it's simply a different scenarios that's operating on a different model of how intelligence explosions proceed. And as far as its proponents are concerned, its arguments haven't been overturned at all, and nothing about how DL works rules it out. But it sure seems like the rough naive view that the Real Experts have grown out of a while ago; and that those who refuse to update simply haven't done that growing-up, haven't realized there's a world outside their chosen field with all these Complicated Factors you need to take into account. It makes it pretty hard to argue against. It's so low-status. ... At least, that's how that argument feels to me, on a social level. (Edit: Uh, to be clear, I'm not saying that there's no other

Got it. To avoid derailing with this object level question, I’ll just say that I think it seems helpful to be explicit about takeoff speeds in macrostrategy discussions. Ideally, specifying how different strategies work over distributions of takeoff speeds.

Thanks for this post. I agree with you that AI macrostrategy is extremely important and relatively neglected. 

However, I'm having some trouble understanding your specific world model. Most concretely: can you link to or explain what your definition of "AGI" is? 

Overall, I expect alignment outcomes to be significantly if not primarily determined by the quality of the "last mile" work done by the first AGI developer and other actors in close cooperation with them in the ~2 years prior to the development of AGI.

This makes me think that in your world... (read more)

3NickGabs
My takeoff speeds are on the somewhat faster end, probably ~a year or two from “we basically don’t have crazy systems” to “AI (or whoever controls AI) controls the world” EDIT: After further reflection, I no longer endorse this. I would now put 90% CI from 6 months to 15 years with median around 3.5 years. I still think fast takeoff is plausible but now think pretty slow is also plausible and overall more likely.

OK, I get your point now better, thanks for clarifying -- and I agree with it. 

In our current society, even if dogs could talk, I bet that we wouldn't allow humans to trade (or at least anywhere close to "free" trade) with them, due to concerns for exploitation. 

I quoted "And if she isn't a good girl, we genetically engineer and manufacture (ie. breed) an ex-wolf who is a good girl."

If genetic engineering a new animal would satisfy human goals, then this would imply that they don't care about their pet's preferences as individuals. 

4SarahNibs
No, it wouldn't imply that, at all. One can very easily care about something's preference as an individual and work to make a new class of thing which will be more useful than the class of thing that individual belongs to.

At the end of the day, no matter how many millions her trainer earns, Lassie just gets a biscuit & ear scritches for being such a good girl. And if she isn't a good girl, we genetically engineer and manufacture (ie. breed) an ex-wolf who is a good girl.

I don't think it's accurate to claim that humans don't care about their pets' preferences as individuals and try to satisfy them. 

To point out one reason that I think this, there are huge markets for pet welfare. There are even animal psychiatrists and there are longevity companies for pets

I'... (read more)

gwern*5135

I also don't think that 'trade' necessarily captures the right dynamic. I think it's more like communism in the sense that families are often communist. But I also don't think that your comment, which sidesteps this important aspect of human-animal relations, is the whole story.

Indeed, 'trade' is not the whole story; it is none of the story - my point is that the human-animal relations, by design, sidestep and exclude trade completely from their story.

Now, how good that actual story is for dogs, or more accurately for the AI/human analogy, wolves, one c... (read more)

3lc
But he didn't say that!

Thanks for this good post. A meta-level observation is that people are grasping at straws like this is evidence that our knowledge of the causes of schizophrenia is quite limited. 

“One day, one of the AGI systems improves to the point where it unlocks a new technology that can reliably kill all humans, as well as destroying all of its AGI rivals. (E.g., molecular nanotechnology.) I predict that regardless of how well-behaved it's been up to that point, it uses the technology and takes over. Do you predict otherwise?”

I agree with this, given your assumptions. But this seems like a fast take off scenario, right? My main question wasn’t addressed — are we assuming a fast take off? I didn’t see that explicitly discussed.

My understanding... (read more)

2Rob Bensinger
I would define hard takeoff as "progress in cognitive ability from pretty-low-impact AI to astronomically high-impact AI is discontinuous, and fast in absolute terms". Unlocking a technology that lets you kill other powerful optimizers (e.g., nanotech) doesn't necessarily require fast or discontinuous improvements to systems' cognition. E.g., humans invented nuclear weapons just via accumulating knowledge over time; the invention wasn't caused by us surgically editing the human brain a few years prior to improve its reasoning. (Though software improvements like 'use scientific reasoning', centuries prior, were obviously necessary.)

Thanks for the write-up. I have very little knowledge in this field, but I'm confused on this point: 

> 34.  Coordination schemes between superintelligences are not things that humans can participate in (eg because humans can’t reason reliably about the code of superintelligences); a “multipolar” system of 20 superintelligences with different utility functions, plus humanity, has a natural and obvious equilibrium which looks like “the 20 superintelligences cooperate with each other but not with humanity”.

Yes. I am convinced that things like ‘oh

... (read more)
9Rob Bensinger
Suppose that many different actors have AGI systems; the systems have terminal goals like 'maximize paperclips', and these goals imply 'kill any optimizers that don't share my goals, if you find a way to do so without facing sufficiently-bad consequences' (because your EV is higher if there are fewer optimizers trying to push the universe in different directions than what you want). The systems nonetheless behave in prosocial ways, because they're weak and wouldn't win a conflict against humans. Instead, the AGI systems participate in a thriving global economy that includes humans as well as all the competing AGIs; and all parties accept the human-imposed legal environment, since nobody can just overthrow the humans. One day, one of the AGI systems improves to the point where it unlocks a new technology that can reliably kill all humans, as well as destroying all of its AGI rivals. (E.g., molecular nanotechnology.) I predict that regardless of how well-behaved it's been up to that point, it uses the technology and takes over. Do you predict otherwise? Alternative scenario: One day, one of the AGI systems unlocks a new technology that can reliably kill all humans, but it isn't strong enough to destroy rival AGI systems. In that case, by default I predict that it kills all humans and then carries on collaborating or competing with the other AGI systems in the new humanless equilibrium. Alternative scenario 2: The new technology can kill all AGI systems as well as all humans, but the AGI made a binding precommitment to not use such technologies (if it finds them) against any agents that (a) are smart enough to inspect its source code and confidently confirm that it has made this precommitment, and (b) have verifiably made the same binding precommitment. Some or all of the other AGI systems may meet this condition, but humans don't, so you get the "AGI systems coordinate, humans are left out" equilibrium Eliezer described. This seems like a likely outcome of multip
Answer by Andy_McKenzie190

It is so great you are interested in this area! Thank you. Here are a few options for cryonics-relevant research: 

- 21st Century Medicine: May be best to reach out to Brian Wowk (contact info here: https://pubmed.ncbi.nlm.nih.gov/25194588/) and/or Greg Fahy (possibly old contact info here: https://pubmed.ncbi.nlm.nih.gov/16706656/)

- Emil Kendziorra at Tomorrow Biostasis may know of opportunities. Contact info here: https://journals.plos.org/plosone/article/authors?id=10.1371/journal.pone.0244980

- Robert McIntyre at Nectome may know of opportunities. C... (read more)

But there’s also a significant utilitarian motivation - which is relevant here because utilitarianism doesn’t care about death for its own sake, as long as the dead are replaced by new people with equal welfare. Indeed, if our lives have diminishing marginal value over time (which seems hard to dispute if you’re taking our own preferences into account at all), and humanity can only support a fixed population size, utilitarianism actively prefers that older people die and are replaced.

I strongly disagree with this. I think the idea of human fungibility is f... (read more)

2meedstrom
To steelman it, maybe he's thinking of how it's commonly seen as a tragedy for a chicken to be alive for only one week, but killing it after some X years is not as much of a tragedy. Initially, this implied to me that the curve of "value of remaining alive' is higher in the beginning of a lifespan. But thinking about it, that's not the same curve as the curve of "value of being alive", which is lowest in the beginning. (If that's confusing, it helps to think of the one curve as the mirror image of the other, i.e. if value of being alive is high later, it means the value of remaining alive "in order to see the later parts of life" is higher early on.) It's also possible to view the value of being alive as a flat line, a positive constant, which could lead to his idea of human fungibility. But to use a different example, if you make me choose between five individuals living one year and one individual living five years, I prefer the latter... Same with two people dying at 25 vs. one dying at 50. Fewer people living longer is better. I can only see this working out if it's not a flat line: value of life increases with each year already lived.
Load More