Thank you for posting about this here so that you can get feedback, and so that other people can know how much people are doing this sort of thing (and by the same token it could be good for people who've already done this sort of thing to say so).
I have a bit of a sinking feeling reading your draft; I'll try to say concrete things about it, but I don't think I'll capture all of what's behind the feeling. I think part of the feeling is about, this just won't work.
Part of it is like, the email seems to come from a mindset that doesn't give weight to curiosity and serious investigation (what Tao does with his time).
I know that to you it isn't the most interesting problem to think about[1] but it really, actually is a very very important and urgent completely open problem. It isn’t simply a theoretical concern, if Demis’ predictions of 10 to 20 years to AGI are anywhere near correct, it will deeply affect you and your family (and everyone else).
I think there's a sort of violence or pushiness here that's anti-helpful. It doesn't acknowledge that Tao doesn't have good reason to trust your judgements about what's "very very important and urgent", and people who go around telling other people what things are "very very important and urgent" in contexts without trust in judgement are often trying to coerce people into doing things they don't want to do. It doesn't acknowledge that people aren't utility maximizing machines, but instead have momentum and joy and specialization and context. (Not to say that Tao doesn't deserve to be informed about the consequences of events happening in the world and his possible effect on those consequences, and not to say that stating beliefs is bad, and not to say that Tao might just be curious and learn about AI risk if it's shown to him in the right way.)
Another thing is the sources recommendation. The links should be to technical arguments about problems in AI alignment and technical descriptions of the overall problem, the sort of thing X-risk and AI-risk thinkers say to each other, not to material prepared with introductoriness in mind.
It is not once but twice that I have heard leaders of AI research orgs say they want you to work on AI alignment.
This is kind of weird and pushy. On the face of it, it looks like either you're confused and think that big high status people to you are also big high status people to Tao and therefore should be able to give him orders about what to work on, and that Tao is even the sort of entity that takes orders; or at least, it looks like you yourself are trying to take orders from big high status people, propagating perceived urgency from them to whoever else, without regard to individual agents's local/private information about what's good for them to do. Like, it looks like you got scared, flailed and grasped for whatever the high status people said they think might be cool, and then wanted to push that. (I'm being blunt here, but to be clear, if something like this is happening, that's very empathizable-with; I don't think "you're bad" or anything like that, and doing stuff that seems like it would have good consequences is generally good.)
If you are ever interested you can start by reading
This is sort of absurd: 1. if Tao were interested, he could likely have lots of conversations with competent AI alignment thinkers, which would be a much better use of his time, and 2. frankly, it seems like you're posturing as someone giving orders to Tao.
I agree with TekhneMakre...it comes across like an average looking unconfident person asking out a gorgeous celeb. Probably a friend approaching him is best, but an email can't hurt. I would get a few people together to work on it...my approach would be to represent truly who we are as a motivated group of people that has the desire to write this email to him by saying something like, "There's a great forum of AI interested and concerned folks that we are a part of, many of us on the younger side, and we fear for the future of humanity from misaligned AI a...
I was trying to rely on Tao’s trust in Demis's judgement, since he is an AI researcher. Mentioning Eliezer is mainly so he has someone to contact if he wants to get hired.
I wanted his thinking to be “this competent entity has spent some of his computational resources verifying that it is important to solve this problem, and now that I’m reminded of that I should also throw mine at it”.
Is he truly mostly interested in what he considers to be mentally stimulating? Not in improving the world, or in social nonsense, or guaranteeing that his family is completel...
I think that to a non-trivial extent, we have a limited supply of such efforts. The more times Terry has been contacted about this, the less likely he is to respond positively.
And to some extent I suppose this is true about reaching out to such figures more generally. Ie. maybe word gets out that we've been doing such outreach and by the time we contact John Doe, even if it's our first time contacting John Doe, we may have exhausted our supply of John Doe's patience.
So then, I don't think such an action is as low cost as it may seem. It costs more than the time it takes to write the email.
What makes more sense to me is to try to traverse through social networks and reach him that way. Figure out which nodes are close to him who he listens to. Note that they might be bloggers like Scott Alexander or someone like Dan Luu. From there think about which of those nodes make sense to pursue. Maybe one, maybe multiple. Then backtrack and think about how we can utilize our current connections to reach those nodes.
I also think it'd be worth brainstorming more creative solutions with a bunch of yoda timers. I'll try one right now.
Some of these ideas seem pretty solid. My sense is that the best path forward is:
This is a little rambly. Sorry. I'll end here.
These are all good ideas, but I also think it’s important not to Chesterton’s Fence too hard. A lot of passionate people avoid doing alignment stuff because they assume it’s already been considered and decided against, even though the field doesn’t have that many people and much of its cultural capital is new.
Be serious, and deliberate, and make sure you’re giving it the best shot if this is the only shot we have, but most importantly, actually do it. There are not many other people trying.
These are a lot of good ideas. I comment above I think a good approach is to truly represent that we are a bunch of younger people who fear for the future...this would appeal to a lot of folks at his level, to know the kids are scared and need his help.
Hi, long-time lurker but first-time poster with a background in math here. I personally agree that it would be a good idea if we were to at least try to get some extremely talented mathematicians to think about alignment. Even if they decide not to, it might still be interesting to see what kinds of objections they have to working on it (e.g. is it because they think it's futile and doomed to failure, because they think AGI is never going to happen, because they think alignment will not be an issue, because they feel they have nothing to contribute, or because it's not technically interesting enough?).
However, I would also like to second TekhneMakre's concerns about the format and content of the email. If you sample some comments on posts on Terry Tao's blog, you will find that there are a number of commenters who would probably best be described as cranks who indefatigably try to convince Terry that their theories are worth paying attention to, that Terry is currently not wisely spending his time, etc. He (sensibly) ignores these comments, and has probably learned for the sake of sanity not to engage with anyone who seems to fit this bill. I am concerned that the email outlined in your post will set off the same response and thus be ignored. AI safety is still a rather fringe idea amongst academics, at least partly because it is speculative and lacking concreteness. It took me years as an academic-adjacent person to be even somewhat convinced that it could be a problem (I still am not totally convinced, but I am convinced it is at least worth looking into). I do not think an email appealing to emotion and anecdotes is likely to convince someone from that background encountering this problem.
I have three alternative suggestions; I'm not sure how good they are, so take them each with a grain of salt:
Firstly, note that Scott Aaronson has said here https://scottaaronson.blog/?p=6288#comment-1928043 that he would provisionally be willing to think about alignment for a year. This seems like it would have several advantages (1) He has already signaled interest, provisionally, so it would be easier to convince him that it might be worth working on, (2) He is already acquainted with many of the arguments for taking AGI seriously, so could start working on the problem more immediately, (3) He is well acquainted with the rationalist community and so would not be put off by rationalist norms or affiliated ideas such as EA (which I believe accounts for the skepticism of at least some academics), (4) Scott's area of work is CS theory, which seems like it would be more relevant to alignment than Tao's fields of interest.
Secondly, there are some academics who take AI safety arguments seriously. Jacob Steinhardt comes to mind, but I'm sure there are a decent number of others, especially given recent progress on AI. If these academics were to contact other top academics asking them to consider working on AI safety, the request would come across as much more credible. They would also know how to frame the problem in such a way to pique the interest of top mathematicians/computer scientists.
Thirdly, note that there are many academics who are open to working on big policy problems that do not directly concern their primary research interests. Terry Tao, I believe, is one of them, as evidenced by https://newsroom.ucla.edu/dept/faculty/professor-terence-tao-named-to-president-bidens-presidents-council-of-advisors-on-science-and-technology . I'm not sure to what extent this is an easier problem or a desirable course of action, but if you could convince some people in politics that this problem is worth taking seriously, it is possible that the government might directly ask these scientists to think about it.
This last point is not a suggestion, but I would like to add one note. Eliezer claims that he was told that you cannot pay top mathematicians to work on problems. I believe this is somewhat false. There are many examples of very talented professors and PhD students leaving academia to work at hedge funds. One example is Abhinav Kumar, who a few years ago was 1 of 4 coauthors on a paper solving the long-open problem on optimal sphere packings in 24 dimensions. He left an Associate Professorship at MIT to work at Renaissance Technologies (a hedge fund). Not exactly in the same vein, but Huawei has recruited 4 Fields medalists to work with them (e.g. see https://www.ihes.fr/en/laurent-lafforgue-en/ for one example) although I'm not certain whether they are working on applied problems. I cannot say whether money is a motivating factor in any given one of these cases, but there are more examples like this, and I think it is fair to say that at least some substantial fraction of all such people involved might have been motivated at least partly by money.
Seems that I wasn't the only person to notice Scott's comment on his blog :) He's just announced that he'll be working on alignment at OpenAI for a year: https://scottaaronson.blog/?p=6484
Oh wow, didn't realise how recent the Huawei recruitment of Field medalists was! This from today. Maybe we need to convince Huawei to care about AGI Alignment :)
Then do you think I should contact Jacob Steinhardt to ask him what I should write to interest Tao and avoid seeming like a crank?
There isn’t much I can do about SA other than telling him to work on the problem in his free time.
Unless something extraordinary happens I’m definitely not contacting anyone in politics. Politicians being interested in AGI is a nightmarish scenario and those news about Huawei don’t help my paranoia about the issue.
My concern is less your email, and more the precedent. Having the rationality community model and encourage obviously undesired forms of contact with high-prestige figures seems like it could lead to intrusions of privacy. One person sending an email is ignorable. If emails, phone calls, unsolicited office visits, etc. start piling up under the banner of “AI risk,” it could feel quite invasive to those on the receiving end. My concern in particular is that people doing as you’re doing may not have the capacity to coordinate their actions. We may not even know whether or how much “randomly emailing Terry Tao about X risk” is going on.
That’s part of the point of the post, to coordinate so that fewer emails are sent. I asked if anyone tried something similar and asked people not to send their own emails without telling the rest of us.
I think Maxim Kontsevich might be a better candidate for an elite mathematician to try to recruit. Check out this 2014 panel with him, Tao and some other eminent mathematicians -- he alone said that he thought HLAI(in math) is plausible in our lifetimes, but also that working on it might be immoral(!) He also mentioned an AI forecast by Kolmogorov that I had never heard of before, so it seems he has some pre-existing interest in the area.
Um. If you want to convince a mathematician like Terry Tao to be interested in AI alignment, you will need to present yourself as a reasonably competent mathematician or related expert and actually formulate an AI problem in such a way so that someone like Terry Tao would be interested in it. If you yourself are not interested in the problem, then Terry Tao will not be interested in it either.
Terry Tao is interested in random matrix theory (he wrote the book on it), and random matrix theory is somewhat related to my approach to AI interpretability and alignment. If you are going to send these problems to a mathematician, please inform me about this before you do so.
My approach to alignment: Given matrices , define a superoperator by setting
, and define . Define the -spectral radius of as . Here, is the usual spectral radius.
Define . Here, is either the field of reals, field of complex numbers, or division ring of quaternions.
Given matrices , define
. The value is always a real number in the interval that is a measure of how jointly similar the tuples are. The motivation behind is that is always a real number in (well except when the denominator is zero) that measures how well can be approximated by -matrices. Informally, measures how random are where a lower value of indicates a lower degree of randomness.
A better theoretical understanding of would be great. If and is locally maximized, then we say that is an LSRDR of . Said differently, is an LSRDR of if the similarity is maximized.
Here, the notion of an LSRDR is a machine learning notion that seems to be much more interpretable and much less subject to noise than many other machine learning notions. But a solid mathematical theory behind LSRDRs would help us understand not just what LSRDRs do, but the mathematical theory would help us understand why they do it.
Problems in random matrix theory concerning LSRDRs:
P.S. By massively downvoting my posts where I talk about mathematics that is clearly applicable to AI interpretability and alignment, the people on this site are simply demonstrating that they need to do a lot of soul searching before they annoy people like Terry Tao with their lack of mathematical expertise.
P.P.S. Instead of trying to get a high profile mathematician like Terry Tao to be interested in problems, it may be better to search for a specific mathematician who is an expert in a specific area related to AI alignment since it may be easier to contact a lower profile mathematician, and a lower profile mathematician may have more specific things to say and contribute. You are lucky that Terry Tao is interested in random matrix theory, but this does not mean that Terry Tao is interested in anything in the intersection between alignment and random matrix theory. It may be better to search harder for mathematicians who are interested in your specific problems.
P.P.P.S. To get more mathematicians interested in alignment, it may be a good idea to develop AI systems that behave much more mathematically. Neural networks currently do not behave very mathematically since they look like the things that engineers would come up with instead of mathematicians.
P.P.P.P.S. I have developed the notion of an LSRDR for cryptocurrency research because I am using this to evaluate the cryptographic security of cryptographic functions.
I have heard about the thing where you commit to a $100m reward for any ML or mathmetician who solves alignment, and simultaneously pay 100 top ML and mathmeticians $1m over the course of a year to do nothing but pursue a solution to alignment (pursuing the bounty in the process). Even if all 100 of them fail, you still selected the best 100 out of every mathmetician who applied for those positions, so a large proportion of them might pursue the problem on their own afterwards in pursuit of the ongoing $100 million bounty. One way or another, many of these influential people will be convinced that the problem is significant and tell their friends, or even contract their friends as consultants to help with the problem.
There's plenty of trust issues, going both ways, but I'm not a grantmaker or lawyer and I think some smart, experienced people could probably figure out how to mitigate most of them.
I really want this to happen.
And why stop at Terry Tao? We could also email other top mathematicians and physicists.
I would put that in Google doc, it will be easier to suggest changes etc
I've seen a lot of discussion of Terence Tao, but not much of 'the overall set of young people who have done impressive technical work but are dramatically less busy and famous'. I also see a lot of discussion of people who are already socially adjacent to the community or have had multiple conversations about AI risk. I'd expect more value from poking people who are less visible, more available, and who haven't spent as much time thinking or talking about AI risk already.
Note that people of Tao's level and prominence likely receive extreme quantities of unsolicited emails, and those remain unread by the intended recipient, possibly not even the subject line. Put yourself in the place of Tao's assistant, why would they let something like what you intend to send through?
Yes, I think the email needs to come from someone with a lot of clout (e.g. a top academic, or a charismatic billionaire; or even a high-ranking government official) if we actually want him to read it and take it seriously.
Maybe reaching Demis Hassabis first is the way to go though, given that he's already thinking about it, and has already mentioned it to Tao (according to the podcast). Does anyone have links to Demis? Would be good to know more about his "Avengers assemble" plan! The main thing is that the assembly needs to happen asap, at least for an initial meeting and "priming of the pump" as it were.
Do you mean website links about his plan? I found nothing.
I’m still not changing the deadlines but I’ve received information that made me want to change the order.
No I mean links to him in person to talk to him (or for that matter, even an email address or any way of contacting him..).
Should also say - good that you are thinking about it P., and thanks for a couple of the links which I hadn't seen before.
Why CEA in particular, and why their community health team in particular? I think it's good to get feedback before sending this email (I wouldn't send the emails in their current form), but I don't think of those groups as particularly relevant for how to contact Terry Tao or Demis.
My understanding was that they were the team to talk to within EA if you're thinking about doing outreach to high-net worth individuals, politicians, projects related to children; or projects involving famous people like this one.
It just occurred to me that Terry Tao is probably one of the less good people to pursue here. He's more of a popular, public figure type. Which, I assume, means that he is contacted by more people, and thus less receptive to things like this. Whereas there are surely other people who are similarly smart but way less popular amongst the general public. And with a quick email to, say, a PhD student in math, it'd probably be pretty easy to find out who these people are.
An EA contacted me who knows Kontsevich and is considering reaching out to him. If you want to coordinate with that person, let me know and I can put you in touch.
Please do! You can DM me their contact info, tell them about my accounts: either this one or my EA Forum one, or ask me for my email address.
I'm not sure about the timing of when the edits in your post were made, but if you want feedback about your planned contact with Demis Hassabis I think you should make a new post about it -- most people who would comment on it may have missed it because they only saw the original unedited post about Tao, which had already received feedback.
I also think that, for the same reason that you chose to let someone else contact Tao instead of you, it may be better to let someone else contact Hassabis (or find someone else to contact him).
The email to Demis has been there since the beginning, I even received feedback on it. I think I will send it next week, but will also try to get to him through some DeepMind employee if that doesn’t work.
Please don't send these types of emails, I expect that they're actively counterproductive for high-profile recipients.
If you want to outreach, there are clear channels which should be used for coordinating it. For example, you could contact DeepMind's alignment team, and ask them if there's anything which would be useful for them.
Any updates? I skimmed through the comments and answers but it seems that we just know someone planned to contact Terrance Tao, and no results have been reported back.
Hey P. Assuming Demis Hassabis reads your email and takes it seriously, why won’t his reaction be “I already have my alignment team, Shane Legg took care of that” ?
Deepmind has had an alignment team for a long time.
Well, if he has, unbeknownst to me, already hired the “Terence Taos of the world” like he said on the podcast, that would be great, and I would move on to other tasks. But if he only has a regular alignment team, I don’t think either of us considers that to be enough. I’m just trying to convince him that it’s urgent and we can’t leave it for later.
This has been discussed several times in the past, see:
But I’m not aware of anyone that has actually even tried to do something like this.
Of special interest is this comment by Eliezer about Tao:
So if anyone has contacted him or people like him (instead of regular college professors), I’d like to know how that went.
Otherwise, especially for people that aren’t merely pessimistic but measure success probability in log-odds, sending that email is a low cost action that we should definitely try.
So you (whoever is reading this) have until June 23rd to convince me that I shouldn’t send this to his @math.ucla.edu address:Edit: I’ve been informed that someone with much better chances of success will be trying to contact him soon, so the priority now is to convince Demis Hassabis (see further below) and to find other similarly talented people.
Title: Have you considered working on AI alignment?
Body:
You can do any of:
What you probably shouldn’t do is to send your own email without telling the rest of us. His attention is a limited resource and bothering him with many different emails might reduce his sympathy for the cause.
And other than him, how many people do you think have a comparable chance of solving the problem or making significant progress? And how do we identify them? By the number of citations? Prizes won? I would like to have a list like that along with conditions under which each alignment org would hire each person. The probability of convincing Tao might be low, but with, say, 100 people like him the chances of finding someone might be decent.
I’m pretty sure that most of them haven’t heard about alignment, or have and just discarded it as something not worth thinking about. I don’t think this means that they couldn’t do great alignment work if they tried, maybe getting them to seriously think about the problem at all is the hard part, and after that their genius simply generalises to this new area.
Relatedly, does anyone here know why Demis Hassabis isn’t assembling his dream team right now? The same as above applies, but until
the 1st of JulyJune 23rd:Title: Are you sure you should wait before creating your dream alignment team?
Body:
This should be obvious to him, but just in case.
This linked here before, but I was told it was a bad idea.