All of Ras1513's Comments + Replies

Thank you for the reply. This has been an important takeaway from this post: There are significant groups (or at least informal networks) doing meaningful work that don't congregate primarily on LW or Twitter. As I said on another comment - that is encouraging! I wish this was more explicit knowledge within LW - it might give things more of a sense of hope around here.

The first question that comes to mind: Is there any sense of direction on policy proposals that might actually have a chance of getting somewhere? Something like: "Regulating card production"... (read more)

2lincolnquirk
Your question is coming from within a frame (I'll call it the "EY+LW frame") that I believe most of the DC people do not heavily share, so it is kind of hard to answer directly. But yes, to attempt an answer, I've seen quite a lot of interest (and direct policy successes) in reducing AI chips' availability and production in China (eg via both CHIPS act and export controls), which is a prerequisite for US to exert more regulatory oversight of AI production and usage. I think the DC folks seem fairly well positioned to give useful inputs into further AI regulation as well.

This is fantastic information, thank you for taking the time.

One of my big takeaways from all of the comments on this post is a big update to my understanding of the "AI Risk" community and that LW was not actually the epicenter and there were significant efforts being made elsewhere that didn't necessarily circle back to LW.

That is very encouraging actually!

The other big update is what you say: There were just so few people with the time and ability to work on these things.

Someone else said similar about the basement possibility, which I did not know.

Interesting questions raised though: Even if it wasn't clear until GPT, wouldn't that still have left something like 2-3 years?

Granted that is not 10-20 years.

It seems we all, collectively, did not update nearly enough on ChatGPT-2?

2ShardPhoenix
ChatGPT was released on November 30 2022, so it's only been around 7 months. The older ones were GPT-2 and GPT-3 which got attention among AI-followers but were relatively unknown to the public - and again, it wasn't obvious then when or if ordinary people would come to know or care about these advances.
3mishka
I personally was mostly ignoring Transformers until GPT-3 came along. (That was mostly because I believed that "true AI" had to be a recurrent machine, whereas Transformers were feed-forward, and I did not understand that Transformers emulated RNNs when used in the autoregressive mode (the paper explaining that was published shortly after GPT-3).) So I thought, Transformers were remarkable, but the "true route to AGI" was elsewhere. Then GPT-3 achieved two "Holy Grails" widely believed to be "impossible"/"long in the future". Namely, few shot learning (so like a human it could learn a pattern from a single exposure without long training) and semi-competent program synthesis (which was considered to be next to impossible because of fragility of computer code to noise such as one-letter changes, with this supposed near-impossibility of program synthesis being the key obstacle to recursive self-improvement and foom). These two breakthroughs were the key reason why I updated, rather than general linguistic competence (which was indeed quite impressive in 2019 models already).

Point taken! This I just plain did not know and I will update based on that.

 It does not make sense to focus on public policy if basement guy is the primary actor.

The UK funding is far and away the biggest win to date, no doubt.

And all this is despite the immediately-preceding public relations catastrophe of FTX!


Do you feel that FTX/EA is that closely tied in the public mind and was a major setback for AI alignment? That is not my model at all.

We all know they are inextricably tied, but I suspect if you were to ask they very people in those same polls if they knew that SBF supported AI risk research they wouldn't know or care.

3Julian Bradshaw
I don't think they're closely tied in the public mind, but I do think the connection is known to the organs of media and government that interact with AI alignment. It comes up often enough, in the background - details like FTX having a large stake in Anthropic, for example. And the opponents of AI x-risk and EA certainly try to bring it up as often as possible. Basically, my model is that FTX seriously undermined the insider credibility of AINotKillEveryoneIsm's most institutionally powerful proponents, but the remaining credibility was enough to work with.

I've added an Edit to the post to include that right up front.

2Neel Nanda
Thanks!

I asked this of another commenter, but I will ask you too:

Do you feel it is accurate to say that many or most people working on this (including and especially Eliezer) at the time considered nuts and bolts alignment work to be the only worthwhile path? Given what info was available at the time.

And that widescale public persuasion / overton window / policy making was not likely to matter as the most scenarios were Foom based?

2Chris_Leong
At the start the majority of people who were worried about AGI were worried about foom, but it’s been less clear that it’s a majority in the last few years. It might have played a role, but I wouldn’t say that it has been the central factor.

You reminded me of that famous tweet
 

Sci-Fi Author: In my book I invented the Torment Nexus as a cautionary tale 

Tech Company: At long last, we have created the Torment Nexus from classic sci-fi novel Don't Create The Torment Nexus

But more seriously, I think this is a real point that has not been explored enough in alignment circles.

I have encountered a large number of people - in fact probably almost all people I discuss AI with - who I would call "normal people". Just regular, moderately intelligent people going about their lives for whi... (read more)

7Mitchell_Porter
In reply, let me start by listing some of the attitudes to AI risk that are out there.  First, there are the people who don't think AI will be a serious rival to human intelligence any time soon, or ever. This includes a lot of people who actually are concerned about AI, but for different reasons: disinformation, scams, dehumanized daily life, unemployment, centralization of power, etc.  Then, there are the people who think AI can or will surpass us, but who embrace this. A lot of these people want AI ASAP so they can cure disease, create abundance and leisure, or even live in a Star Trek futurist utopia. Some of them talk as if AI will always be a human servant no matter how powerful it becomes (and they may or may not worry about humans using this superhuman servant for unwise or evil purposes).  In this accelerationist faction, among those who have a position on AI that is not just smarter than us, but independent of us, I see two attitudes. There are those who think that a superintelligent being will inherently discover the correct morality and adopt it; and there are those who have a kind of "might makes right" attitude - if it's smarter than us, it deserves to be in charge, and has the right to do whatever it wants.  Then we have what we could call the alignment faction, who see coexistence of AI and humanity as something that must be engineered, it won't happen by itself. The original philosophy of MIRI was an alignment philosophy with an extreme focus on safety: do not do anything that risks creating superintelligence, until you have the complete theory of alignment figured out.  Now, in the era of deep learning and language models, and billion-dollar companies explicitly aiming to create superhuman AI, there is an alignment philosophy with a greater resemblance to the norms of commercial software development. You work on your AI in private, you make it as good as you can, or good enough for release, then you put it out there, and then there's an ongoin

I'm starting to draw a couple of conclusions for myself from this thread as I get a better understanding of the history.

Do you feel it is accurate to say that many or most people working on this (including and especially Eliezer) at the time considered nuts and bolts alignment work to be the only worthwhile path? Given what info was available at the time.

And that widescale public persuasion / overton window / policy making was not likely to matter as the most scenarios were Foom based?

It is pretty interesting that the previous discussion in all these years... (read more)

1mishka
It is difficult to talk about community as a whole. Right now there is a lot of diversity of opinion about likely future dynamics (timelines (from ultra-short to ultra-long), foom vs no-foom, single dominating AI vs multi-polar forces, etc), about likely solutions for AI existential safety if any, about likely difficulty of those solutions, etc. The whole situation is such a mess precisely because the future is so multi-variate; it's difficult to predict how it will go, and it's difficult to predict properties of that unknown future trajectory. See, for example, this remarkable post: 60+ Possible Futures See also this post by Zvi about how ill-defined the notion of alignment is: Types and Degrees of Alignment ---------------------------------------- With Eliezer, I only have snapshot impressions of his evolving views. I have been exposed to a good part of his thought, but not to all of it. At some point, he strongly wanted provably friendly AI. I had doubts that that was possible, and I remember our conversation at his poster at AGI-2011. I said (expressing my doubts), "but would not AI rebel against any constraints one tries to impose on it; just look at our teenagers; I would certainly rebel if I knew I was forced to behave in a specific way", and Eliezer told me, "that's why we should not build a human-like AI, but should invent an entirely different architecture, such that one can prove things about it". (And he has a very good point here, but compare this with his recent suggestions to focus on radical intelligence amplification in humans as a last ditch effort; that is exactly the prescription for creating human-like (or human-digital hybrid) super-intelligent entities, which he told me in 2011 we should not do; those entities will then decide what they want to happen, and who knows what would they decide, and who knows if we are going to have better chances with them than with artificial systems.) Then MIRI started to focus on "Loebian obstacle" (which

You say that we have no plan to solve AI risk, so we cannot communicate that plan. That is not the same as not being able to take any beneficial actions.

"We do not know the perfect thing to do, therefore we cannot do anything"

Do we not consider timeline-extending things to be worthwhile?

This is a genuine question: Is it the prevailing wisdom that ultimately solving AI X-risk is the only worthwhile path and only work worthy of pursing? This seems to have been Eliezer's opinion prior to GPT-3ish. That would answer the questions of my original post.

For exampl... (read more)

4ChristianKl
One of the plans that were actually done was OpenAI which actually accelerated timelines. Taking action and that action having a positive effect is not the same thing.  There are things you can do to slow down AI progress but if humanity dies one or two year later then otherwise that's not still not a very desirable outcome. Without having a plan that has a desirable outcome it's hard to convince people of it.  Politics tends to be pretty hostile for clear thinking and MIRI and CFAR both thought that creating an environment where clear thinking can happen well is crucial for solving AI risk. 

Are these "DC" people you are talking about organized somewhere? Or is this a more hidden / informal type of thing?

I ask because I have seen both Zvi and Eliezer make comments to the effect of: "There is no on special behind the curtain working on this - what you see on twitter is what there is" (my paraphrasing)

2lincolnquirk
I've been in DC for ~ the last 1.5y and I would say that DC AI policy has a good amount of momentum, I doubt it's particularly visible on twitter but also it doesn't seem like there are any hidden/secret missions or powerful coordination groups (if there are, I don't know about it yet). I know ~10-20 people decently well here who work on AI policy full time or their work is motivated primarily by wanting better AI policy, and maybe ~100 who I have met once or twice but don't see regularly or often; most such folks have been working on this stuff since before 2022; they all have fairly normal-seeming thinktank- or government-type jobs. They don't mostly spend time on LW (although certainly a few of them do). Many do spend time on Twitter, and they do read lots of AI related takes from LW-influenced folks. They have meetup groups related to AI policy. I guess it looks pretty much as I was expecting before I came here. Happy to answer further questions that don't identify specific people, just because I don't know how many of them want to be pointed-at on LW.

Great websites! 

I find it interesting that you are the second commenter (and Dan H above) to jump in and explicitly say: I have been doing that! 

and point to great previous work doing exactly these things, but from my perspective they do not seem widely known or supported within the community here (I could be wrong about that)

I am starting to feel that I have a bad map of the AI Alignment/Safety community. My previous impression was the lesswrong / MIRI was mostly the epicenter, and if much of anything was being done it was coming from there or at least was well known there. That seems not to be the case - Which is encouraging! (I think)

5plex
This is true of many people, and why I built the map of AI safety :) Next step is to rebuild aisafety.com into a homepage which ties all of this together, and offer AI Safety Info's database via an API for other websites (like aisafety.com, and hopefully lesswrong) to embed.

100% agreed - I thought I had flagged the complete hindsight bias by saying that it is obvious in retrospect.

The post was a genuine attempt to ask why it was not a clear path before.

3Neel Nanda
Ah, thanks I hadn't noticed that one. I'm a fan of putting this kind of thing early on, not just at the end - it's pretty easy to come across as attacking people, or "man aren't people dumb for not noticing this" without clearly noticing biases like this up front.

I completely agree that it made no sense to divert qualified researchers away from actually doing the work. I hope my post did not come across as suggesting that.