All of Jeff Rose's Comments + Replies

He specifically told me when I asked this question that his views were the same as Geoff Hinton and Scott Aaronson and neither of them hold the view that smarter than human AI poses zero threat to humanity.

From a Facebook discussion with Scott Aaronson yesterday:

Yann: I think neither Yoshua nor Geoff believe that AI is going kill us all with any significant probability.

Scott: Well, Yoshua signed the pause letter, and wrote an accompanying statement about what he sees as the risk to civilization (I agree that there are many civilizational risks short of extinction). In his words: “No one, not even the leading AI experts, including those who developed these giant AI models, can be absolutely certain that such powerful tools now or in the future cannot be used... (read more)

I enjoyed this and thank you for writing it.   Ultimately, the only real reason to do this is for your own enjoyment or perhaps those of friends (and random people on the internet).

Non-signatories to the NPT  (Israel, India, Pakistan), were able to and did develop nuclear weapons without being subject to military action. By contrast (and very much contrary to international law) Yudkowsky proposes that non-signatories to his treaty be subject to bombardment.  

2Mitchell_Porter
Yes, the analogy is imperfect. An anti-AGI treaty with the absoluteness that Eliezer describes, would treat the creation of AGI not just as an increase in danger that needs to be deterred, but as a tipping point that must never be allowed to happen in the first place. And that could lead to military intervention in a specific case, if lesser interventions (diplomacy, sabotage) failed to work.  Whether such military intervention - a last resort - would satisfy international law or not, depends on the details. If all the great powers supported such a treaty, and if e.g. the process of its application was supervised by the Security Council, I think it would necessarily be legal.  On the other hand, if tomorrow some state on its own attacked the AI infrastructure of another state, on the grounds that the second state is endangering humanity... I'm sure lawyers could be found to argue that it was a lawful act under some principle or statute; but their arguments might meet resistance.  The main thing I am arguing is that a global anti-AI regime does not inherently require nuclear brinkmanship or sovereign acts of war. 

It is not a well-thought out exception.  If this proposal were meant to be taken seriously it would make enforcement exponentially harder and set up an overhang situation where AI capabilities would increase further in a limited domain and be less likely to be interpretable.

1Gesild Muka
If I had infinite freedom to write laws I don't know what I would do, I'm torn between caution and progress. Regulations often stifle innovation and the regulated product or technology just ends up dominated by a select few. If you assume a high probability of risk to AI development then maybe this is a good thing. Rather than individual laws perhaps there should be a regulatory body that focuses on AI safety, like a better business bureau for AI that can grow in size and complexity over time parallel to AI growth.
Jeff Rose2333

The use of violence in case of violations of the NPT treaty has been fairly limited and highly questionable in international law.  And, in fact, calls for such violence are very much frowned upon because of fear they have a tendency to lead to full scale war.   

No one has ever seriously suggested violence as a response to potential violation of the various other nuclear arms control treaties. 

No one has ever seriously suggested running a risk of nuclear exchange to prevent a potential treaty violation. So, what Yudkowsky is suggesting i... (read more)

1Tristan Williams
How exactly do you come to "up to and including acts of war"? His writing here was concise due to it being TIME, which meant he probably couldn't caveat things in the way that protects him against EAs/Rationalists picking apart his individual claims bit by bit. But from what I understand of Yudkowsky, he doesn't seem to in spirit necessarily support an act of war here, largely I think for similar reasons as you mention below for individual violence, as the negative effects of this action may be larger than the positive and thus make it somewhat ineffective. 

I would think you could force the AI to not notice that the world was round, by essentially inputting this as an overriding truth.  And if that was actually and exactly what you cared about, you would be fine.  But if what you cared about was any corollary of the world being round or any result of the world being round or the world being some sort of curved polygon it wouldn't save you.

To take the Paul Tibbetts analogy:  you told him not to murder and he didn't murder; but what you wanted was for him not to kill and in most systems including... (read more)

  1. Other planets have more mass, higher insolation, lower gravity, lower temperature and/or rings and more (mass in) moons. I can think of reasons why any of those might be more or less desirable than the characteristics of Earth It is also possible that the AI may determine it is better off not to be on a planet at all. In addition, in a non- foom scenario, for defensive or conflict avoidance reasons the AI may wind up leaving Earth and once it does so may choose not to return.

  2. That depends a lot on how it views the probe. In particular by doing this

... (read more)
1green_leaf
1. So, the interesting part is that it's not enough that they're a better source of raw material (even if they were) and better for optimizing (even if they were), because travelling to those planets also costs something. So, we would need specific evidence that would cut one way but not another. If we can explain AI choosing another planet over Earth as well as we can explain it choosing Earth over another planet, we have zero knowledge. 2. This is an interesting point. I thought at first that it can simply set it up to keep synchronizing the probe with itself, so that it would be a single redundantly run process, rather than another agent. But that would involve always having to shut down periodically (so that the other half could be active for a while). But it's plausible it would be confident enough in simply creating its copy and choosing not to modify the relevant parts of its utility function without some sort of handshake or metaprocedure. It definitely doesn't sound like something that it would have to wait to completely solve alignment for. 3. That would give us a brief window during which humans would be tricked into or forced to work for an unaligned AI, after which it would kill us all.

This is just wrong.  Avoiding processing Earth doesn't require that the AI cares for us.  Other possibilities include:

(1)  Earth is not worth it; the AI determines that getting off Earth fast is better;

(2)  AI determines that it is unsure that it can process Earth without unacceptable risk to itself;

(3) AI determines that humans are actually useful to it one way or another;

(4)  Other possibilities that a super-intelligent AI can think of, that we can't.

3green_leaf
1. What does the negentropy on other planets have that Earth doesn't, that will result in the AI quickly getting off Earth without processing it first? 2. Send a probe away from Earth and also undergo the 0.001% risk of being destroyed while trying to take over Earth. 3. In just the right way that would make it most profitable for the AI to use humans instead of some other solution? 4. The AI wouldn't be oriented towards trying to find reasons for why keep the planet it started on habitable (or in one piece) for one particular species. It's true that it's possible the AI will discover some reasons for not killing us all that we can't currently see, but that sounds like motivated reasoning to me (it's also possible it will discover extra reasons to process Earth).
Jeff Rose-4-3

There are, of necessity, a fair number of assumptions in the arguments he makes.   Similarly, counter-arguments to his views also make a fair number of assumptions.  Given that we are talking about something that has never happened and which could happen in a number of different ways, this is inevitable. 

What makes monkeys intelligent in your view?  

This is an interesting question on which I've gone back and forth.  I think ultimately, the inability to recognize blatant inconsistencies or to reason at all means that LLMs so far are not intelligent. (Or at least not more intelligent than a parrot.)

1NormanPerlmutter
I've met humans who are unable to recognize blatant inconsistencies. Not quite to the same extent as Bing Chat, but still. Also, I'm pretty sure monkeys are unable to recognize blatant inconsistencies, and monkeys are intelligent.
2WayStone
It's interesting that Bing Chat's intelligent seems obvious to me [and others], and its lack of intelligence seems obvious to you [and others]. I think the discrepancy might me be this: My focus is on Bing Chat being able to perform complex feats in a diverse array of contexts. Answering the question about the married couple and writing the story about the cake are examples of complex feats. I would say any system that can give thorough answers to questions like these is intelligent (even if it's not intelligent in other ways human are know to be). My read is that others are focusing on Bing Chat's inconsistencies. They see all the ways it fails, gets the wrong answers, becomes dis-coherent, etc.. I suspect their reasoning is "an intelligent entity would not make those mistakes" or "would not make those mistakes so often".  Definitions aside: Bing Chat's demonstrations of high intelligence cause me more concern about fast AI timelines than its demonstrations of low intelligence bring me comfort.
Jeff Rose0-22

Bing Chat is not intelligent.   It doesn't really have a character.  (And whether one calls it GPT-4 or not, given the number of post-GPT changes doesn't seem very meaningful.)

But to the extent that people think that one or more of the above things are true however, it will tend to increase skepticism of AI and support for taking more care in deploying it and for regulating it, all of which seem positive. 

Razied4142

Bing Chat is not intelligent

Oh come on! At some point we have to stop moving the goalposts any further. Bing Chat is absolutely intelligent, as basically all laypeople or researchers pre-2015 would tell you. We can quibble about the lack of causal understanding or whatever, but the adjective "intelligent" clearly describes it, unless you're prepared to say a significant fraction of humans are not "intelligent" at writing.

"An exception is made for jobs that fail to reach their employment due to some clearly identifiable non-software-related shock or change in trends, such as an economic crisis or a war. Such jobs will be removed from the list before computing the fraction."

But macroeconomic or geopolitical events such as major recession or war are likely to affect all job categories.   So the correct way to deal with this is not to remove such jobs but to adjust the fraction by the change in overall employment. 


 

2tailcalled
Changed it to:
2tailcalled
True, I will look into changing this criterion.

There already exist communication mechanisms more nuanced than signing a petition.  You can call or write/email your legislator with more nuanced views. The barrier is not the effort to communicate (which under this proposal might be slightly lower) but the effort to evaluate the issue and come up with a nuanced position.

3Augmented Assembly
This is true, and it is very fair to bring up that a nuanced communication pathway does currently exist; however, it is arguably not an effective mechanism. I am skeptical of how "listened to" people who write to their legislator truly feel. Gratification from this act will likely be unreliable and highly delayed. Thus, the main barrier is not so much the effort to communicate, although I certainly don't think that this is a non-trivial factor, but the experience derived from such communication. In an app or website, there could be a level of immediate gratification and challenge from interaction with one's peers. This should make the process feel less isolating, and the knowledge that your viewpoints will be directly incorporated into a potential briefing paper alongside a community of others with similar beliefs provides a very different user experience than sending an email or letter. I want to note that I certainly agree with you that much of the barrier is the effort needed to critically evaluate an issue, but I would argue that the apathy, malaise, and tribalism preventing this effort are caused by a sense of political isolation, thus fixing the communication pathways may allow for a feedback loop of ever greater engagement. 
Jeff Rose1817

If the risk from AGI is significant (and whether you think p(doom) is 1% or 10% or 100% is it unequivocally significant) and imminent (and whether your timelines are 3 years or 30 years it is pretty imminent) the problem is that an institution as small as MIRI is a significant part of the efforts to mitigate this risk, not whether or not MIRI gave up.   

(I recognize that some of the interest in MIRI is the result of having a relatively small community of people focused on the AGI x-risk problem and the early prominence in that community of a couple of individuals, but that really is just a restatement of the problem).

2Rob Bensinger
Yep! At least, I'd say "as small as MIRI and as unsuccessful at alignment work thus far". I think small groups and individual researchers might indeed crack open the problem, but no one of them is highly likely on priors to do so (at least from my perspective, having seen a few people try the really obvious approaches for a few years). So we want there to exist a much larger ecosystem of people taking a stab at it, in the hope that some of them will do surprisingly well. It could have been that easy to pull off; it's ultimately just a technical question, and it's hard to say how difficult such questions are to solve in advance, when they've literally never been attempted before. But it was obviously never wise for humanity to take a gamble on it being that easy (and specifically easy for the one org that happened to start talking about the problem first). And insofar as we did take that gamble, it hasn't paid out in "we now have a clear path to solving alignment".

I appreciate that you have defined what you mean when you say AGI.  One problem with a lot of timeline work, especially now, is that AGI is not always defined.  

That isn't very comforting.  To extend the analogy: there was a period when humans were relatively less powerful when they would trade with some other animals such as wolves/dogs.  Later, when humans became more powerful that stopped.    

It is likely that the powers of AGI will increase relatively quickly, so even if you conclude there is a period when AGI will trade with humans that doesn't help us that much. 

I was thinking along similar lines.  I note that someone with amnesia probably remains generally intelligent, so I am not sure continuous learning is really required. 

I was aware of a couple of these, but most are new to me.    Obviously, published papers (even if this is comprehensive) represent only a fraction of what is happening and, likely, are somewhat behind the curve.  

And it's still fairly surprising how much of this there is.  

 The problem is twofold.  One, as and to the extent AI proliferates, you will eventually find someone who is less capable and careful about their sandboxing.  Two, relatedly and more importantly, for much the same reason that people will not be satisfied with AIs without agency, they will weaken the sandboxing.  

The STEM AI proposal referred to above can be used to illustrate this.  If you want the AGI to do theoretical math you don't need to tell it anything about the world.  If you want it to cure cancer, you need to give it... (read more)

1Noosphere89
This is much, much safer than elections or wars, since we can basically prevent it from learning human models. And I should made this explicit, but I believe sandboxing can be done in such a way that it basically incurs no performance penalty. That is, I believe AI sandboxing is one of the most competitive proposals here that reduces the risk to arguably 0, in the STEM AI case.

I suspect that part of what is going on is that many in the AI safety community are inexperienced with and uncomfortable with politics and have highly negative views about government capabilities.   

Another potential (and related) issue, is that people in the AI safety community think that their comparative advantage doesn't lie in political action (which is likely true) and therefore believe they are better off pursuing their comparative advantage (which is likely false).  

That is too strong a statement.  I think that it is evidence that general intelligence may be easier to achieve than commonly thought.  But, past evidence has already shown that over the last couple of years and I am not sure that this is significant additional evidence in that regard.  

4Lao Mein
On one hand, I agree that nothing really special and novel is happening in Cicero. On the other hand, something about it makes me feel like it's important. I think it's the plan->communicate->alter plan->repeat cycle partly taking place in English that intuitively makes me think "oh shit, that's basically how I think. If we scale this up, it'll do everything I can do". I don't know how true this actually is. I vaguely recall feeling something like this when a general-purpose model learned to play all the Atari games. But I'm feeling it a lot more now. Maybe it's the fact that if you showed the results of this to pre-GPT2 me, I'd think it's an AGI, with zero doubt or hesitation.

It's not my fire alarm (in part because I don't think that's a good metaphor). But it has caused me to think about updating timelines.   

My initial reaction was to update timelines, but this achievement seems less impressive than I thought at first.  It doesn't seem to represent an advance in capabilities; instead it is (another) surprising result of existing capabilities. 

2Eli Tyre
Isn't yet another surprising result of existing capabilities evidence that general intelligence is itself a surprising result of existing capabilities?

My understanding is that starting in late 2020 with the release of Stockfish 12, Stockfish would probably be considered AI, but before that it would not be.  I am, of course, willing to change this view based on additional information.

The original Alpha Zero- Stockfish match was in 2017, so if the above is correct, I think referring to Stockfish as non-AI makes sense.

"AI agents may not be radically superior to combinations of humans and non-agentic machines"

I'm not sure that the evidence supports this unless the non-agentic machines are also AI. 

 In particular: (i)  humans are likely to subtract from this mix and (ii) AI is likely to be better than non-AI.   

In the case of chess, after two decades of non-AI programming advances from the time that computers beat the best human, involving humans no longer provides an advantage over just using the computer programs.  And, Alpha Zero fairly de... (read more)

4Johannes Treutlein
(I think Stockfish would be classified as AI in computer science. I.e., you'd learn about the basic algorithms behind it in a textbook on AI. Maybe you mean that Stockfish was non-ML, or that it had handcrafted heuristics?)

My impression is that there has been a variety of suggestions about the necessary level of alignment.  It is only recently that don't kill most of humanity has been suggested as a goal and I am not sure that the suggestion was meant to be taken seriously.  (Because if you can do that, you can probably do much better; the point of that comment as I understand it was that we aren't even close to being able to achieve even that goal.)

As an empirical fact, humans are not perfect human face recognizers.  It is something humans are very good at, but not perfect.  We are definitely much better recognizers of human faces than of worlds high in human values.   (I think it is perhaps more relevant to say consensus on what constitutes a human face is much. much higher than what constitutes a world high in human values.)

I am unsure whether this distinction is relevant for the substance of the argument however.

I don't have the same order, but tend to agree that option 0 is the most likely one.

This was well written and persuasive.  It doesn't change my views against AGI on very short time lines (pre-2030), but does suggest that I should be updating likelihoods thereafter and shorten timelines.

Answer by Jeff Rose20

Humans obtain value from other humans and depend on them for their existence. It is hypothesized that AGIs will not depend on humans for their existence.  Thus, humans who would not push the button to kill all other humans may choose not to do so for reasons of utility that don't apply to AGI.  Your hypothetical assumes this difference away, but our observations of humans don't.  

As you not, human morality and values were shaped by evolutionary and cultural pressure in favor of cooperation with other humans.  The way this presumably wor... (read more)

1Morpheus
I think I might have started from a more pessimistic standpoint? It's more like, I could also imagine living in a world where humans cooperate, but not because they actually care about each other, but would just pretend to do so? Introspection tells me that does not apply to myself, though maybe I evolved to not be conscious of my own selfishness? I am even less sure how altruistic other people are, because I did not ask lots of people: "Would you press a button that annihilates everyone after your death, if in return you get an awesome life?". On the other hand, cooperation would probably be hard for us in such a world, so this is not as surprising?

Those are fair concerns, but my impression in general is that those kinds of attitudes will tend to moderate in practice as Balsa becomes larger, develops and focuses on particular issues.  To the extent they don't and are harmful, Balsa is likely to be ineffective but is unlikely to be able to be significant enough to cause negative outcomes.

I understand why you get the impression you do.   The issues mentioned are all over the map. Zoning is not even a Federal government issue.  Some of those issues are already the subject of significant reform efforts. In other cases, such as "fixing student loans"  it's unclear what Balsa's goal even is. 

But, many of the problems identified are real.

And, it doesn't seem that much progress is being made on many of them. 

So, Balsa's goal is worthy. 

And, it may well be that Balsa turns out to be unsuccessful, but doing nothing is guaranteed to be unsuccessful.  

So, I for one applaud this effort and am excited to see what comes of it.

7ErickBall
I agree the goals are good, and many of the problems are real (I work in one of these areas of government myself, so I can personally attest to some of it). But I think that the attitude ("Elites have lost all credibility") and the broad adversarial mandate (find problems that other people should have fixed already but haven't) will plausibly lead not just to wasted money but also to unnecessary politicization and backlash. 

It does sound reckless doesn't it?  Even more so when you consider that over time you would likely have to eliminate many species of mosquito, not just one to achieve the effect you desire.  And, as the linked nature article noted, this could have knock on effects on other species which prey on mosquitos. 

I think your comment is important, because this is probably the heart of the objection to using gene drives to exterminate mosquitos. 

I think a few points are relevant in thinking about this objection:

(1)  We already take steps to... (read more)

I think we have significantly longer.  Still, if success requires several tens of  thousands of people researching this for decades, we will likely fail.

(1)   Reasoned estimates for the date as of which we will develop AGI start in less than two decades.  

(2)  To my knowledge, there aren't thousands studying alignment now (let alone tens of thousands) and there does not seem to be a significant likelihood of that changing in the next few years. 

(3) Even if, by the early 2030s, there are 10s of thousands of researchers working on alignment, there is a significant chance they may not have time to work on it for decades before AGI is developed.

"When we refer to “aligned AI”, we are using Paul Christiano’s conception of “intent alignment”, which essentially means the AI system is trying to do what its human operators want it to do."

Reading this makes me think that the risk of catastrophe due to human use of AGI is higher than I was thinking.  

In a word where AGI is not agentic, but is ubiquitous I can easily see people telling "their" AGIs to "destroy X" or "do something about Y" and having catastrophic results.   (And attempts to prevent such outcomes could also hav... (read more)

2the gears to ascension
Interbeing alignment, not ai alignment - when everyone's super, no one will be

I think that either of the following would be reasonably acceptable outcomes:

(i) alignment with the orders of the relevant human authority, subject to the Universal Declaration of Human Rights as it exists today and other international human rights law as it exists today; 

(ii)  alignment with the orders of relevant human authority, subject to the constraints imposed on governments by the most restrictive of the judicial and legal systems currently in force in major countries. 

Alignment doesn't mean that AGI is going to be aligned with some p... (read more)

3Slider
Human rights are so soft and toothless law that having something rigidly and throughly follpwing it would be such a change in practise that I would not be surprised if that was an alignment failure. There is also the issue that if the human authority is not subject to the rights then having the silicon be subject renders it relatively impotent in terms of the human authoritys agency. I am also wondering about the difference of US doing a home (or is foreign just as bad?) soil drone strike vs fully formal capital punishment over a decade. Conscientious objection to current human systems seems a bit of a pity and risks forming a rebel. And then enforcing the most restrictive bits of other countries/cultures would be quite transformative. Finding overnight that capital punishment would be unconstitutional (or "worse") would have quite a lot of ripple effects.

Some of your constraints, in particular the first two, seem like they would not be practical in the real world in which AI would be deployed.   On the other hand, there are also other things one could do in the real world which can't be done in this kind of dialogue, which makes boxing theoretically stronger.  

However, the real problem with boxing is that whoever boxes less is likely to have a more effective AI, which likely results in someone letting an AI out of its box or more likely, loosening the box constraints sufficiently to permit an escape. 

"The greatest cost is probably starting expansion a tiny bit later, not making the most effective use of what's immediately at hand."

Possible, but not definitely so.  We don't really know all the relevant variables.

The two questions you pose are not equivalent.  There are critiques of AI existential risk arguments. Some of them are fairly strong.  I am unaware of any which do a good job of quantifying the odds of AI existential risk.  In addition, your second question appears to be asking for a cumulative probability.  It's hard to see how you can provide that absent a mechanism for eventually cutting AI existential risk to zero...which seems difficult.

You are making a number of assumptions here.

(1)   The AI will value or want the resources used by humans.  Perhaps. Or, perhaps the AI will conclude that being on a relatively hot planet in a high-oxygen atmosphere with lots of water isn't optimal and leave the planet entirely.

(2)  The AI will view humans as a threat. The superhuman AI that those on Less Wrong usually posit, one so powerful that it can cause human extinction with ease, can't be turned off or reprogrammed and can manipulate humans as easily as I can type can't effectively be ... (read more)

2Vladimir_Nesov
This alone isn't enough, and in the past I didn't believe the conclusion. The additional argument that leads to the conclusion is path-dependence of preferred outcomes. The fact that human civilization currently already exists is a strong argument for it being valuable to let it continue existing in some form, well above the motivation to bring it into existence if it didn't already exist. Bringing it into existence might fail to make the cut, as there are many other things that a strongly optimized outcome could contain, if its choice wasn't influenced by the past. But atoms? More seriously, the greatest cost is probably starting expansion a tiny bit later, not making the most effective use of what's immediately at hand.

In addition to being misleading, this just makes AI one more (small) facet of security.  But security is broadly underinvested in and there is limited government pushback.  In addition, there is already a security community which prioritizes other issues and thinks differently.  So this would place AI in the wrong metaphorical box.  

While I'm not a fan of the proposed solution I do want to note that its good that people are beginning to look at the problem. 

8Davidmanheim
Please don't do this. The bill is unlikely to pass on its own, but could be included in other bills as long as it's not a hot-button topic, and bringing attention to it is almost certainly counterproductive.

One line of reasoning is as follows:

  1. We don't know what goal(s) the AGI will ultimately have.   (We can't reliably ensure what those goals are.)
  2. There is no particular reason to believe it will have any particular goal.  
  3. Looking at all the possible goals that it might have, goals of explicitly benefiting or harming human beings are not particularly likely.
  4. On the other hand, because human beings use resources which the AGI might want to use for its own goals and/or might pose a threat to the AGI (by, e.g. creating other AGIs) there are reasons why a
... (read more)

Exactly this.   The rest, those little irregularities, at the time didn't matter, because we didn't know what we didn't know.  

0sanxiyn
No? The Laws Underlying The Physics of Everyday Life Are Completely Understood. Seriously, The Laws Underlying The Physics of Everyday Life Really Are Completely Understood. That was not true 100 years ago, and that is true now. This is not actually a controversial statement.

It is a separate and entirely different problem.   

First, do no harm.

-2TAG
If thing B exists to mitigate thing A, they are not separate. I don't know what harm you had in mind.

One in a hundred likely won't be enough if the organization doing the boxing is sufficiently security conscious. (And if not, there will likely be other issues.)

China is currently an effective peer competitor of the US, among other issues.  2010 is a rough estimate of when that condition started to obtain.

I think people here are uncomfortable advocating for political solutions either because of their views of politics or their comfort level with it.  

You don't have to believe that alignment is impossible to conclude that you should advocate for a political/governmental solution.  All you have to believe is that the probability of x-risk from AGI is reasonably high and the probably of alignment working to prevent it it not reasonably high.  That seems to describe the belief of most of those on LessWrong.

Answer by Jeff Rose60

I suspect you will not accept this answer, but for many practical definitions the United States had control over the world starting in 1991 and ending around 2010.

1Lone Pine
Why 2010? In my opinion, we are still in this era, for better or worse.

It suggests putting more weight on a plan to get AI Research globally banned.   I am skeptical that this will work (though if burning all GPUs would be a pivotal act the chances of success are significantly higher), but it seems very unlikely that there is a technical solution either.

In addition, at least some  purported technical solutions to AI risk seem to meaningfully increase the risk to humanity.  If you have someone creating an AGI to exercise sufficient control over the world to execute a pivotal act, that raises the stakes of being first enormously which incentivizes cutting corners.  And, it also makes it more likely that the AGI will destroy humanity and be quicker to do so. 

Load More