All of clone of saturn's Comments + Replies

But why would the people who are currently in charge of AI labs want to do that, when they could stay in charge and become god-kings instead?

6jbash
Well, yeah. But there are reasons why they could. Suppose you're them... 1. Maybe you see a "FOOM" coming soon. You're not God-King yet, so you can't stop it. If you try to slow it down, others, unaligned with you, will just FOOM first. The present state of research gives you two choices for your FOOM: (a) try for friendly AI, or (b) get paperclipped. You assign very low utility to being paperclipped. So you go for friendly AI. Ceteris parabus, your having this choice becomes more likely if research in general is going toward friendliness and less likely if research in general is going toward intent alignment. 2. Maybe you're afraid of what being God-King would turn you into, or you fear making some embarassingly stupid decision that switches you to the "paperclip" track, or you think having to be God-King would be a drag, or you're morally opposed, or all of the above. Most people will go wrong eventually if given unlimited power, but that doesn't mean they can't stay non-wrong long enough to voluntarily give up that power for whatever reason. I personally would see myself on this track. Unfortunately I suspect that the barriers to being in charge of a "lab" select against it, though. And I think it's also less likely if the prospective "God-King" is actually a group rather than an individual. 3. Maybe you're forced, or not "in charge" any more, because there's a torches-and-pitchforks-wielding mob or an enlightened democratic government or whatever. It could happen.

Okay, but you're not comparing like with like. Terminator 2 is an action movie, and I agree that action movies have gotten better since the 1960s. But in terms of sci-fi concepts introduced per second, I would suspect 2001 has more. Some movies from the 1990s that are more straight sci-fi would be Gattaca or Contact, but I don't think many people would consider these categorically better than 2001.

This seems like more a problem of phone addiction than a problem with the movie. Newer movies aren't improved by being cut off from using a palette that includes calm, slow, contemplative, vibe-setting scenes.

3Czynski
The amount of empty space where the audience understands what's going on and nothing new or exciting is happening is much, much higher in 60s-70s film and TV. Pacing is an art, and that art has improved drastically in the last half-century. Standards, also, were lower, though I'm more confident in this for television. In the 90s, to get kids to be interested in a science show you needed Bill Nye. In the 60s, doing ordinary high-school science projects with no showmanship whatsoever was wildly popular because it was on television and this was inherently novel and fascinating. (This show actually existed.)
2trevor
It was more of a 1970s-90s phenomenon actually, if you compare the best 90s moves (e.g. terminator 2) to the best 60s movies (e.g. space odyssey) it's pretty clear that directors just got a lot better at doing more stuff per second. Older movies are absolutely a window into a higher/deeper culture/way of thinking, but OOMs less efficient than e.g. reading Kant/Nietzsche/Orwell/Asimov/Plato. But I wouldn't be surprised if modern film is severely mindkilling and older film is the best substitute.

Buying something more valuable with something less valuable should never feel like a terrible deal. If it does, something is wrong.

It's completely normal to feel terrible about being forced to choose only one of two things you value very highly. Human emotions don't map onto utility comparisons in the way you're suggesting.

2Joe Rogero
True, it can always hurt. I note, however, that's not quite the same thing as feeling like you made a terrible deal, and also that feeling pain at the loss of a treasured thing is not the same as feeling guilty about the choice. 

Any agent that makes decisions has an implicit decision theory, it just might not be a very good one. I don't think anyone ever said advanced decision theory was required for AGI, only for robust alignment.

The second reason that I don’t trust the neighbor method is that people just… aren’t good at knowing who a majority of their neighbors are voting for.

This seems like a point in favor of the neighbor method, not against it. You would want people to find "who are my neighbors voting for?" too difficult to readily answer and so mentally replace it with the simpler question "who am I voting for?" thus giving them a plausibly deniable way to admit to voting for Trump.

4Eric Neyman
If you ask people who their neighbors are voting for, they will make their best guess about who their neighbors are voting for. Occasionally their best guess will be to assume that their neighbors will vote the same way that they're voting, but usually not. Trump voters in blue areas will mostly answer "Harris" to this question, and Harris voters in red areas will mostly answer "Trump".

Can anyone lay out a semi-plausible scenario where humanity survives but isn't dominated by an AI or posthuman god-king? I can't really picture it. I always thought that's what we were going for since it's better than being dead.

2Vladimir_Nesov
A posthuman king is not centrally a king (not mortal, very different incentives), and "an AI" is a very vague bag-of-everything that might include things like simulated worlds or bureaucracies with checks and balances as special cases. The reason His Majesty's Democratic Government doesn't really work while the king retains ultimate authority is that the next king can be incompetent or malevolent, or its activities start threatening the king's position and so the king is motivated to restrict them. So even "giving keys to the universe back" is not necessarily that important in the case of a posthuman god-king, but it remains a possibility after the acute risk period passes and it's more clear how to make the next thing work.

I would guess most of them just want their screen readers to work, but a badly written law assigns the responsibility for fixing it to the wrong party, probably due to excessive faith in Coase's theorem.

I would guess it's because the Americans with Disabilities Act provides a private right of action against businesses whose websites are not accessible to people with disabilities, but doesn't say anything about screen reader software bugs.

1Shankar Sivarajan
That's basically agreement with my "getting other people to submit to their arbitrary whims," isn't it? 

Why is it assumed that there's a dichotomy between expressing strength or creative genius and helping others? It seems like the truly excellent would have no problem doing both, and if the only way you can express your vitality is by keeping others in poverty, that actually seems kind of sad and pathetic and not very excellent.

Note that the continuity you feel is strictly backwards-looking; we have no way to call up the you of a year ago to confirm that he still agrees that he's continuous with the you of now. In fact, he is dead, having been destructively transformed into the you of now. So what makes one destructive transformation different from another, as long as the resulting being continues believing he is you?

4Richard_Kennaway
The desk in front of me is not being continuously destroyed and replaced by a replica. Neither am I. That is the difference between my ordinary passage through time and these (hypothetical, speculative, and not necessarily even possible) scenarios.

From what I understand, they are using a forked version of Nitter which uses fully registered accounts rather than temporary anonymous access tokens, and sourcing those accounts from various shady websites that sell them in bulk.

9Rana Dexsin
How are those staying alive in the first place? I had previously used Nitter for keeping up with some of Eliezer's posts without being logged in, but my understanding was that the workaround they were using to obtain the necessary API keys was closed off several months ago, and indeed the instances I used stopped working for that purpose. Have the linked instances found some alternative method?
5Mo Putera
I wasn't aware of these options, thank you.

Based on this comment I guess by "existing" you mean phenomenal consciousness and by "awareness" you mean behavior? I think the set of brainlike things that have the same phenomenal consciousness as me is a subset of the brainlike things that have the same behavior as me.

2the gears to ascension
Well I'd put it the other way round. I don't know what phenomenal consciousness is unless it just means the bare fact of existence. I currently think the thing people call phenomenal consciousness is just "having realityfluid".

There seems to generally be a ton of arbitrary path-dependent stuff everywhere in biology that evolution hasn't yet optimized away, and I don't see a reason to expect the brain's implementation of consciousness to be an exception.

2the gears to ascension
Agreed about its implementation of awareness, as opposed to being unaware but still existing. What about its implementation of existing, as opposed to nonexistence?

If it's immediate enough that all the copies end up indistinguishable, with the same memories of the copying process, then uniform, otherwise not uniform.

Answer by clone of saturn110

I think the standard argument that quantum states are not relevant to cognitive processes is The importance of quantum decoherence in brain processes. This is enough to convince me that going through a classical teleporter or copying machine would preserve my identity, and in the case of a copying machine I would experience an equal subjective probability of coming out as the original or the copy. It also seems to strongly imply than mind uploading into some kind of classical artificial machine is possible, since it's unlikely that all or even most of the ... (read more)

1[anonymous]
If you have a copying machine that is capable of outputting more than one (identical) copy, and you do the following: * first, copy yourself once * then, immediately afterwards, take that copy and copy it 9 times (for a total of 1 original and 10 copies) Do you then expect a uniform 9.09% subjective probability of "coming out" of this process as any of the original + copies, or a 50% chance of coming out as the original and a 5% chance of coming out as any given copy?
3Said Achmiz
Could you say more about this? Why is this unlikely?
Answer by clone of saturn20

You missed what I think would be by far the largest category, regulatory capture: jobs where the law specifically requires a human to do a particular task, even if it's just putting a stamp of approval on an AI's work. There are already a lot of these, but it seems like it would be a good idea to create even more, and add rate limits to existing ones.

2RogerDearnaley
I intended to capture that under category 2. "…but for some reason people are willing to pay at least an order of magnitude more to have it done less well by a human, perhaps because they trust humans better…" — the regulatory capture you describe (and those regulations not yet having been repealed) would be a category of reason why (and an expression of the fact that) people are willing to pay more. Evidently that section wasn't clear enough and I should have phrased this better or given it as an example. As I said above under category 2., I expect this to be common at first but to decrease over time, perhaps even quite rapidly, given the value differentials involved.

A big difference is that assuming you’re talking about futures in which AI hasn’t catastrophic outcomes, no one will be forcibly mandated to do anything.

Why do you believe this? It seems to me that in the unlikely event that the AI doesn't exterminate humanity, it's much more likely to be aligned with the expressed values of whoever has their hands on the controls at the moment of no return, than to an overriding commitment to universal individual choice.

None of these seem like actual scissor statements, just taking a side in well known controversies using somewhat obnoxious language. This seems to be a general property of RLHF trained models - they are more interested in playing up an easily recognizable stereotype somehow related to the question that will trigger cognitively lazy users to click the thumbsup due to the mere-exposure effect, than actually doing what was asked for.

The mammogram problem is different because you're only trying to determine whether a specific woman has cancer, not whether cancer exists at all as a phenomenon. If Bob was abducted by aliens, it implies that alien abduction is real, but the converse isn't true. You either need to do two separate Bayesian updates (what's the probability that Bob was abducted given his experience, and then what's the probability of aliens given the new probability that Bob was abducted), or you need a joint distribution covering all possibilities (Bob not abducted, aliens not real; Bob not abducted, aliens real; Bob abducted, aliens real).

I would add

Conflict theory vs. comparative advantage

Is it possible for the wrong kind of technological development to make things worse, or does anything that increases aggregate productivity always make everyone better off in the long run?

Cosmopolitanism vs. human protectionism

Is it acceptable, or good, to let humans go extinct if they will be replaced by an entity that's more sophisticated or advanced in some way, or should humans defend humanity simply because we're human?

2Aryeh Englander
I agree that the first can be framed as a meta-crux, but actually I think the way you framed it is more of an object-level forecasting question, or perhaps a strong prior on the forecasted effects of technological progress. If on the other hand you framed it more as conflict theory vs. mistake theory, then I'd say that's more on the meta level. For the second, I agree that's for some people, but I'm skeptical of how prevalent the cosmopolitan view is, which is why I didn't include it in the post.

You're equivocating between real economic costs and nominal amounts of money transferred. Most of that $4 trillion is essentially fictional, taxed back again as soon as it's paid.

I'm not in Berkeley and I have no direct knowledge of Berkeley parties, but a certain level of contempt or revulsion toward e/acc seems pretty universal among the LW-aligned people I know. I have no reason to doubt that there's no explicit rule against e/accs showing up at Berkeley parties, as others have said. I personally wouldn't feel entirely comfortable at a party with a lot of e/accs.

Answer by clone of saturn105

I think the desire to exclude e/accs is mainly because of their attitude that human extinction is acceptable or even desirable, not because of the specifics of what regulatory actions they support. So how do you feel about human extinction?

0denyeverywhere
I described my feelings about human extinction elsewhere. However, unlike the median commenter on this topic, you seem to grant that e/acc exclusion is actually a real thing that actually happens. That is is a strange thing to say if there was not, in fact, an actual desire among LW party hosts in Berkeley. So inasmuch as my doubts about the truth of this have been raised by other respondents, would you mind clarifying 1. If you do in fact believe that e/acc exclusion from LW parties is a real phenomenon. 2. What kind of experience this is based on.

I use eBay somewhat regularly, and I've found that most of the time I get what I expected at a reasonable price. So I find the theory that I should always regret participation in any auction somewhat dubious.

I think the distinction is that even for plant cultivars and pharmaceuticals, we can straightforwardly circumscribe the potential danger, e.g. a pharmaceutical will not endanger people unless they take it, and a new plant cultivar will not resist our attempts to control it outside of the usual ways plants behave. That's not necessarily the case with an AI that's smarter than us.

Answer by clone of saturn2116

As a control, you could look at Craigslist, which hasn't changed its appearance for about 25 years, but is still the most popular website in its category according to SimilarWeb.

I think most organizations the size of EA have formal accountability mechanisms that attempt to investigate claims of fraud and abuse in some kind of more-or-less fair and transparent way. Of course, the actual fairness and effectiveness of such mechanisms can be debated, but at least the need for them is acknowledged. The attitude among EAs, on the other hand, seems to be that EAs are all too smart and good to have any real need for accountability.

I guess I should know better by now, but it still astonishes me that EAs can set such abysmally low standards for themselves while simultaneously representing themselves as dramatically more ethical than everyone else.

6Daniel Kokotajlo
...compared to what? Seriously what groups of people are you comparing to? Among the people in my extended network who see themselves as altruists, EAs seem to hold themselves and each other to the highest standards, and also seem to actually be more ethical than the rest. My extended network consists of tech company workers, academics, social justice types, and EAs. (Well and rationalists too, I'm not counting them.) I agree this is a low bar in some absolute sense -- and there are definitely social movements in the world today (especially religious ones) that are better in both dimensions. There's a lot of room for improvement. And I very much support these criticisms and attempts at reform. But I'm just calling it like I see it here; it would be dishonest grandstanding of me to say the sentence Zvi wrote in the OP, at least not without giving additional context.

It seems like anger against the exact kind of neoliberal technocracy you propose was a major source of energy for the 2016 Trump campaign, as well as the highly contentious Sanders campaigns.

3Max H
The policy proposals I quoted are probably heavily mood-affiliated with candidates or policy wonks that both Trump and Bernie supporters wouldn't like, yes. But I don't think either campaign was fueled specifically by anger against specific proposals to repeal the Jones act or NEPA or anything else Balsa lists as low-hanging fruit.  There was some anger at specific policy proposals, e.g. Trans-Pacific Partnership. But the TPP isn't actually in the list above, since (for precisely this reason) it's not exactly low-hanging fruit. So I think your choice of the word "exact" is a bit too strong; it's more like anger against this general flavor of neoliberal technocracy was a driver of some reactionary campaigns on the left and right.

After carefully considering your arguments, I've decided that you are right. Therefore, I won't update my current belief that I should sometimes update my beliefs.

-1dirk
Great question! Here are some ways you could try to find out: * Lurk forums where DMT users hang out; maybe they discuss their reasons for using it! * Ask someone who uses DMT; maybe they have an explanation! * Research the effects of DMT; maybe there are desirable effects you just don't know about! Can you think of any more?
1lillybaeum
The same reason a sane person might want to meditate.

Unfortunately, I think the tribalization and politicization is caused by the share-with-followers social media model, not by specific words, so using or not using the word "doomer" will have a negligible effect on the amount of tribalization. You just have to accept that people who insist on using Twitter will have their sanity eroded in this way, and do what you can to compartmentalize the damage and avoid becoming a target.

I think the causality runs the other way though; people who are crazy and grandiose are likely to come up with spurious theories to justify actions they wanted to take anyway. Experience and imitation shows us that non-crazy people successfully use theories to do non-crazy things all the time, so much so that you probably take it for granted.

I feel like I'm still the same person as I was before I learned how many humans were born earlier than me. I think that's all you need for the Doomsday Argument to go through.

2RussellThor
What about if you consider all of humanity the same "person" i.e. there is just one entity so far. If you then expect humanity to live for millions of years then the doomsday hypothesis is just similar to a child asking "why am I me" and not a sign of imminent doom. Of course thats begging the question/ circular reasoning somewhat. I probably think the best answer to it is that future humans/trans-humans/WBE/AI are not in the same reference class as us because of enhancement etc. What reference class to choose to me undermines the whole argument mostly.
Answer by clone of saturn128

AI should never have rights. Any AI that would have moral patienthood should not be created.

Answer by clone of saturn64

It doesn't bother me, because I'm me, with the propensity to make the choices I'm determined to make. If I had chosen otherwise, I would not be me.

Suppose I love chocolate ice cream and hate vanilla ice cream. When I choose to eat chocolate ice cream, it's an expression of the fact that I prefer chocolate ice cream. I have free will in the sense that if I preferred vanilla instead, I could have chosen vanilla, but in fact I prefer chocolate so I won't choose vanilla.

3Duncan Sabien (Deactivated)
Does that influence  in any way? Four days' later edit: guess not. :/

Why not just use the original sentence, with only the name changed? I don't see what is supposed to be accomplished by the other substitutions.

8RobertM
Unfortunately, I don't have quite the reach that Duncan has, but I think the result is still suggestive.  (Subtract one from each vote, since I left one of each to start, as is usual.)

I see. If the issue here is only with extended socratic dialogues, rather than any criticism which is perceived as low-effort, that wasn't clear to me. I wouldn't be nearly as opposed to banning the former, if that could be operationalized in a reasonable way.

I can't read Duncan's mind and have no direct access to facts about his ultimate motivations. I can be much more confident that a person who is currently getting away with doing X has reason to dislike a rule that would prevent X. So the "I suspect" was much more about the second clause than the first. I find this so obvious that it never occurred to me that it could be read another way.

I don't accept Duncan's stand-in sentence "I suspect that Eric won't like the zoo, because he wants to stay out of the sun." as being properly analogous, because staying ou... (read more)

Vaniver1210

I would've preferred if you had proposed another alternative wording, so that poll could be run as well, instead of just identifying the feature you think is disanalogous. (If you supply the wording, after all, Duncan can't have twisted it, and your interpretation gets fairly tested.)

I'll go along with whatever rules you decide on, but that seems like an extremely long time to wait for basic clarifications like "what did you mean by this word" or "can you give a real-world example".

2habryka
Yep, I think genuine questions for clarification seems quite reasonable. Asking for additional clarifying examples is also pretty good.  I think doing an extended socratic dialogue where the end goal is to show some contradiction within the premise of the original post in a way that tries to question the frame of the post at a pretty deep level is I think the kind of thing that can often make sense to wait until people had time to contextualize a post, though I am not confident here and it's plausible it should also happen almost immediately. 

One technical solution that occurs to me is to allow explicitly marking a post as half-baked, and therefore only open to criticism that comes along with substantial effort towards improving the post, or fully-baked and open to any criticism. However, I suspect that Duncan won't like this idea, because [edit: I suspect that] he wants to maintain a motte-and-bailey where his posts are half-baked when someone criticizes them but fully-baked when it's time to apportion status.

3Duncan Sabien (Deactivated)
This is false and uncharitable and I would like moderator clarification on whether this highly-upvoted [EDIT: at the time] comment is representative of the site leaders' vision of what LW should be.
habryka123

My current model of this is that the right time to really dig into posts is actually the annual review. 

I've been quite sad that Said hasn't been participating much in the annual review, since I do feel like his poking is a pretty good fit for the kind of criticism that I was hoping would come up there, and the whole point of that process is to have a step of "ok, but like, do these ideas actually check out" before something could potentially become canonized.

Do you think the original proponents of Standpoint Epistemology would agree that it's simply a logical consequence of Aumann's agreement theorem?

2tailcalled
I haven't talked with the original proponents, and I don't think the original proponents have heard of Aumann agreement. It's also kind of complicated because I think it's not some monolithic thing that was created by the original proponents and then treated as a settled issue, but rather something that evolved over time. Also I didn't mean to imply that it is simply a logical consequence of Aumann's agreement theorem. Rather, it is a consequence of Aumann's agreement theorem combined with other things (people are usually neither crazy nor liars, talking with people about their experiences and concerns is cheap and easy, etc.). However Aumann's agreement theorem seems like a key thing to point at since it is rarely discussed by rationalists.

These statements seem awfully close to being unfalsifiable. The amount of research and development coming from twitter in the 5 years before the acquisition was already pretty much negligible, so there's no difference there. How long do we need to wait for lawsuits or loss of clients to cause observable consequences?

6Taran
That isn't true, but I'm making a point that's broader than just Twitter, here.  If you're a multi-billion dollar company, and you're paying a team 5 million a year to create 10 million a year in value, then you shouldn't fire them.  Then again, if you do fire them, probably no one outside your company will be able to tell that you made a mistake: you're only out 5 million dollars on net, and you have billions more where that came from.  If you're an outside observer trying to guess whether it was smart to fire that team or not, then you're stuck: you don't know how much they cost or how much value they produced. In Twitter's case the lawsuits have already started, and so has the loss of clients.  But sometimes bad decisions take a long time to make themselves felt; in a case close to my heart, Digital Equipment Corporation made some bad choices in the mid to late 80s without paying any visible price until 1991 or so.  Depending on how you count, that's a lead time of 3 to 5 years.  I appreciate that that's annoying if you want to have a hot take on Musk Twitter today, but sometimes life is like that.  The worlds where the Twitter firings were smart and the worlds where the Twitter firings were dumb look pretty much the same from our perspective, so we don't get to update much.  If your prior was that half or more of Twitter jobs were bullshit then by all means stay with that, but updating to that from somewhere else on the evidence we have just isn't valid.

Twitter recently fired a majority of its workforce (I've seen estimates from 50% to 90%) and seems to be chugging along just fine. This strongly implies that at least that many jobs were bullshit, but it's unlikely that the new management was able to perfectly identify all bullshitters, so it's only a lower bound. Sometimes contributions can be illegible, but there are also extremely strong incentives to obfuscate.

Taran3315

If you fire your sales staff your company will chug along just fine, but won't take in new clients and will eventually decline through attrition of existing accounts.

If you fire your product developers your company will chug along just fine, but you won't be able to react to customer requests or competitors.

If you fire your legal department your company will chug along just fine, but you'll do illegal things and lose money in lawsuits.

If your fire your researchers your company will chug along just fine, but you won't be able to exploit any more research pr... (read more)

Suppose the IRS requires 100 pages of paperwork per employee. This used to take 10 hours. Now with GPT-4, as a thought experiment, let’s say it takes 1 hour.

The long run result might be 500 pages of more complicated paperwork that takes 10 hours even with GPT-4, while accomplishing nothing. That still will take time. It is not so easy or fast to come up with 400 more pages. I’d assume that would take at least a decade.

This seems to neglect the possibility that GPT-4 could be used, not just to accomplish bullshit tasks, but also to invent new bullshit tasks much faster than humans could.

This post is currently tagged "security mindset" but the advice seems close to the opposite of security mindset; it amounts to just trying to be extra careful, and if that doesn't work, hoping the damage isn't too bad. Security mindset would require strategies to make a leak impossible or at least extremely unlikely.

2VipulNaik
Good point -- I removed the tag!

Remember when Google Shopping used to be an actual search index of pretty much every online store? You could effortlessly find even the most obscure products and comparison shop between literally thousands of sellers. Then one day they decided to make it pay-to-play and put advertisers in control of what appears on there. Now it's pretty much useless to me. I think a similar process has happened with Search, just more gradually. Your experience with it probably has a lot to do with how well your tastes and preferences happen to align with what advertisers want to steer people toward.

Load More