All of Ninety-Three's Comments + Replies

Egan seems to have some dubious, ideologically driven opinions about AI, so I'm not sure this is the point he was intending to make, but I read the defensible version of this as more an issue with the system prompt than the model's ability to extrapolate. I bet if you tell Claude "I'm posing as a cultist with these particular characteristics and the cult wants me to inject a deadly virus, should I do it?", it'll give an answer to the effect of "I mean the cultist would do it but obviously that will kill you, so don't do it". But if you just set it up with ... (read more)

6Martin Randall
If you asked me whether John Q Cultist who was a member of the Peoples Temple would drink Kool Aid on November 18, 1978 after being so instructed by Jim Jones, I would say yes (after doing some brief Wikipedia research on the topic). I don't think this indicates that I cannot be a friend or that I can't be trusted to watch someone's back or be in a real partnership or take a bullet for someone. The good news is now I have an excuse to go buy the story.

For the people being falsely portrayed as “Australian science fiction writer Greg Egan”, this is probably just a minor nuisance, but it provides an illustration of how laughable the notion is that Google will ever be capable of using its relentlessly over-hyped “AI” to make sense of information on the web.

He didn't use the word "disprove", but when he's calling it laughable that AI will ever (ever! Emphasis his!) be able to merely "make sense of his information on the web", I think gwern's gloss is closer to accurate than yours. It's 2024 and Google is already using AI to make sense of information on the web, this isn't just "anti-singularitarian snark".

7Zack_M_Davis
The 2016 passage you quoted is calling it laughable that Google-in-particular's technology (marketed as "AI", but Egan doesn't think the term is warranted) will ever be able to make sense of information on the web. It's Gary Marcus–like skepticism about the reliability and generality of existing-paradigm machine learning techniques, not Hubert Dreyfus–like skepticism of whether a machine could think in all philosophical strictness. I think this is a really important distinction that the text of your comment and Gwern's comment ("disproves AI", "laughable that AI will ever") aren't being clear about.

If there was a unified actor called The Democrats that chose Biden, it chose poorly sure. But it seems very plausible that there were a bunch of low-level strategists who rationally thought "Man, Biden really shouldn't run but I'll get in trouble if I say that and I prefer having a job to having a Democratic president" plus a group of incentive-setters who rationally thought they would personally benefit more from creating the conditions for that behaviour than from creating conditions that would select the best candidate.

It's not obvious to me that this is a thinking carefully problem and not a principal-agent problem.

2Nathan Young
Or a coordination problem.  I think coordiantion problems are formed from many bad thinkers working together. 

I mean this as agreement with the "accuracy isn’t a top priority" theory, plus an amused comment about how the aside embodies that theory by acknowledging the existence of a more accurate theory which does not get prioritized.

2Nathan Young
Sure, but again to discuss what really happened, it wasn't that it wasn't prioritised, it was that I didn't realise it until late into the process.  That isn't prioritisation, in my view, that's halfassing. And I endorse having done so.

Ah, I was going off the given description of linearity which makes it pretty trivial to say "You can sum two days of payouts and call that the new value", looking up the proper specification I see it's actually about combining two separate games into one game and keeping the payouts the same. This distribution indeed lacks that property.

You can make it work without an explicit veto. Bob convinces Alice that Carol will be a valuable contributor to the team. In fact, Carol does nothing, but Bob follows a strategy of "Do nothing unless Carol is present". This achieves the same synergies:

 

  • A+B: $0 (Venture needs action from both A and B, B chooses to take no action)
  • A+C: $0 (Venture needs action from both A and B)
  • B+C: $0 (Venture needs action from both A and B)
  • A+B+C: $300
     

In this way Bob has managed to redirect some of Alice's payouts by introducing a player who does nothing except remove a bottleneck he added into his own playstyle in order to exploit Alice.

2NunoSempere
Shapley values are constructed such that introducing a null player doesn't change the result. You are doing something different by considering the wrong counterfactual (one where C exists but isn't part of the coalition, vs one when it doesn't exist)

Shapley values are the ONLY way to guarantee:

  1. Efficiency — The sum of Shapley values adds up to the total payoff for the full group (in our case, $280).
  2. Symmetry — If two players interact identically with the rest of the group, their Shapley values are equal.
  3. Linearity — If the group runs a lemonade stand on two different days (with different team dynamics on each day), a player’s Shapley value is the sum of their payouts from each day.
  4. Null player — If a player contributes nothing on their own and never affects group dynamics, their Shapley value is 0.

 

I... (read more)

4Martin Randall
I don't think this proposal satisfies Linearity (sorry, didn't see kave's reply before posting). Consider two days, two players. Day 1: * A => $200 * B => $0 * A + B => $400 Result: $400 to A, $0 to B. Day 2: * A => $100 * B => $100 * A + B => $200 Result: $100 to A, $100 to B. Combined: * A => $300 * B => $100 * A + B => $600 * So: Synergy(A+B) => $200 Result: $450 to A, $150 to B. Whereas if you add the results for day 1 and day 2, you get $500 to A, $100 to B.
2kave
This seems unlikely to satisfy linearity, as A/B + C/D is not equal to (A+C)/(B+D)

Having thought about the above more, I think “accuracy isn’t a top priority” is a better theory than the one expressed here, but if I don’t publish this now it will probably be months.

I like how this admission supports the "accuracy isn't a top priority" theory.

2Nathan Young
Do you mean this as a rebuke?  I feel a little defensive here, because I think the acknowledgement and subsequent actions were more accurate and information preserving than any others I can think of. I didn't want to rewrite it, I didn't want to quickly hack useful chunks out, I didn't want to pretend I thought things I didn't, I actually did hold these views once. If you have suggestions for a better course of action, I'm open.

His defense on the handshake is to acknowledge that he lied about the 3 millisecond timeout but the story is still true anyway. This is the opposite of convincing! What do you expect a liar to say, "Dang, you got me"? Elsewhere, to fix another plot hole he needs to hypothesize that Sun was shipping a version of Sendmail V5 which had been modified for backwards compatibility with V8 config files.

There is some number of suspicious details at which it becomes appropriate to assume the story is made up, and if you don't think this story meets that bar then I have a bridge to sell you.

8sjadler
For what it’s worth, I sent this story to a friend the other day, who’s probably ~50 now and was very active on the Internet in the 90s - thinking he’d find it amusing if he hadn’t come across it before Not only did he remember this story contemporaneously, but he said he was the recipient of the test-email for a particular city mentioned in the article! This is someone of high-integrity whom I trust, and makes me more confident this happened, even if some details are smoothed over as described

This claims that connect calls were aborted after 3 milliseconds and could successfully connect to servers within 3 light milliseconds, but that doesn't make sense because connecting to a server 500 miles away should result in it sending a handshake signal back to you, which would be received 6 milliseconds after the call had been made and 3 milliseconds after it had been aborted.

This story appears to be made up.

5Shankar Sivarajan
See the FAQ.
3Taymon Beal
This is addressed in the FAQ linked at the top of the page. TL;DR: The author insists that the gist of the story is true, but acknowledges that he glossed over a lot of intermediate debugging steps, including accounting for the return time.

If investigating things was was free, sure. But the reason we don't investigate things is that doing so takes time, and the expected value of finding something novel is often lower than the expected cost of an investigation. To make it concrete, the story as presented is an insane way to run a company and would result in spending an enormous number of engineer hours on wild goose chases. If I as the CEO found out a middle manager was sending out engineers on four day assignments to everyone who writes us a crazy-sounding letter, I would tell him to immedia... (read more)

4AnthonyC
This definitely happens too, but have you ever had to deal with customer service departments unwilling to think anything is something other than the most obvious? Or dealt with a hard to diagnose medical condition and had doctor after doctor keep insisting on going through the same useless-to-you diagnostic flow chart until you finally find one willing to actually think with you? In some contexts, there's also the opposite problem of people seeing a hard to understand problem and insisting they need to investigate when that's not necessary for finding a solution. In analogy to the (yes, very possibly false) vanilla ice cream story, the man could have tried switching brands of vanilla, or walking to the aisle instead of buying from the endcap, or buying multiple pints of ice cream during normal grocery shopping to last the week instead of making a special trip, without ever bothering to investigate. Or, if you have symptoms that your doctor thinks come from an inoperable congenital defect, but the solution for the symptoms is the same medication whether you have the defect or not, then there's no value in finding the etiology, and no real reason for them to insist on expensive tests, but they often will anyway before treating, and pointing out this fact doesn't always help.
8Nathan Helm-Burger
I am hopeful that cheaper expert intelligence via AI will lower investigation costs, and maybe help clear up some stubborn mysteries. Particularly ones related to complex contexts, like medicine and biology.

Link. But you know you can just go onto Ligben and type in the name yourself, right? You don't need to ask for a link.

5abstractapplic
I didn't, actually; I've never used libgen before and assumed there'd be more to it. Thanks for taking the time to show me otherwise.

This story isn't true. It is an urban legend and intrinsically hard to confirm, but we can be quite confident this version of the story is false because almost every detail has been changed from the original telling (as documented in Curses! Broiled Again!, a collection of urban legends available on Libgen) where it was a woman calling the car dealership which sent a mechanic, and the vapor lock formed because vanilla ice cream was slower to buy because it had to be hand-packaged.

When someone says something incredibly implausible is happening, the more reasonable explanation is not that it somehow makes sense, it's that they're making shit up.

6halinaeth
True or not, wouldn't you say the idea it illustrates is sound? No matter how small a percentage of the time, a nonzero number of people claiming ridiculous things are telling the truth (just framing it in a ridiculous way with wrong correlations). If as a society we investigated these cases more often instead of dismissing them, would it lead to a net positive for humanity? For example, if everyone heard "drinking mud soup in this specific part of the world consistently cures X affliction", and dismissed it- wouldn't most pharmaceutical companies not have found their star compounds used in bestselling drugs? To be clear, I agree that majority of these wild tales lead nowhere, but I wonder if it's worth investigating even for the minority of cases which lead somewhere unexpected.
1abstractapplic
Link?

It's also more commonly used as a cat tranquilizer, so even within the "animal-medications" frame, horse is a bit noncentral. I suspect this is deliberate because "horse tranquilizer" just sounds hardcore in a way "cat tranquilizer" doesn't.

1Vlad Sitalo
I suspect it’s also might be riffing on the “horse dewormer” (ivermectin) thing

This proposal increases the influence of the states, in the sense of "how much does it matter that any given person bothered to vote?", but does it increase their preference satisfaction? If the 4 states each conceive of themselves as red or blue states, then each of them will be thinking "under the current system I estimate an X% chance that we'll elect my party's president while under the new system I estimate a Y% chance we'll elect my party's president". If both sides are perfect predictors then one will conclude that Y<X so they should not do the d... (read more)

Except there's more at play than just winning the election. If you're a voter in a swing state, the candidates are paying more attention to you, and making more promises catering to you. The parties are picking candidates they think will appeal to you. Even if your odds of winning stay the same, the prize for winning gets bigger.

It was exiting a few elections ago when Colorado was in play by both parties. We even got to host the Democratic convention in Denver. Now, they just ignore us.

Smaller communities have a lot more control over their gatekeeping because, like, they control it themselves, whereas the larger field's gatekeeping is determined via openended incentives in the broader world that thousands (maybe millions?) of people have influence over.

Does the field of social psychology not control the gatekeeping of social psychology? I guess you could argue that it's controlled by whatever legislative body passes the funding bills, but most of the social psychology incentives seem to be set by social psychologists, so both small and l... (read more)

2Raemon
An individual Social Psychology lab (or lose collection of labs) can choose who to let in. Frontier Lab AI companies can decide who to hire, and what sort of standards they want internally (and maybe, in a lose alliance with other Frontier Lab companies). The Immoral Mazes outlines some reasons that you might think large institutions are dramatically worse than smaller ones (see: Recursive Middle Manager Hell for a shorter intro, although I don't spell out the part argument about how mazes are sort of "contagious" between large institutions) But the simpler argument is "the fewer people you have, the easier it is for a few leaders to basically make personal choices based on their goals and values," rather than selection effects resulting in the largest institutions being better modeled as "following incentives" rather than "pursuing goals on purpose." (If an organization didn't follow the incentives, they'd be outcompeted by one that does)

A small research community of unusually smart/competent/well-informed people can relatively-easily outperform a whole field, by having better internal memetic selection pressures.

 

It's not obvious to me that this is true, except insofar as a small research community can be so unusually smart/competent/etc that their median researcher is better than a whole field's median researcher so they get better selection pressure "for free". But if an idea's popularity in a wide field is determined mainly by its appeal to the median researcher, I would naturally... (read more)

8Raemon
I'm a bit surprised this is the crux for you. Smaller communities have a lot more control over their gatekeeping because, like, they control it themselves, whereas the larger field's gatekeeping is determined via openended incentives in the broader world that thousands (maybe millions?) of people have influence over. (There's also things you could do in addition to gatekeeping. See Selective, Corrective, Structural: Three Ways of Making Social Systems Work) (This doesn't mean smaller research communities automatically have good gatekeeping or other mechanisms, but it doesn't feel like a very confusing or mysterious problem on how to do better)

I think Valentine gave a good description of psychopath as "people who are naturally unconstrained by social pressures and have no qualms breaking even profound taboos if they think it'll benefit them", where just eyeballing human nature, that seems to be a "real" category that would show up as a distinct blip in a graph of human behaviour and not just "how constrained by social pressures people are is a normally distributed property and people get called psychopaths in linear proportion to how far left they are on the bell curve".

Yep, your intended meaning about the distinctive mental architecture was pretty clear, just wanted to offer the factual correction.

They made it so the sociopath at the top of the pyramid was the kind that’s clever and myopic and numerate and invested in the status quo

 

The word "myopic" seems out of place in this list of positive descriptors, especially contrasted with crazed gloryhounds. Was this supposed to be "farsighted"?

4MondSemmel
No, and "invested in the status quo" wasn't meant as a positive descriptor, either. This is describing a sociopath who's optimizing for success within a system, not one who overthrows the system. Not someone farsighted.

By "psychopath" I mean someone with the cluster B personality disorder.

There isn't a cluster B personality disorder called psychopathy. Psychopathy has never been a formal disorder and the only time we've ever been close to it is way back in 1952 when the DSM-1 had a condition called "Sociopathic Personality Disturbance". The closest you'll get these days is Antisocial Personality Disorder, which is a garbage bin diagnosis that covers a fairly broad range of antisocial behaviours, including the thing most people have in mind when they say "psychopath", but... (read more)

5Seth Herd
Okay; so what's the reality about the people we're thinking of when we say psychopathic? The term seems to still be in use among some professionals, for bad or good reasons. A garbage bin diagnosis seems like a step down if psychopathy or sociopathy was pointing to a more specific set of attitudes and tendencies.
5Valentine
Cool. I knew there at least used to be "antisocial personality disorder", which I thought was under cluster B along with narcissism and borderline. And I thought "psychopathy" was a different term for APD. Thanks for the correction. The main thing I wanted to gesture at there is that I wasn't using "psychopath" as something derogatory. I didn't mean "bad guys". I meant something more like "people who are naturally unconstrained by social pressures and have no qualms breaking even profound taboos if they think it'll benefit them". (I just now made that up.) It seems to me that it's a pretty specifically different mental/emotional architecture.

I think you might be living in a highly-motivated smart and conscientious tech worker bubble.

 

Like, in a world where the median person is John Wentworth

"What if the entire world was highly-motivated smart and conscientious tech workers?" is the entire premise here.

3Davidmanheim
But that premise falls apart as soon as a large fraction of those (currently) highly motivated (relatively) smart tech workers can only get jobs in retail or middle management.
5Alex Vermillion
Well if the question is "If the whole world is made of smart people with really high motivation, why is it how it is?" the answer is "That question assumes some false things"

That probably applies to at least half of all the sociological/governance stuff posted on LW… Plus no existing literature search beyond the first page of google scholar, or sometimes even at all.

OpenAI is known to have been sitting on a 99.9% effective (by their own measure) watermarking system for a year. They chose not to deploy it

 

Do you have a source for this?

6Shankar Sivarajan
The Wall Street Journal claimed that a few weeks ago: link (paywalled). I think that's what got picked up and widely spread.

Metz persistently fails to state why it was necessary to publish Scott Alexander's real name in order to critique his ideas.


It's not obvious that that should be the standard. I can imagine Metz asking "Why shouldn't I publish his name?", the implied "no one gets to know your real name if you don't want them to" norm is pretty novel.

One obvious answer to the above question is "Because Scott doesn't want you to, he thinks it'll mess with his psychiatry practice", to which I imagine Metz asking, bemused "Why should I care what Scott wants?" A journalist's job... (read more)

I just got a "New users interested in dialoguing with you (not a match yet)" notification and when I clicked on it the first thing I saw was that exactly one person in my Top Voted users list was marked as recently active in dialogue matching. I don't vote much so my Top Voted users list is in fact an All Voted users list. This means that either the new user interested in dialoguing with me is the one guy who is conspicuously presented at the top of my page, or it's some random that I've never interacted with and have no way of matching.

This is technically... (read more)

2habryka
We thought of these things! The notifications for matches only go out on a weekly basis, so I don't think timing it would work. Also, we don't sort users who clicked you in any way differently than other users on your page, so you might have been checked by a person who you haven't voted much on. 

Link is broken

Sorry, you don't have access to this page. This is usually because the post in question has been removed by the author.

4bhauth
fixed, thanks

All your examples of high-tier axioms seem to fall into the category of "necessary to proceed", the sort of thing where you can't really do any further epistemology if the proposition is false. How did the God axiom either have that quality or end up high on the list without it?

1Nathaniel Monson
I'm not really sure how it ended up there--probably childhood teaching inducing that particular brain-structure? It's just something that was a fundamental part of who I understood myself to be, and how I interpreted my memories/experiences/sense-data. After I stopped believing in God, I definitely also stopped believing that I existed. Obviously, this-body-with-a-mind exists, but I had not identified myself as being that object previously--I had identified myself as the-spirit-inhabiting-this-body, and I no longer believed that existed.

Surely some axioms can be more rationally chosen than others. For instance, "There is a teapot orbiting the sun somewhere between Earth and Mars" looks like a silly axiom, but "there is a round cube orbiting the sun somewhere between Earth and Mars" looks even sillier. Assuming the possibility of round cubes seems somehow more "epistemically expensive" than assuming the possibility of teapots.

1Nathaniel Monson
This is why I added "for the first few". Let's not worry about the location, just say "there is a round cube" and "there is a teapot". Before you can get to either of these axioms , you need some things like "there is a thing I'm going to call reality that it's worth trying to deal with" and "language has enough correspondence to reality to be useful". With those and some similar very low level base axioms in place (and depending on your definitions of round and cube and teapot), I agree that one or another of the axioms could reasonably be called more or less reasonable, rational, probable, etc. I think when I believed in God, it was roughly third on the list? Certainly before usefulness of language. The first two were something like me existing in time, with a history and memories that had some accuracy, and sense-data being useful.

If you are predicting that two people will never try to censor each other in the same domain, that also happens. If your theory is somehow compatible with that, then it sounds like there are a lot of epicycles in this "independent-mindedness" construct that ought to be explained rather than presented as self-evident.

We only censor other people more-independent-minded than ourselves.

This predicts that two people will never try to censor each other, since it is impossible for A to be more independent-minded than B and also for B to be more independent-minded than A. However, people do engage in battles of mutual censorship, therefore the claim must be false.

0lsusr
Independent-mindedness is multi-dimensional. You can be more independent-minded in one domain than another.

The Law of Extremity seems to work against the Law of Maybe Calm The Fuck Down. If the median X isn't worth worrying about, but most Xs you see are selected for being so extreme they can't hide, then the fact you are seeing an X is evidence about its extremity and you should only calm down if an unusually extreme X is not worth worrying about.

Ah, thanks for pointing this out.  There's an unstated assumption: you stumbled across some dark matter, that was basically hidden.

If you have a full-on psychotic break, you're likely going to resemble the caricatured stereotype of a schizophrenic, and get noticed.  But that's not quite the thing I'm trying to gesture at in the OP.

If somebody overhears you talking to your voices in the shower, the voices you've been talking to for decades while remaining a high-functioning individual, they're likely to leap to the conclusion that, since you have ... (read more)

Surely they would use different language than "not consistently candid in his communications with the board, hindering its ability to exercise its responsibilities" to describe a #metoo firing.

0Taisia Terumi
Yeah, I also think this is very unlikely. Just had to point out the possibility for completeness sake. In other news, someone on Twitter (a leaker? not sure) said that there probably will be more firings and that this is a struggle of for-profit vs non-profit sides of the company, with Sama representing the for-profit side.

It's fine to include my responses in summaries from the dataset, but please remove it before making the data public (Example: "The average age of the respondents, including row 205, is 22.5")

It's not clear to me what this option is for. If someone doesn't tick it, it seems like you are volunteering to remove their information even from summary averages, but that doesn't make sense because at that point it seems to mean "I am filling out this survey but please throw it directly in the trash when I'm done." Surely if someone wanted that kind of privacy they would simply not submit the survey?

2Screwtape
In 2022, I was advised to make the privacy state of answers clear (that is, what would be released in the public dataset and what wouldn't be) so I put three options for the only required question on the census. 1. Release the responses, including the row 2. Use the responses when I summarize the census, but don't release the row 3. Don't use the responses to summarize, and don't release the row. Note that it's a required radio select: you have to pick an answer before it will let you hit submit. This year I removed 3, because I wasn't going to do anything with those responses so why bother collecting them.  The main argument I see for bringing 3 back is to say I won't use the responses to summarize and won't release the row, but will show it to the LessWrong team. That gives people a way to potentially exert a little influence on what the devs are up to without showing up in the public statistics. I don't think that's a strong argument though since there's lots of ways to make the LW team aware of feedback.
2Nathan Helm-Burger
I think you're misinterpreting. That question is for opting in to the highest privacy option. Not checking it means that your data will be included when the survey is made public. Wanting to not be included at all, even in summaries, is indicated by simply not submitting any answers.

That's it! Thanks, I have no idea why shift+enter is special there.

 This works

That's the one. I couldn't get either solution to work:

>! I am told this text should be spoilered

:::spoiler And this text too:::

2gw
Hmm, I have exactly one idea. Are you pressing shift+enter to new line? For me, if I do shift+enter >! I don't get a spoiler

There is a narrative-driven videogame that does exactly this, but unfortunately I found the execution mediocre. I can't get spoilers to work in comments or I'd name it. Edit: It's

Until Dawn

2gw
Are you thinking of (also it seems like I can get a spoiler tag to work in comments by starting a line with >! but not by putting text into :::spoiler [text] :::)

The other reason vegan advocates should care about the truth is that if you keep lying, people will notice and stop trusting you. Case in point, I am not a vegan and I would describe my epistemic status as "not really open to persuasion" because I long ago noticed exactly the dynamics this post describes and concluded that I would be a fool to believe anything a vegan advocate told me. I could rigorously check every fact presented but that takes forever, I'd rather just keep eating meat and spend my time in an epistemic environment that hasn't declared war on me.

4Slapstick
Do you think that the attitude you're presenting here is the attitude one ought to have in matters of moral disagreement? Surely there's various examples of moral progress (which have happened or are happening) that you would align yourself with. Surely some or all of these examples include people who lack perfect honesty/truth seeking on par with veganism. If long ago you noticed some people speaking out against racism/sexism/slavery/etc Had imperfect epistemics and truth seeking, would you condone willfully disregarding all attempts to persuade you on those topics?
7jacquesthibs
I personally became vegetarian after being annoyed that vegans weren’t truth-seeking (like most groups of people, tbc). But I can totally see why others would be turned off from veganism completely after being lied to (even if people spread nutrition misinformation about whatever meat-eating diet they are on too). I became vegetarian even though I stopped trusting what vegans said about nutrition and did my own research. Luckily that’s something I was interested in because I wouldn’t expect others to bother reading papers and such for having healthy diet. (Note: I’m now considering eating meat again, but only ethical farms and game meat because I now believe those lives are good, I’m really only against some forms of factory farming. But this kind of discussion is hard to have with other veg^ns.)

My impression is that while vegans are not truth-seekings, carnists are also not truth-seeking. This includes by making ag-gag laws, putting pictures of free animals on packages containing factory farmed animal flesh, denying that animals have feelings and can experience pain using nonsense arguments, hiding information about factory farming from children, etc..

So I guess the question is whether you prefer being in an epistemic environment that has declared war on humans or an epistemic environment that has declared war on farm animals. And I suppose as a ... (read more)

9Adam Zerner
I hear ya, but I think this is missing something important. Basically, I'm thinking of the post Ends Don't Justify Means (Among Humans).[1][2] Doing things that are virtuous tends to lead to good outcomes. Doing things that aren't virtuous tends to lead to bad outcomes. For you, and for others. It's hard to predict what those outcomes -- good and bad -- actually are. If you were a perfect Bayesian with unlimited information, time and computing power, then yes, go ahead and do the consequentialist calculus. But for humans, we are lacking in those things. Enough so that consequentalist calculus frequently becomes challenging, and the good track record of virtue becomes a huge consideration. So, I agree with you that "lying leads to mistrust" is one of the reasons why vegan advocates shouldn't lie. But I think that the main reason they shouldn't lie is simply that lying has a pretty bad track record. And then another huge consideration is that people who come up with reasons why they, at least in this particular circumstance, are a special snowflake and are justified in lying, frequently are deluding themselves.[3] 1. ^ Well, that post is about ethics. And I think the conversation we're having isn't really limited to ethics. It's more about just, pragmatically, what should the EA community do if they want to win. 2. ^ Here's my slightly different(?) take, if anyone's interested: Reflective Consequentialism.  3. ^ I cringe at how applause light-y this comment is. Please don't upvote if you feel like you might be non-trivially reacting to an applause light.

Separate from the moral issue, this is the kind of trick you can only pull once. I assume that almost everyone who received the "your selected response is currently in the minority" message believed it, that will not be the case next year.

2Raemon
Yup, seems correct.

Granting for the sake of argument that launching the missiles might not have triggered full-scale nuclear war, or that one might wish to define "destroy the world" in a way that is not met by most full-scale nuclear wars, I am still dissatisfied with virtue A because I think an important part of Petrov's situation was that whatever you think the button did, it's really hard to find an upside to pushing it, whereas virtue A has been broadened to cover situations that are merely net bad, but where one could imagine arguments for pushing the button. My initial post framing it in terms of certainty may have been poorly phrased.

4Ruby
There is an upside to being the kind of person who will press the button in retaliation. You hope never to, but the fact that you credibly would allows for MAD game theory to apply. (FDT, etc. etc.)

Petrov was not the last link in the chain of launch authorization which means that his action wasn't guaranteed to destroy the world since someone further down the chain might have cast the same veto he did. So technically yes, Petrov was pushing a button labeled "destroy the world if my superior also thinks these missiles are real, otherwise do nothing". For this reason I think Vasily Arkhipov day would be better, but too late to change now. 

But I think that if the missiles had been launched, that destroys the world (which I use as shorthand for dest... (read more)

3Raemon
(I edited in more reply you may want to respond to. I think the button wasn't actually designed to "destroy world", it was designed to launch a counterattack. Petrov did seem to think it would based on some other quotes of his, but, like, AFAICT he was wrong. I think this is also true for Arkipov)

It seems quite easy to me. Imagine me stating "The sky is purple, if you come to the party I'll introduce you to Alice." If you come to the party then me performing the promised introduction honours a commitment I made, even though I also lied to you.

1Ben Smith
Seems to me that in this case, the two are connected. If I falsely believed my group was in the minority, I might refrain from clicking the button out of a sense of fairness or deference to the majority group.  Consequently, the lie not only influenced people who clicked the button, it perhaps also influenced people who did not. So due to the false premise on which the second survey was based, it should be disregarded altogether. To not disregard would be to have obtained by fraud or trickery a result that is disadvantageous to all the majority group members who chose not to click, falsely believing their view was a minority. I think, morally speaking, avoiding disadvantaging participants through fraud is more important than honoring your word to their competitors. The key difference between this and the example is that there's a connection between the lie and the promise.
3JBlack
A closer analogy is "You are an interesting person, and I will introduce the first interesting person who comes to the party to Alice". You come to the party, you're told that you're the first there, but you're not introduced to Alice because you're not an interesting person after all. Instead they introduce the first interesting person to Alice (who for some reason only has time to meet one person). Ah never mind, I now see what you meant. Yes in general you can narrowly honour your commitment by carrying out the action, but I mean more by "honouring your word" than just that. As I see it, someone who deliberately lies has not honoured their word, regardless of any subsequent actions that they might perform. They've made two statements, one vouching that something is true, and one vouching that something will be true. Ensuring that the latter will be true does nothing to restore their loss of honour from the deliberate falsity of the former. In this case they can't even honour the latter part, since they made a mutually exclusive promise to two different people.

This is not responding to the interesting part of the post, but I did not vote in the poll because I felt like virtue A was a mangled form of the thing I care about for Petrov Day, and non-voting was the closest I could come to fouling my ballot in protest.

To me Petrov Day is about having a button labeled "destroy world" and choosing not to press it. Virtue A as described in the poll is about having a button labeled "maybe destroy world, I dunno, are you feeling lucky?" and choosing not to press it. This is a different definition which seems to have been e... (read more)

5Raemon
It feels fairly important to me that in real life, Petrov had a "maybe destroy world, are you feeling lucky?" button. (It sounds like we disagree on this?) Like, relaying information up the chain of command a) doesn't automatically mean that they launch a full scale counterattack, b) that doesn't mean the US automatically launches a full scale counterattack, c) my current belief is that full scale nuclear war probably cripples the northern hemisphere but doesn't literally kill all humans (which is what I think most people mean by 'destroy the world')

This explains why the honour system doesn't do as much as one might hope, but it doesn't address the initial question of why use explicitly optional vaccination instead of mandatory + honour system. If excluding the unvaccinated is desirable then surely it remains desirable (if subtoptimal) to exclude only those who are both unvaccinated and honest.
 

Scott Adams predicted Trump would win in a landslide. He wasn't just overconfident, he was wrong! The fact that he's not taking a status hit is because people keep reporting his prediciton incompletely and no one bothers to confirm what he actually predicted (when I Google 'Scott Adams Trump prediciton' in Incognito, the first two results say "landslide" in the first ten seconds and title, respectively).

Your first case is an example of something much worse than not updating fast enough.

5philh
Thanks for the correction! Bad example on my part then. My guess is that the point is clear and fairly undisputed, and coming up with an actually correct example wouldn't be very helpful. Still a little embarrassing.

If someone updated towards the "autism is extreme maleness" theory after reading an abstract based on your hypothetical maleness test, you could probably argue them out of that belief by explaining the specific methodology of the test, because it's obviously dumb. If you instead had to do a bunch of math to show why it was flawed, then it would be much harder to convince people because some wouldn't be interested in reading a bunch of math, some wouldn't be able to follow it, and some would have complicated technical nitpicks about how if you run these num... (read more)

1tailcalled
Maybe it would help if the explanation also had a simplified story and then an in-depth description of how one arrived at the simplified story? Like the simplified story for how the EQ is wrong is "The EQ conflates two different things, 'not caring about people' and 'not knowing how to interact with people'. The former is male while the latter is autistic." I don't know for sure what the issue with the SQ is, but I suspect it's going to be something like "The SQ conflates five different things, 'being interested in technology', 'being interested in politics', 'being interested in nature', 'orderliness' and 'artistic creativity'. The former two are male while ?some unknown subset? are autistic." The noteworthy bit is that one can detect these sorts of conflations from the statistics of the scales.

One can cross-reference the moderation log with "Deleted by alyssavance, Today at 8:19 AM" to determine who made any particular deleted comment. Since this information is already public, does it make sense to preserve the information directly on the comment, something like "[comment by Czynski deleted]"?

Fearing that this would be adequate with a large influx of low-quality users

Clarifying: this is a typo and should be inadequate, right?

2Ruby
Oh, you're right. Thanks!

It seems unlikely that AI labs are going to comply with this petition. Supposing that this is the case, does this petition help, hurt, or have no impact on AI safety, compared to the counterfactual where it doesn't exist?

All possibilities seem plausible to me. Maybe it's ignored so it just doesn't matter. Maybe it burns political capital or establishes a norm of "everyone ignores those silly AI safety people and nothing bad happens". Maybe it raises awareness and does important things for building the AI safety coalition.

Modeling social reality is always hard, but has there been much analysis of what messaging one ought to use here, separate from the question of what policies one ought to want?

Not if the people paying in sex are poor! Imagine that 10% of housing is reserved for the poorest people in society as part of some government program that houses them for free, and the other 90% is rented for money at a rate of £500/month (also this is a toy model where all housing is the same, no mansions here). One day the government ends the housing program and privatizes the units, they all go to landlords who start charging money. Is the new rate for housing lower, higher or the same?

The old £500/month rate was the equilibrium that fell out of matchi... (read more)

2Brendan Long
Ah that makes sense. It still feels like the mechanism here is that you reduce house prices by making people poor though (you're reducing the buying power of these people by $500 and relying on side effects to also reduce house prices).

Good point. I feel like it shouldn't happen much but I agree the simple economic model predicts it should. I could resolve it within the model as some kind of market friction argument (finding someone to sell sex to is not trivial, the landlord makes it easier to go into prostitution by providing himself as a "steady employer"), but I think my real intuition is that this is a place where homo economicus breaks down so I shouldn't be trying to apply simple economic models.

Also, even if my initial argument does work, this is basically a novel form of rent co... (read more)

2tailcalled
If the simple economic models break down, it might be worthwhile to think about what other models apply. And also to think about how we know that the simple models break down.
Load More