All of Nick_Tarleton's Comments + Replies

I don't feel a different term is needed/important, but n=1, due to some uses I've seen of 'lens' as a technical metaphor it strongly makes me think 'different mechanically-generated view of the same data/artifact', not 'different artifact that's (supposed to be) about the same subject matter', so I find the usage here a bit disorienting at first.

The Y-axis seemed to me like roughly 'populist'.

The impressive performance we have obtained is because supervised (in this case technically "self-supervised") learning is much easier than e.g. reinforcement learning and other paradigms that naturally learn planning policies. We do not actually know how to overcome this barrier.

What about current reasoning models trained using RL? (Do you think something like, we don't know, and won't easily figure out, how to make that work well outside a narrow class of tasks that doesn't include 'anything important'?)

1Cole Wyeth
Yes, that is what I think.  Edit: The class of tasks doesn't include autonomously doing important things such as making discoveries. It does include becoming a better coding assistant. 

Few people who take radical veganism and left-anarchism seriously either ever kill anyone, or are as weird as the Zizians, so that can't be the primary explanation. Unless you set a bar for 'take seriously' that almost only they pass, but then, it seems relevant that (a) their actions have been grossly imprudent and predictably ineffective by any normal standard + (b) the charitable[1] explanations I've seen offered for why they'd do imprudent and ineffective things all involve their esoteric beliefs.

I do think 'they take [uncommon, but not esoteric, ... (read more)

Violence by radical vegans and left-anarchists has historically not been extremely rare. Nothing in Zizians' actions strike me as particularly different (in kind if not in competency) than, say, the Belle Époque illegalists like the Bonnot Gang, or the Years of Lead leftist groups like the Red Army Fraction or the Weather Underground.

I don't think it's an outright meaningless comparison, but I think it's bad enough that it feels misleading or net-negative-for-discourse to describe it the way your comment did. Not sure how to unpack that feeling further.

4Nathan Helm-Burger
Well, I upvoted your comment, which I think adds important nuance. I will also edit my shortform to explicitly say to check your comment. Hopefully, the combination of the two is not too misleading. Please add more thoughts as they occur to you about how better to frame this.

https://artificialanalysis.ai/leaderboards/providers claims that Cerebras achieves that OOM performance, for a single prompt, for 70B-parameter models. So nothing as smart as R1 is currently that fast, but some smart things come close.

2Nathan Helm-Burger
Yeah, I just found a cerebras post which claims 2100 serial tokens/sec.

I don't see how it's possible to make a useful comparison this way; human and LLM ability profiles, and just the nature of what they're doing, are too different. An LLM can one-shot tasks that a human would need non-typing time to think about, so in that sense this underestimates the difference, but on a task that's easy for a human but the LLM can only do with a long chain of thought, it overestimates the difference.

Put differently: the things that LLMs can do with one shot and no CoT imply that they can do a whole lot of cognitive work in a single forwar... (read more)

2Nathan Helm-Burger
Yeah, of course. Just trying to get some kind of rough idea at what point future systems will be starting from.

I don't really have an empirical basis for this, but: If you trained something otherwise comparable to, if not current, then near-future reasoning models without any mention of angular momentum, and gave it a context with several different problems to which angular momentum was applicable, I'd be surprised if it couldn't notice that  was a common interesting quantity, and then, in an extension of that context, correctly answer questions about it. If you gave it successive problem sets where the sum of that quantity was applicable, the integr... (read more)

3TsviBT
It's a good question. Looking back at my example, now I'm just like "this is a very underspecified/confused example". This deserves a better discussion, but IDK if I want to do that right now. In short the answer to your question is * I at least would not be very surprised if gippity-seek-o5-noAngular could do what I think you're describing. * That's not really what I had in mind, but I had in mind something less clear than I thought. The spirit is about "can the AI come up with novel concepts", but the issue here is that "novel concepts" are big things, and their material and functioning and history are big and smeared out. I started writing out a bunch of thoughts, but they felt quite inadequate because I knew nothing about the history of the concept of angular momentum; so I googled around a tiny little bit. The situation seems quite awkward for the angular momentum lesion experiment. What did I "mean to mean" by "scrubbed all mention of stuff related to angular momentum"--presumably this would have to include deleting all subsequent ideas that use angular moment in their definitions, but e.g. did I also mean to delete the notion of cross product? It seems like angular momentum was worked on in great detail well before the cross product was developed at all explicitly. See https://arxiv.org/pdf/1511.07748 and https://en.wikipedia.org/wiki/Cross_product#History. Should I still expect gippity-seek-o5-noAngular to notice the idea if it doesn't have the cross product available? Even if not, what does and doesn't this imply about this decade's AI's ability to come up with novel concepts? (I'm going to mull on why I would have even said my previous comment above, given that on reflection I believe that "most" concepts are big and multifarious and smeared out in intellectual history. For some more examples of smearedness, see the subsection here: https://tsvibt.blogspot.com/2023/03/explicitness.html#the-axiom-of-choice)

It seems right to me that "fixed, partial concepts with fixed, partial understanding" that are "mostly 'in the data'" likely block LLMs from being AGI in the sense of this post. (I'm somewhat confused / surprised that people don't talk about this more — I don't know whether to interpret that as not noticing it, or having a different ontology, or noticing it but disagreeing that it's a blocker, or thinking that it'll be easy to overcome, or what. I'm curious if you have a sense from talking to people.)

These also seem right

  • "LLMs have a weird, non-human shape
... (read more)
7TsviBT
More recently I've mostly disengaged (except for making kinda-shrill LW comments). Some people say that "concepts" aren't a thing, or similar. E.g. by recentering on performable tasks, by pointing to benchmarks going up and saying that the coarser category of "all benchmarks" or similar is good enough for predictions. (See e.g. Kokotajlo's comment here https://www.lesswrong.com/posts/oC4wv4nTrs2yrP5hz/what-are-the-strongest-arguments-for-very-short-timelines?commentId=QxD5DbH6fab9dpSrg, though his actual position is of course more complex and nuanced.) Some people say that the training process is already concept-gain-complete. Some people say that future research, such as "curiosity" in RL, will solve it. Some people say that the "convex hull" of existing concepts is already enough to set off FURSI (fast unbounded recursive self-improvement). True; I think I've heard some various people discussing how to more precisely think of the class of LLM capabilities, but maybe there should be more. It's often awkward discussing these things, because there's sort of a "seeing double" that happens. In this case, the "double" is: "AI can't FURSI because it has poor sample efficiency... 1. ...and therefore it would take k orders of magnitude more data / compute than a human to do AI research." 2. ...and therefore more generally we've not actually gotten that much evidence that the AI has the algorithms which would have caused both good sample efficiency and also the ability to create novel insights / skills / etc." The same goes mutatis mutandis for "can make novel concepts". I'm more saying 2. rather than 1. (Of course, this would be a very silly thing for me to say if we observed the gippities creating lots of genuine novel useful insights, but with low sample complexity (whatever that should mean here). But I would legit be very surprised if we soon saw a thing that had been trained on 1000x less human data, and performs at modern levels on language tasks (allowing it

To be more object-level than Tsvi:

o1/o3/R1/R1-Zero seem to me like evidence that "scaling reasoning models in a self-play-ish regime" can reach superhuman performance on some class of tasks, with properties like {short horizons, cheap objective verifiability, at most shallow conceptual innovation needed} or maybe some subset thereof. This is important! But, for reasons similar to this part of Tsvi's post, it's a lot less apparent to me that it can get to superintelligence at all science and engineering tasks.

Also the claim that Ziz "did the math" with relation to making decisions using FDT-ish theories

IMO Eliezer correctly identifies a crucial thing Ziz got wrong about decision theory:

... the misinterpretation "No matter what, I must act as if everyone in the world will perfectly predict me, even though they won't." ...

also:

i think "actually most of your situations do not have that much subjunctive dependence" is pretty compelling personally

it's not so much that most of the espoused decision theory is fundamentally incorrect but rather that subjunc

... (read more)
7A_donor
Oh... huh. @Eliezer Yudkowsky, I think I figured it out. In a certain class of altered state,[1] a person's awareness includes a wider part of their predictive world-model than usual. Rather than perceiving primarily the part of the self model which models themselves looking out into a world model, the normal gating mechanisms come apart and they perceive much more of their world-model directly (including being able to introspect on their brain's copy of other people more vividly). This world model includes other agents. Those models of other agents in their world model are now existing in an much less sandboxed environment. It viscerally feels like there is extremely strong entanglement between their actions and those of the agents that might be modelling them, because their model of the other agents is able to read their self-model and vice versa, and in that state they're kinda running it right on the bare-metal models themselves. Additionally, people's models of other people generally use themselves as a template. If they're thinking a lot about threats and blackmail and similar, it's easy for that to leak into expecting others are modelling this more than they are. So their systems strongly predict that there is way more subjunctive dependence than is real, due to how the brain handles those kind of emergencies.[2] Add in the thing where decision theory has counterintuitive suggestions and tries to operate kinda below the normal layer of decision process, plus people not being intuitively familiar with it, and yea, I can see why some people can get to weird places. Not reasonably predictable in advance, it's a weird pitfall, but in retrospect fits. Maybe it's a good idea to write an explainer for this to try and mitigate this way people seem to be able to implode. I might talk to some people. 1. ^ The schizophrenia/psychosis/psychedelics-like cluster, often caused by being in extreme psychological states like those caused by cults and extreme pe

Let's look at preference for eating lots of sweets, for example. Society tries to teach us not to eat too much sweets because it's unhealthy, and from the perspective of someone who likes eating sweets, this often feels coercive. Your explanation applied here would be that upon reflection, people will decide "Actually, eating a bunch of candy every day is great" -- and no doubt, to a degree that is true, at least with the level of reflection that people actually do.

However when I decided to eat as much sweet as I wanted, I ended up deciding that sweets

... (read more)
2jimmy
The part of OP you quoted only covers part of what I'm saying. It's not just that we can be pressured into doing good things, it's also that we have no idea what our intrinsic desires will become as we learn more about they interact with each other and the world, and there is a lot of legitimate change in intrinsic preferences which are more reflectively stable upon sufficiently good reflection, but which nevertheless revert to the shallower preferences upon typical reflection because reflection is hard and people are bad at it. "Reflectively stable in absence of coercive pressure" is very difficult to actually measure, so it's more of a hypothetical construct which is easy to get wrong -- especially since "absence of coercive pressure" can't actually exist, so we have to figure out which kinds of coercive pressure we're going to include in our hypothetical.
2Benquo
"As calculated prior" is not quite correct, "reflectively stable absent coercive pressure" is a better formulation.

I can easily imagine an argument that: SBF would be safe to release in 25 years, or for that matter tomorrow, not because he'd be decent and law-abiding, but because no one would trust him and the only crimes he's likely to (or did) commit depend on people trusting him. I'm sure this isn't entirely true, but it does seem like being world-infamous would have to mitigate his danger quite a bit.

More generally — and bringing it back closer to the OP — I feel interested in when, and to what extent, future harms by criminals or norm-breakers can be prevented just by making sure that everyone knows their track record and can decide not to trust them.

4Ben Pace
I think having an easily-findable reputation makes it harder to do crimes, but being famous makes it easier. Many people are naive & gullible, or are themselves willing to do crime, and would like to work with him. I expect him to get opportunities for new ventures on leaving prison, with unsavory sorts. I definitely support track-records being more findable publicly. Of course there's some balance in that the person who publishes it has a lot of power over the person being written about, and if they exaggerate it or write it hyperbolically then they can impose a lot of inappropriate costs on the person that they're in a bad position to push back on.

Though — I haven't read all of his recent novels, but I think — none of those are (for lack of a better word) transhumanist like Permutation City or Diaspora, or even Schild's Ladder or Incandescence. Concretely: no uploads, no immortality, no artificial minds, no interstellar civilization. I feel like this fits the pattern, even though the wildness of the physics doesn't. (And each of those four earlier novels seems successively less about the implications of uploading/immortality/etc.)

In practice, it just requires hardware with limited functionality and physical security — hardware security modules exist.

An HSM-analogue for ML would be a piece of hardware that can have model weights loaded into its nonvolatile memory, can perform inference, but doesn't provide a way to get the weights out. (If it's secure enough against physical attack, it could also be used to run closed models on a user's premises, etc.; there might be a market for that.)

4cqb
Indeed! I was very close to writing a whole bit about TEEs, enclaves, and PUFs in my last comment, but I figured that it also boils down to "just don't give it permission" so I left it out. I actually think designing secure hardware is incredibly interesting and there will probably be an increase in demand for secure computing environments and data provenance in the near future.

This doesn't work. (Recording is Linux Firefox; same thing happens in Android Chrome.)

An error is logged when I click a second time (and not when I click on a different probability):

[GraphQL error]: Message: null value in column "prediction" of relation "ElicitQuestionPredictions" violates not-null constraint, Location: line 2, col 3, Path: MakeElicitPrediction instrument.ts:129:35
7jimrandomh
Sorry about that, a fix is in progress. Unmaking a prediction will no longer crash. The UI will incorrectly display the cancelled prediction in the leftmost bucket; that will be fixed in a few minutes without you needing to re-do any predictions.

How can I remove an estimate I created with an accidental click? (Said accidental click is easy to make on mobile, especially because the way reactions work there has habituated me to tapping to reveal hidden information and not expecting doing so to perform an action.)

1Ben Pace
You click the estimate a second time.

If specifically with IQ, feel free to replace the word with "abstract units of machine intelligence" wherever appropriate.

By calling it "IQ", you were (EDIT: the creator of that table was) saying that gpt4o is comparable to a 115 IQ human, etc. If you don't intend that claim, if that replacement would preserve your meaning, you shouldn't have called it IQ. (IMO that claim doesn't make sense — LLMs don't have human-like ability profiles.)

2Logan Zoellner
  gpt4o is not literally equivalent to a 115 IQ human.   Use whatever word you want for the concept "score produced when an LLM takes an IQ test".  

Learning on-the-fly remains, but I expect some combination of sim2real and muZero to work here.

Hmm? sim2real AFAICT is an approach to generating synthetic data, not to learning. MuZero is a system that can learn to play a bunch of games, with an architecture very unlike LLMs. This sentence doesn't typecheck for me; what way of combining these concepts with LLMs are you imagining?

4Logan Zoellner
Imagine you were trying to build a robot that could: 1. Solve a complex mechanical puzzle it has never seen before 2. Play at an expert level a board game that I invented just now. Both of these are examples of learning-on-the-fly.  No amount of pre-training will ever produce a satisfying result. The way I believe a human (or a cat) solves 1. is they: look at the puzzle, try some things, build a model of the toy in their head, try things on the model in their head, eventually solve the puzzle.  There are efforts to get robots to follow the same process, but nothing I would consider "this is the obvious correct solution" quite yet. The way to solve 2. (I think) is simply to have the LLM translate the rules of the game into a formal description and then run muZero on that. Ideally there is some unified system that takes out the "translate into another domain and do your training there" step (which feels very anti-bitter-lesson).  But I confess I haven't the slightest idea how to build such a system.

I don't think it much affects the point you're making, but the way this is phrased conflates 'valuing doing X oneself' and 'valuing that X exist'.

Among 'hidden actions OpenAI could have taken that could (help) explain his death', I'd put harassment well above murder.

Of course, the LessWrong community will shrug it off as a mere coincidence because computing the implications is just beyond the comfort level of everyone on this forum.

Please don't do this.

1daijin
I made a v2 of this shortform that answers your point with an example from recent history.

I've gotten things from Michael's writing on Twitter, but also wasn't distinguishing him/Ben/Jessica when I wrote that comment.

2habryka
Makes sense. My agree-react and my sense of Niplav's comments were specifically about Michael's writing/podcasts.

I can attest to something kind of like this; in mid-late 2020, I

  • already knew Michael (but had been out of touch with him for a while) and was interested in his ideas (but hadn't seriously thought about them in a while)
  • started doing some weird intense introspection (no drugs involved) that led to noticing some deeply surprising things & entering novel sometimes-disruptive mental states
  • noticed that Michael/Ben/Jessica were talking about some of the same things I was picking up on, and started reading & thinking a lot more about their online writi
... (read more)

I have understood and become convinced of some of Michael's/Ben's/Jessica's stances through a combination of reading their writing and semi-independently thinking along similar lines, during a long period of time when I wasn't interacting with any of them, though I have interacted with all of them before and since.

2niplav
Thank you, that's useful evidence!
2habryka
(I think Jessica and Ben have both been great writers and I have learned a lot from both of them. I have also learned a bunch of things from Michael, but definitely not via his writing or podcasts or anything that wasn't in-person, or second-hand in-person. If you did learn something from the Michael podcasts or occasional piece of writing he has done, like the ones linked above, that would be a surprise to me)

... those posts are saying much more specific things than 'people are sometimes hypocritical'?

"Can crimes be discussed literally?":

  • some kinds of hypocrisy (the law and medicine examples) are normalized
  • these hypocrisies are / the fact of their normalization is antimemetic (OK, I'm to some extent interpolating this one based on familiarity with Ben's ideas, but I do think it's both implied by the post, and relevant to why someone might think the post is interesting/important)
  • the usage of words like 'crime' and 'lie' departs from their denotation, to excl
... (read more)
2Unnamed
Does it bother you that this is not what's happening in many of the examples in the post? e.g., With "the American hospital system is built on lies."
-3Viliam
After sleeping on it, it seems to me that the topic were are talking about is "staring into the abyss": whether, when, and how to do it properly, and for what outcome. The easiest way is to not do it at all. Just pretend that everything is flowers and rainbows, and refuse to talk about the darker aspects of reality. This is what we typically do with little children. A part of that is parental laziness: by avoiding difficult topics we avoid difficult conversations. But another part is that children are not cognitively ready to process nontrivial topics, so we try to postpone the debates about darker things until later, when they get the capability. Some lazy parents overdo it; some kids grow up living in a fairy tale world. Occasional glimpses of darkness can be dismissed as temporary exceptions to the general okay-ness of the world. "Grandma died, but now she is happy in Heaven." At this level, people who try to disrupt the peace are dismissed relatively gently, accused of spoiling the mood and frightening the kids. When this becomes impossible because the darkness pushes its way beyond our filters, the next lazy strategy is to downplay the darkness. Either it is not so bad, or there is some silver lining to everything. "Death gives meaning to life." "The animals don't mind dying so that we can have meat to eat; they understand it is their role in the system." "Slavery actually benefits the blacks; they do not have the mental capacity to survive without a master." "What doesn't kill you, makes you stronger." At this point the pushback against those trying to disrupt the peace is stronger; people are aware that their rationalizations are fragile. Luckily, we can reframe the rationalizations as a sign of maturity, and dismiss those who disagree with us as immature. "When you grow up, you will realize that..." Another possible reaction is trying to join the abyss. Yes, bad things happen, but since they are inevitable, there is no point worrying about that. Heck, if
4Viliam
Thank you for the summary. I guess my bubble is more cynical than the average population, so I may underestimate how shocking similar thoughts would be for them. I might also add the replication crisis in science, etc. Yes. The world is bad. Almost everything is broken. People don't want to admit it. Those who understand how things work are often ashamed for their role in the system. Some respond by attacking those who point it out, or even those who merely refuse to participate in the same way. ...it still feels like I am waiting for the other shoe to drop. All these things, they make me feel sad. I feel bad about all the wasted opportunity to live better. I don't model most people as evil, just... overwhelmed by all the things that are wrong, and their inability to do something against it. Well, many are too stupid to care. Some people are genuinely evil. Many are going along with the flow, wherever it takes them. Most of human behavior is probably determined by habit; if you grow up in a dysfunctional environment, it will become your normal, but if your twin grew up in a better environment, they would recognize it as better. Even now, there are people who name the unpleasant truths. People who spend a large part of their life fighting against some specific dysfunction. But the world is complicated, and problems too numerous. So, what is that extra insight that I could gain by contemplating these things while taking drugs and listening to Vassar whispering dark thoughts into my ears? Me too, but nothing specific. Maybe it's like when you are high and you believe that you have amazing insights, but when you write them down and read them again when you are sober, there is nothing.

Embarrassingly, that was a semi-unintended reaction — I would bet a small amount against that statement if someone gave me a resolution method, but am not motivated to figure one out, and realized this a second after making it — that I hadn't figured out how to remove by the time you made that comment. Sorry.

It sounds to me like the model is 'the candidate needs to have a (party-aligned) big blind spot in order to be acceptable to the extremists(/base)'. (Which is what you'd expect, if those voters are bucketing 'not-seeing A' with 'seeing B'.)

(Riffing off from that: I expect there's also something like, Motive Ambiguity-style, 'the candidate needs to have some, familiar/legible(?), big blind spot, in order to be acceptable/non-triggering to people who are used to the dialectical conflict'.)

1deepthoughtlife
It seems I was not clear enough, but this is not my model. (I explain it to the person who asked if you want to see what I meant, but I was talking about parties turning their opponents into scissors statements.) That said, I do believe that it is a possible partial explanation that sometimes having an intentional blind spot can be seen as a sign of loyalty by the party structure.

if my well-meaning children successfully implement my desire never to die, by being uploaded, and "turn me on" like this with sufficient data and power backups but lack of care; or if something else goes wrong with the technicians involved not bothering to check if the upload was successful in setting up a fully virtualized existence complete with at least emulated body sensations, or do not otherwise check from time to time to ensure this remains the case;

These don't seem like plausible scenarios to me. Why would someone go to the trouble of running an upload, but be this careless? Why would someone running an upload not try to communicate with it at all?

2Richard_Kennaway
I don't want to aggravate the OP's problem, but a reason that immediately occurs to me is that the people running the upload have a statutory duty to do so but don't actually care. Consider the treatment of the elderly incapable in some care homes.

A shell in a Matrioshka brain (more generally, a Dyson sphere being used for computation) reradiates 100% of the energy it captures, just at a lower temperature.

2Charlie Steiner
Yeah, the energy radiated to infinity only gets reduced if it's being used for something long-term, like disassembling the sun or sending off energy-intensive intergalactic probes.

The AI industry people aren't talking much about solar or wind, and they would be if they thought it was more cost effective.

I don't see them talking about natural gas either, but nuclear or even fusion, which seems like an indication that whatever's driving their choice of what to talk about, it isn't short-term cost-effectiveness.

I doubt it (or at least, doubt that power plants will be a bottleneck as soon as this analysis says). Power generation/use varies widely over the course of a day and of a year (seasons), so the 500 GW number is an average, and generating capacity is overbuilt; this graph on the same EIA page shows generation capacity > 1000 GW and non-stagnant (not counting renewables, it declined slightly from 2005 to 2022 but is still > 800 GW):

This seems to indicate that a lot of additional demand[1] could be handled without building new generation, at least ... (read more)

Confidentiality: Any information you provide will not be personally linked back to you. Any personally identifying information will be removed and not published. By participating in this study, you are agreeing to have your anonymized responses and data used for research purposes, as well as potentially used in writeups and/or publications.

Will the names (or other identifying information if it exists, I haven't taken the survey) of the groups evaluated potentially be published? I'm interested in this survey, but only willing to take it if there's a confide... (read more)

[This comment is no longer endorsed by its author]Reply

The hypothetical bunker people could easily perform the Cavendish experiment to test Newtonian gravity, there just (apparently) isn't any way they'd arrive at the hypothesis.

1VojtaKovarik
Good point. Also, for the purpose of the analogy with AI X-risk, I think we should be willing to grant that the people arrive at the alternative hypothesis through theorising. (Similarly to how we came up with the notion of AI X-risk before having any powerful AIs.) So that does break my example somewhat. (Although in that particular scenario, I imagine that sceptic of Newtonian gravity would came up with alternative explanations for the observation. Not that this seems very relevant.)

As a counterpoint, I use Firefox as my primary browser (I prefer a bunch of little things about its UI), and this is a complete list of glitches I've noticed:

  • The Microsoft account login flow sometimes goes into a loop of asking me for my password
  • Microsoft Teams refuses to work ('you must use Edge or Chrome')
  • Google Meet didn't used to support background blurring, but does now
  • A coworker reported that a certain server BMC web interface didn't work in Firefox, but did in Chrome (on Mac) — I found (on Linux, idk if that was the relevant difference) it broke the same way in both, which I could get around by deleting a modal overlay in the inspector

(I am not a lawyer)

The usual argument (e.g.) for warrant canaries being meaningful is that the (US) government has much less legal ability to compel speech (especially false speech) than to prohibit it. I don't think any similar argument holds for private contracts; AFAIK they can require speech, and I don't know whether anything is different if the required speech is known by both parties to be false. (The one relevant search result I found doesn't say there's anything preventing such a contract; Claude says there isn't, but it could be thrown out on grou... (read more)

9Adam Scholl
Yeah, the proposal here differs from warrant canaries in that it doesn't ask people to proactively make statements ahead of time—it just relies on the ability of some people who can speak, to provide evidence that others can't. So if e.g. Bob and Joe have been released, but Alice hasn't, then Bob and Joe saying they've been released makes Alice's silence more conspicuous.

Upvoted, but weighing in the other direction: Average Joe also updates on things he shouldn't, like marketing. I expect the doctor to have moved forward some in resistance to BS (though in practice, not as much as he would if he were consistently applying his education).

And the correct reaction (and the study's own conclusion) is that the sample is too small to say much of anything.

(Also, the "something else" was "conventional treatment", not another antiviral.)

1Jay Molstad
Well, we can say that 27/30 (90%) patients improved. With a very high level of confidence, we can say that this disease is less fatal than Ebola (which would have killed 26 or so).

I find the 'backfired through distrust'/'damaged their own credibility' claim plausible, it agrees with my prejudices, and I think I see evidence of similar things happening elsewhere; but the article doesn't contain evidence that it happened in this case, and even though it's a priori likely and worth pointing out, the claim that it did happen should come with evidence. (This is a nitpick, but I think it's an important nitpick in the spirit of sharing likelihood ratios, not posterior beliefs.)

Ben Pace*190

Yeah. I regularly model headlines like this as being part of the later levels of simulacra. The article argued that it should backfire, but it also said that it already had. If the article catches on, then it will become true to the majority of people who read it. It's trying to create the news that it's reporting on. It's trying to make something true by saying it is.

I think a lot of articles are like that these days. They're trying to report on what's part of social reality, but social reality depends on what goes viral on twitter/fb/etc, so they work to

... (read more)
4Raemon
I'd say this isn't just a nitpicking, it's pretty directly challenging the core claim. Or at least, if the essay didn't want to be making that it's core claim, it should have picked a different title. (I say that while generally endorsing the article)

if there's a domain where the model gives two incompatible predictions, then as soon as that's noticed it has to be rectified in some way.

What do you mean by "rectified", and are you sure you mean "rectified" rather than, say, "flagged for attention"? (A bounded approximate Bayesian approaches consistency by trying to be accurate, but doesn't try to be consistent. I believe 'immediately update your model somehow when you notice an inconsistency' is a bad policy for a human [and part of a weak-man version of rationalism that harms people who try to follo

... (read more)
9orthonormal
The next paragraph applies there: you can rectify it by saying it's a conflict between hypotheses / heuristics, even if you can't get solid evidence on which is more likely to be correct. Cases where you notice an inconsistency are often juicy opportunities to become more accurate.

On the other hand:

We found that viable virus could be detected... up to 4 hours on copper...

1Spiracular
Extract from that paper HCOV-19 is a (unusual?) name for the nCOV/SARS-CoV-2 virus responsible for COVID19. It's the one in red.
Answer by Nick_Tarleton190

Here's a study using a different coronavirus.

Brasses containing at least 70% copper were very effective at inactivating HuCoV-229E (Fig. 2A), and the rate of inactivation was directly proportional to the percentage of copper. Approximately 103 PFU in a simulated wet-droplet contamination (20 µl per cm2) was inactivated in less than 60 min. Analysis of the early contact time points revealed a lag in inactivation of approximately 10 min followed by very rapid loss of infectivity (Fig. 2B).

5Nick_Tarleton
On the other hand:

That paper only looks at bacteria and does not knowably carry over to viruses.

I don't see you as having come close to establishing, beyond the (I claim weak) argument from the single-word framing, that the actual amount or parts of structure or framing that Dragon Army has inherited from militaries are optimized for attacking the outgroup to a degree that makes worrying justified.

6Benquo
This definitely doesn't establish that. And this seems like a terrible context in which to continue to elaborate on all my criticisms of Duncan's projects, so I'm not going to do that. My main criticisms of Dragon Army are on the Dragon Army thread, albeit worded conservatively in a way that may not make it clear how these things are related to the "army" framing. If you want to discuss that, some other venue seems right at this point, this discussion is already way too broad in scope.
3John_Maxwell
Since Benquo says he thinks sports are good, I'd be curious whether he is also worried about sports teams with names that suggest violence. Many teams are named after parties in a violent historical conflict or violent animals: Patriots, Braves, Panthers, Raptors, Bulls, Sharks, Warriors, Cavaliers, Rangers, Raiders, Blackhawks, Predators, Tigers, Pirates, Timberwolves...

Doesn't work in incognito mode either. There appears to be an issue with lesserwrong.com when accessed over HTTPS — over HTTP it sends back a reasonable-looking 301 redirect, but on port 443 the TCP connection just hangs.

Similar meta: none of the links to lesserwrong.com currently work due to, well, being to lesserwrong.com rather than lesswrong.com.

2Raemon
hmm. I can fix these links, but fyi if you clear your browser cache they should work for you. (If not, lemme know)

Further-semi-aside: "common knowledge that we will coordinate to resist abusers" is actively bad and dangerous to victims if it isn't true. If we won't coordinate to resist abusers, making that fact (/ a model of when we will or won't) common knowledge is doing good in the short run by not creating a false sense of security, and in the long run by allowing the pattern to be deliberately changed.

3clone of saturn
I don't think it's that simple. First, if abusers and victims exist then the situation just is actively dangerous. Hypocrisy is unavoidable but it's less bad if non-abusers can operate openly and abusers need to keep secrets than vice versa. Second, I don't think the pattern can be deliberately changed except by creating a sense of security that starts out false but becomes true once enough people have it.

This post may not have been quite correct Bayesianism (... though I don't think I see any false statements in its body?), but regardless there are one or more steel versions of it that are important to say, including:

  • persistent abuse can harm people in ways that make them more volatile, less careful, more likely to say things that are false in some details, etc.; this needs to be corrected for if you want to reach accurate beliefs about what's happened to someone
  • arguments are soldiers; if there are legitimate reasons (that people are responding to) to a
... (read more)

IMO, the "legitimate influence" part of this comment is important and good enough to be a top-level post.

4Sniffnoy
OK, give me some time and maybe I'll post it, expanded with some related notions that are less relevant to the original context but which I think are worth writing about...

This is simply instrumentally wrong, at least for most people in most environments. Maybe people and an environment could be shaped so that this was a good strategy, but the shaping would actually have to be done and it's not clear what the advantage would be.

Load More