While I certainly have thoughts on all of this, let me point out one aspect of this system which I think is unusually dangerous and detrimental:
The ability (especially for arbitrary users, not just moderators) to take moderation actions that remove content, or prevent certain users from commenting, without leaving a clearly and publicly visible trace.
At the very least (if, say, you’re worried about something like “we don’t want comments sections to be cluttered with ‘post deleted’”), there ought to be a publicly viewable log of all moderation actions. (Consider the lobste.rs moderation log feature as an example of how such a thing might work.) This should apply to removal of comments and threads, and it should definitely also apply to banning a user from commenting on a post / on all of one’s posts.
Let me say again that I consider a moderation log to be the minimally acceptable moderation accountability feature on a site like this—ideally there would also be indicators in-context that a moderation action has taken place. But allowing totally invisible / untraceable moderation actions is a recipe for disaster.
Edit: For another example, note Scott’s register of bans/warnings, which ...
I'm also mystified at why traceless deletition/banning are desirable properties to have on a forum like this. But (with apologies to the moderators) I think consulting the realpolitik will spare us the futile task of litigating these issues on the merits. Consider it instead a fait accompli with the objective to attract a particular writer LW2 wants by catering to his whims.
For whatever reason, Eliezer Yudkowsky wants to have the ability to block commenters and have the ability to do traceless deletion on his own work, and he's been quite clear this is a condition for his participation. Lo and behold precisely these features have been introduced, with suspiciously convenient karma thresholds which allow EY (at his current karma level) to traceless delete/ban on his own promoted posts, yet exclude (as far as I can tell) the great majority of other writers with curated/front page posts from being able to do the same.
Given the popularity of EY's writing (and LW2 wants to include future work of his), the LW2 team are obliged to weigh up the (likely detrimental) addition of these features versus the likely positives of his future posts. Going for the latter is probably the right judgement call to make, but let's not pretend it is a principled one: we are, as the old saw goes, just haggling over the price.
Yeah, I didn't want to make this a thread about discussing Eliezer's opinion, so I didn't put that front and center, but Eliezer only being happy to crosspost things if he has the ability to delete things was definitely a big consideration.
Here is my rough summary of how this plays into my current perspective on things:
1. Allowing users to moderate their own posts and set their own moderation policies on their personal blogs is something I wanted before we even talked to Eliezer about LW2 the first time.
2. Allowing users to moderate their own front-page posts is not something that Eliezer requested (I think he would be happy with them just being personal posts), but is a natural consequence of wanting to allow users to moderate their own posts, while also not giving up our ability to promote the best content to the front-page and to curated
3. Allowing users to delete things without a trace was a request by Eliezer, but is also something I thought about independently to deal with stuff like spam and repeated offenders (for example, Eugine has created over 100 comments on one of Ozy's posts, and you don't want all of them to show up as deleted stubs). I expect we wouldn't have built the future as it currently stands without Eliezer, but I hadn't actually considered a moderation logs page like the one Said pointed out, and I actually quite like that idea, and don't expect Eliezer to object too much to it. So that might be a solution that makes everyone reasonably happy.
I really like the moderation log idea - I think it could be really good for people to have a place where they can go if they want to learn what the norms are empirically. I also propose there be a similar place which stores the comments explaining why posts are curated.
(Also note that Satvik Beri said to me I should do this a few months ago and I forgot and this is my fault.)
I actually quite like the idea of a moderation log, and Ben and Ray also seem to like it. I hadn't really considered that as an option, and my model is that Eliezer and other authors wouldn't object to it either, so this seems like something I would be quite open to implementing.
Yeah, I agree it doesn't create the ideal level of transparency. In my mind, a moderation log is more similar to an accounting solution than an educational solution, where the purpose of accounting is not something that is constantly broadcasted to the whole system, but is instead used to backtrack if something has gone wrong, or if people are suspicious that there is some underlying systematic problem going on. Which might get you a lot of the value that you want, for significantly lower UI-complexity cost.
Dividing the site to smaller sub-tiefs where individual users have ultimate moderation power seems to have been a big part of why Reddit (and to some extent, Facebook) got so successful, so I'm having big hopes for this model.
I don't have an opinion on the moderation policy, but I did want to say thanks for all the hard work in bringing the new site to life.
LessWrong 1.0 was basically dead, and 2.0 is very much alive. Huge respect and well-wishes.
Just want to say this moderation design addresses pretty much my only remaining aversion to posting on LW and I will be playing around with Reign of Terror if I hit the karma. Also really prefer not to leave public traces.
My primary desire to remove the trace is that there are characters so undesirable on the internet that I don't want to be reminded of their existence every time I scroll through my comments section, and I certainly don't want their names to be associated with my content. Thankfully, I have yet to receive any comments anything close to this level on LW, but give a quick browse through the bans section of SlateStarCodex and you'll see they exist.
I am in favor of a trace if it were on a moderation log that does not show up on the comment thread itself.
I think giving people the right and responsibility to unilaterally ban commenters on their posts is demanding too much of people's rationality, forcing them to make evaluations when they're among most likely to be biased, and tempting them with the power to silence their harshest or most effective critics. I personally don't trust myself to do this and have basically committed to not ban anyone or even delete any comments that aren't obvious spam, and kind of don't trust others who would trust themselves to do this.
Banning someone does not generally silence their harshest critics. It just asks those critics to make a top-level post, which generally will actually have any shot at improving the record and discourse in reasonable ways compared to nested comment replies.
The thing that banning does is make it so the author doesn't look like he is ignoring critics (which he hasn't by the time he has consciously decided to ban a critic).
which generally will actually have any shot at improving the record and discourse in reasonable ways compared to nested comment replies
I would believe this iff banned users were nonetheless allowed (by moderator fiat) to type up a comment saying "I have written a response to this post at [insert link]," which actually shows up in the comment section of the original post.
Otherwise, I'd suspect a large part of the readers of the post will not even know there is such a response disagreeing with the original (because they just stumbled upon the post, or they clicked on a link to it from elsewhere on LW, or were just browsing some tag or the "recent activity" tab, etc).
(Not to mention that posts don't even have a sticker at the bottom saying "the author has banned the following users from commenting on their posts: [...]", which should absolutely appear if the point of allowing authors to ban commenters was actually to improve the record and discourse.
You have to know to click on a whole different link (which basically gets advertised precisely nowhere on the front page) to gather that info, and unironically I don't even currently remember what that link is... and I think I'...
The thing that banning does is make it so the author doesn’t look like he is ignoring critics (which he hasn’t by the time he has consciously decided to ban a critic).
… of course banning has this effect, but this is obviously a massively misleading appearance in that case, so why would we want this? You seem to be describing a major downside of allowing authors to ban critics!
Like, suppose that I read a post, see that the author responds to all critical comments, and think: “Cool, this dude doesn’t ignore critics; on the contrary, he replies coherently and effectively to all critical comments; well done! This makes me quite confident that this post in particular doesn’t have any glaring flaws (since someone would point them out, and the author would have no good answer); and, more generally, it makes me think that this author is honest and unafraid to face criticism, and has thought about his ideas, and so is unlikely to be spouting nonsense.”
And then later I find out that actually, this author has banned one or more users from his posts after those users posted seriously critical comments for which the author had no good responses; and I think “… oh.”
Because that initial impres...
It's a localized silencing, which discourages criticism (beyond just the banned critic) and makes remaining criticism harder to find, and yes makes it harder to tell that the author is ignoring critics. If it's not effective at discouraging or hiding criticism, then how can it have any perceived benefits for the author? It's gotta have some kind of substantive effect, right? See also this.
the alternative of having people provide a platform to anyone without any choice in the matter, whose visibility gets multiplied proportional to their own reach and quality, sucks IMO a lot more.
On the contrary, this is an actively good thing, and should be encouraged. “If you write about your ideas on a public discussion forum, you also thereby provide a platform to your critics, in proportion to the combination of your own reach and the critics’ popularity with the forum’s membership[1]” is precisely the correct sort of dynamic to enable optimal truth-seeking.
Think about it from the reader’s perspective: if I read some popular, interesting, apparently-convincing idea, the first thing—the first thing!—that I want to know, after doing a “first-pass” evaluation of the idea myself, is “what do other people think about it”? This is not a matter of majoritarianism, note, but rather:
I think there are few people who have beliefs as considered on this topic as I do! And maybe no one who has as much evidence as I do (which doesn't mean I am right, people come to dumb beliefs while being exposed to lots of evidence all the time).
I've conducted informal surveys, have done hundreds of user interviews, have had conversations about their LessWrong posting experienes with almost every core site contributor over the years, have had conversations with hundreds of people who decided not to post on LW but instead post somewhere else, conversations with people who moved from other platforms to LW, conversations with people who moved from LW to other platforms, and many more.
I have interviewed people in charge of handling these tradeoffs at many of the other big content platforms out there, as well as dozens of people who run smaller forums and online communities. I have poured over analytics and stats and graphs trying to understand what causes people to write here instead of other places, and what causes them to grow as both a commenter and writer.
All of these form a model of how things work that suggests to me that yes, it is true that bad commenters dri...
Why can't they just ignore those they don't want to engage with, maybe with the help of a mute or ignore feature (which can also mark the ignored comments/threads in some way to notify others)?
I get a sense that you (and Said) are really thinking of this as a 1-1 interaction and not a group interaction, but the group dynamics are where most of my crux is.
I feel like all your proposals are like “have a group convo but with one person blanking another person” or “have a 6-person company where one person just ignores another person” and all of my proposals are “have 2 group convos where the ignored isn’t with the person they’re ignoring”, and I feel like the former is always unnatural and never works and the latter is entirely natural and works just fine.
If you ignore a person in a group convo, it’s really weird. Suppose the person makes some additional comment on a topic you’re discussing and now other people continue it. You are still basically interacting with that person’s comment; you couldn’t ignore it, because it directed the flow of conversation. Or instead perhaps you randomly stop engaging with threads and then ppl learn not to reply to that person because you won’t engage, and then they start to get subtly socially excluded in ways they didn’t realize were happening. These are confusing and unnatural and either can make the conversation “not worth it” or “net costly” to one of the participants.
Here's my understanding of the situation. The interested parties are:
Does this sound about right?
[Update: The guidelines above say "Before users can mo...
Said's comment that triggered this debate is 39/34, at the top of the comments section of the post and #6 in Popular Comments for the whole site, but you want to allow the author to ban Said from future commenting, with the rationale "you should model karma as currently approximately irrelevant for managing visibility of comments". I think this is also wrong generally as I've often found karma to be very helpful in exposing high quality comments to me, and keeping lower quality comments less visible toward the bottom, or allowing me to skip them if they occur in the middle of threads.
I almost think the nonsensical nature of this justification is deliberate, but I'm not quite sure. In any case, sigh...
For the record, I had not read that instance of banning, and it is only just at this late point (e.g. after basically the whole thread has wrapped [edit: the whole thread, it turned out, had not wrapped]) did I read that thread and realize that this whole thread was downstream of that. All my comments and points so far were not written with that instance in mind but on general principle.
(And if you're thinking "Surely you would've spoken to Habryka at work about this thread?" my response is "I was not at work! I am currently on vacation." Yes, I have chosen to — and enjoyed! — spending my vacation arguing the basic principles of moderation, criticism, and gardening.)
Initial impressions re: that thread:
Plenty of authors are “willing to engage” with “critics”—as long as the “critics” are the sort that take as an axiom that the author’s work is valuable, important, and interesting, and that the author himself is intelligent, well-intentioned, well-informed, and sane; and as long as their “criticism” is of the sort that says “how very fascinating your ideas are; I would love to learn more about your thinking, but I have not yet grasped your thesis in its fullness, and am confused; might you deign to enlighten me?” (in other words, “here’s a prompt for you to tell us more about your amazing ideas”). (You might call this “intellectual discussion as improv session”—where, as in any improv, the only allowed replies are “yes, and…”.)
It is challenging and unpleasant to be in an interaction with someone who is exuding disgust and contempt for you, and it's not a major weakness in people that they disprefer conversations like that.
A good thing to do in such situations is to do post-level responses rather than comment-level replies. I've seen many post-level back-and-forths where people disrespect the other person's opinions (e.g. Scott Alexander / Robin Hanson on healthcare, Scott Alexander...
It is challenging and unpleasant to be in an interaction with someone who is exuding disgust and contempt for you, and it’s not a major weakness in people that they disprefer conversations like this.
There are two mistakes here, I’d say.
First: no, it absolutely is a “major weakness in people” that they prefer to avoid engaging with relevant criticism merely on the basis of the “tone”, “valence”, etc., of the critics’ words. It is, in fact, a huge weakness. Overcoming this particular bias is one of the single biggest personal advances in epistemic rationality that one can make.
(Actually, I recently read a couple of tweets by @Holly_Elmore, with whom I certainly haven’t always agreed, but who describes this sort of thing very aptly.)
Second: you imply a false dichotomy between the “improv session” sort of faux-criticism I describe, and “exuding disgust and contempt”. Those are not the only options! It is entirely possible to criticize someone’s ideas, very harshly, while exhibiting (and experiencing) no significant emotionally-valenced judgment of the person themselves.
One of the best things that I’ve read recently was “ArsDigita: From Start-Up to Bust-Up” by Philip Greenspun, which...
but it just seems absurd to suggest that paragraphs like this is not equivalent to calling people "stupid" or "evil":
It's... obviously not equivalent to saying people are dumb or evil?
It is equivalent to saying people have soft egos. But that doesn't mean they are dumb or evil. I know plenty of smart and good people who have trouble receiving any meaningful criticism. Heck, I used to (in my opinion) be one of those people when I was younger!
I suspect the proportion of people with soft egos is significantly larger than the proportion of people who are stupid and evil.
FWIW I regularly read a barely-veiled contempt/derision into Said's comments for many people on LessWrong, including in the passage that Habryka quotes. My guess is that we should accept that some people strongly read this and some people do not, and move on with the conversation, rather than insist that there is an 'obvious' reading of intent/emotion.
(To be clear I am willing to take the side of the bet that the majority of people will read contempt/derision for other commenters into Said's comments, including the one you mention. Open to setting up a survey on this if you are feel confident it will not show this.)
"contempt for their weak egos"
Look, man, it's definitely "contempt for them" not just "contempt for their weak egos'".
It's not like Said is walking around distinguishing between people's ego's and the rest of their personality or identity. If someone wanted to communicate "contempt for your weak ego, because of how it prevents you from having good epistemic/contributing meaningfully to a truth-seeking forum" you would use very different words. You would say things like "I have nothing against you as a whole, but I do have something against this weak ego of yours, which I think is holding you back".
In as much as you are just trying to say "contempt for them, because of their weak egos", then sure, whenever someone acts contemptuous they will have some reason. In this case the reason is "I judge your ego to be weak" but that doesn't really change anything.
In my mind things aren't neatly categorized into "top N reasons", but here are some quick thoughts:
(I.) I am generally very averse to having any UI element that shows on individual comments. It just clutters things up quickly and requires people to scan each individual comment. I have put an enormous amount of effort into trying to reduce the number of UI elements on comments. I much prefer organizing things into sections which people can parse once, and then assume everything has the same type signature.
(II.) I think a core thing I want UI to do in the space is to hit the right balance between "making it salient to commenters that they are getting more filtered evidence" and "giving the author social legitimacy to control their own space, combined with checks and balances".
I expect this specific proposal to end up feeling like a constant mark of shame that authors are hesitant to use because they don't feel the legitimacy to use it, and most importantly, make it very hard for them to get feedback on whether others judge them for how they use it, inducing paranoia and anxiety, which I think would make the feature largely unused. I think in that world it isn't really hel...
To reduce clutter you can reuse the green color bars that currently indicate new comments, and make it red for muted comments.
No, the whole point of the green bars is to be a very salient indicator that only shows in the relatively rare circumstance where you need it (which is when you revisit a comment thread you previously read and want to find new comments). Having a permanent red indicator would break in like 5 different ways:
To be clear, I still appreciate the suggestion, but I don't think it's a good one in this context....
As an aside, I think one UI preference I suspect Habryka has more strongly than Wei Dai does here is that the UI look the same to all users. For similar reasons why WYSIWYG is helpful for editing, when it comes to muting/threading/etc it’s helpful for ppl to all be looking at the same page so they can easily model what others are seeing. Having some ppl see a user’s comments but the author not, or key commenters not, is quite costly for social transparency, and understanding social dynamics.
So like, do you distrust writers using substack? Because substack writers can just ban people from commenting. Or more concretely, do you distrust Scott to garden his own space on ACX?
It's normally out of my mind, but whenever I'm reminded of it, I'm like "damn, I wonder how many mistaken articles I read and didn't realize it because the author banned or discouraged their best (would be) critiques." (Substack has other problems though like lack of karma that makes it hard to find good comments anyway, which I would want to fix first.)
Giving authors the ability to ban people they don't want commenting is so common that it feels like a Chesterton's Fence to me.
It could also just be a race to the bottom to appeal to unhealthy motivation, kind of like YouTube creating Shorts to compete with TikTok.
Comment threads are conversations! If you have one person in a conversation who can't see other participants, everything gets confusing and weird.
The additional confusion seems pretty minimal, if the muted comments are clearly marked so others are aware that the author can't see them. (Compare to the baseline confusion where I'm already pretty unsure who has read which other comments.)
I just don't get how this is worse than making it so that certain perspectives are completely missing from the comments.
I quite dislike the idea of people being able to moderate their content in this fashion - that just isn't what a public discussion is in my view - but thanks for being transparent about this change.
Here's a hypothesis for the crux of the disagreement in this comments section:
There's a minor identity crisis about whether LW is/should primarily be a community blog or a public forum.
If it is to be a community blog, then the focus is in the posts section, and the purpose of moderation should be to attract all the rationality bloggers to post their content in one place.
If it is to be a public forum/reddit (I was surprised at people referring to it like so), then the focus is in the comments section, and the main purpose of moderation should be to protect all viewpoints and keep a bare minimum of civility in a neutral and open discussion.
What was the logic behind having a karma threshold for moderation? What were you afraid would happen if low karma people could moderate, especially on their personal blog?
Does allowing users to moderate mean the moderation team of the website will not also be moderating those posts? If so, that seems to have two implications: one, this eases the workload of the moderation team; two, this puts a lot more responsibility on the shoulders of those contributors.
Ah, sorry, looks like I forgot to mention that in the post above. There is a checkbox you can check on your profile that says "I'm happy for LW site moderators to help enforce my policy", which then makes it so that the sitewide moderators will try to help with your moderation.
We will also continue enforcing the frontpage guidelines on all frontpage posts, in addition to whatever guidelines the author has set up.
I'm still somewhat uncomfortable with authors being able to moderate front-page comments, but I suppose it could be an interesting experiment to see if they use this power responsibly or if it gets abused.
I think that there should also be an option to collapse comments (as per Reddit), instead of actually deleting them. I would suggest that very few comments are actually so bad that they need to be deleted, most of the time it's simply a matter of reducing the incentive to incite controversy in order to get more people replying to your comment.
Anyway, I'm really hoping that it encourages some of the old guard to post more of their content on Less Wrong.
I think this is extremely bad. Letting anyone, no matter how prominent, costlessly remove/silence others is toxic to the principle of open debate.
At minimum, there should be a substantial penalty for banning and deleting comments. And not a subtraction, a multiplication. My first instinct would be to use the fraction of users you have taken action against as a proportional penalty to your karma, for all purposes. Or, slightly more complex, take the total "raw score" of karma of all users you've taken action against, divide by the total "...
If a comment of yours is ever deleted, you will automatically receive a PM with the text of your comment, so you don’t lose the content of your comment.
My intuition is that it would be better to allow users to see posts of their own that were deleted in a grayed out way instead of going through the way of sending an PM.
If there's a troll, sending a troll a PM that one of their post got deleted creates a stronger invitation to respond. That especially goes for deletions without giving reasons.
In addition I would advocate that posts that are deleted ...
Will there be a policy on banned topics, such as e.g. politics, or will that be left to author discretion as part of moderation? Perhaps topics that are banned from promotion / front page (regardless of upvotes and comments) but are fine otherwise?
If certain things are banned, can they please be listed and defined more explicitly? This came up recently in another thread and I wasn't answered there.
[I will move this into meta in a few days, but this seemed important enough to have around on the frontpage for a bit]
Here is a short post with some of the moderation changes we are implementing. Ray, Ben and me are working on some more posts explaining some of our deeper reasoning, so this is just a list with some quick updates.
Even before the start of the open beta, I intended to allow trusted users to moderate their personal pages. The reasoning I outlined in our initial announcement post was as follows:
“We want to give trusted authors moderation powers for the discussions on their own posts, allowing them to foster their own discussion norms, and giving them their own sphere of influence on the discussion platform. We hope this will both make the lives of our top authors better and will also create a form of competition between different cultures and moderation paradigms on Lesswrong.”
And I also gave some further perspectives on this in my “Models of Moderation” post that I posted a week ago.
We now finally got around to implement the technology for this. But the big question that has been on my mind while working on the implementation has been:
Me, Ray, Ben and Vaniver talked for quite a while about the pros and cons, and considered a bunch of perspectives, but the two major considerations on our mind were:
After a good amount of internal discussion, as well as feedback from some of the top content contributors on LW (including Eliezer), we settled on allowing users above 2000 karma to moderate their own frontpage posts, and allow users above 100 karma to moderate their personal blogs. This strikes me as the best compromise between the different considerations we had.
Here are the details about the implementation:
I also want to allow users to create private comments on posts, that are only visible to themselves and the author of the post, and allow authors to make comments private (as an alternative to deleting them). But that will have to wait until we get around to implementing it.
We tested this reasonably thoroughly, but there is definitely a chance we missed something, so let us know if you notice any weird behavior around commenting on posts, or using the moderation tools, and we will fix it ASAP.