Level 4: Contributions not asking 101-level stuff
There should be a clear distinction between asking 101-level things, being unaware of 101-level things, and criticising 101-level things. Criticising-while-aware is very valuable, being unaware is sometimes fine (not everything is really relevant, and an unaware frame takes a fresh look). Passively asking things that could've been learned from standard introductions might need to happen in a different place, though I don't currently notice it on LW with any regularity.
Strong agree about the distinction.
It's not overwhelming yet, but I feel like I am seeing a real increase in users asking questions and sharing ideas of things that are addressed in standard intros. In the last week or so we started blocking first posts and comments that are strongly in this category (you'll soon be able to see them in the Rejected Content section of the site) and redirecting people to the Intro AI Open threads.
Hello Ruby, moderators and all,
I must say I feel grateful seeing the moderator's quest towards improving things, especially through dialogue and transparency. It is engaging, and it makes it more clear what LW is, and what I would have to do to contribute. Thanks a lot for working on this and doing the internal as well as external work needed to jump over the hidden iceberg. Massive thumbs up.
I also wanted to give my reflections and thoughts on the post.
From my own experience, getting down-karmaed and also critical feedback on my first posts, was a jarring experience. I have since understood that my posts weren't up to par, and I am working to improve, but I still believe I would have, with some effort, have improved and been able to contribute much more if given a hand and some friendly advice. I would therefore decidedly prefer a clear rejection instead over a death by a thousand cuts. (or as one Norwegian king supposedly said: Kill me, my liege, but not with grout. (grøt))
From my perspective as a new, but sincere person, I would have loved if my posts simply got rejected, and landed somewhere where people could, if they so desired, give me that friendly hand, some advice and a welcome.
Therefore, from the options above, I strongly prefer more moderation before things are posted. Not only for the betterment of signal-to-noise, but as a safeguard that protects new users from the potentially corroding touch of down-karma and negative feedback.
I would also prefer if posts ending up in the rejection pile weren't also subjected to additional 'punishment' by getting downvoted and critiqued. The way it stands now, posts can turn into porcupines, and that is really not a healthy look. You have already mentioned great ideas, like having a 101-zone to send people that are new. And with regard to posts, writing 'good enough' is hard, but reading the posts of others is a start.
A spur of the moment might be to add a simple icon on rejected posts that you believe could improve from constructive feedback. I mean, wouldn't it be great if posts that landed in the rejected folder, got constructive feedback, and a couple of weeks later landed on the front page?
Acknowledging the importance of Comments:
In Killing Socrates, my view on it is that even though moderating posts is very important, I see more how the commenting-culture and skill could improve. In my comment here on that post, I contrast two comments, and how they get opposing voting 'scores', which seems odd considering the guidelines.
You can see a lot of good posts on the front page, but there is no 'good comments' front-page. I see the post by Duncan_Sabien as exemplifying the importance of acknowledging/taking seriously the mutual dependent relationship between builders and commentors, that both of them are, in essence, builders. From that point of view, well-written, acknowledging the author, gives good constructive feedback and/or colorful, fun, empathic and compassionate etc. comments, would benefit from receiving the same scrutiny as posts, but also to be highlighted. The goal isn't to have every comment be stellar or even the same visibility as posts, but to see these as two sides of the same coin.
When I studied Social Work, I also started to notice the lesser value that is put on inner growth, soft skills and things like culture. Even though a big reason people quit their jobs, get sick and/or underperform, isn't because they have little pay, do not like their work or are not appreciated in society - it is often related to the climate where they work. Especially, how their superiors are acting, has far-reaching consequences.
Comments are, in my opinion, the continuation of posts. If they are better than the post, they elevate it, if they are worse, they diminish its worth, and if they are at the same level, they preserve and uphold it. The same goes for good discussions in the comment-section. Good exchanges are a goal in and of themselves, and more so when they are happening in the vicinity of important posts.
I hope this is a somewhat useful comment, and I want to repeat that I appreciate your effort and the process you are making in steering LW clear of a potential iceberg. Thank you, I appreciate it.
Kindly,
Caerulea-Lawrence
I super appreciate this in-depth sharing of your experience and thoughts.
I'm curious to hear from other newish users about whether they'd have preferred "rejection" vs "approval-but-downvoting/comments."
One thing that I think is missing (maybe just beyond the scope of this post) is thinking about newcomers with a positive frame: how do we help them get up to speed, be welcomed, and become useful contributors?
You could imagine periodic open posts, for instance, where we invite 101-style questions, post your objection to AI risks, etc where more experienced folks could answer those kind of things without cluttering up the main site. Possibly multiple more specific such threads if there's enough interest.
Then you can tell people who try to post level 1-3 stuff that they should go to those threads instead, and help make sure they get attention.
I'm sure there are other ideas as well -- the main point is that we should think of both positive as well as negative actions to take in response to an influx of newbies.
Yeah I think the AI Questions Open Thread series has been good for this (and I've been directing people to that where appropriate).
I've also been mulling over how to make a good "okay, you're new, you want to contribute to the AI discussion on LessWrong, what stuff is useful to do?" post.
Oh, oops. I originally intended to include my thoughts about newcomers in this post and must have gotten distracted before I did so. I just add a whole section on that. Thanks for flagging.
Periodic open posts is one of my current favorite ideas of what to do here.
A hard thing about trying to be transparent about our moderation decisions and actions is that this also requires publicly calling out a user or their content. So you get more transparency but also more embarrassment. I don't have any good solution to this.
Maybe you could not display usernames in the rejected posts section (though this might conflict with transparency if a user feels they are being personally targeted).
I think that's solved by a user claiming ownership of their post if they feel it's necessary and the mod team can dispute if a user inaccurately claims it.
Perhaps a kind "101 Friendly" tag for posts that can be optionally applied. I could imagine it being applied to posts like The basic reasons I expect AGI ruin which signals there isn't an expectation of prior familiarity with basic arguments. We could literally tie to this a user group permissions category, of certain users are restricted to commenting on posts with that tag.
If you consider that post 101 friendly, I suggest there may be some curse of knowledge going on. Perhaps you expect the average 101 question to come from a university student with a background in computer science? Otherwise I suggest asking people without such a background (e.g. your parents or grandparents, or random members of the public) to read such a post; I expect they will not consider it comprehensible.
It's not reasonable to start site content from zero, of course, but that post is very, very, very far from zero.
Anyway, considering that you want to consider 101 questions in moderation decisions, I really wonder what the target audience for this 101 content is supposed to be, and what kinds of prerequisites you expect them to have. Are highschool students fine? University students from soft sciences? Etc.
To suggest a counterexample to what could be considered 101 content, consider that the first post in the original sequences was The Simple Truth, a parable about a shepherd, written to explain what it means for something to be "true". Now that's starting from zero.
I'm curious about whether I should change my shortform posting behaviour in response to higher site quality standards. I currently perceive it to be an alright place to post things that are quick and not aiming to be well-written or particularly useful for others to read because it doesn't clutter up the website the way a post or comment on other people's posts would.
I'm still interested in Shortform being the place for established users to freely post their rough thoughts. I'd say you're good and shouldn't worry.
The rough thoughts of a person who has the site culture, knows all the basics, has a decent chance of having some valuable such that it's good for people to post them. On on the other hand, some people's high effort stuff is not very interesting, e.g. from new users. I wouldn't want Shortform to become super low SNR because we put low quality users indiscriminately, and by extension, have that content all over the site (Recent Discussion, All Posts, etc) which it would be right now unless we changed some things.
[Raemon for a while was thinking to send users not quite up to par to Shortform, but I've told him I don't endorse that use case currently. We'll have to think about it more/perhaps change various parts of the site, e.g. what goes in Recent Discussion, before endorsing that.]
send users not quite up to par to Shortform [...] I don't endorse that use case currently
+1. Shortform is harder to filter (no titles) and gets less scrutiny, so directing new users in there runs a higher-than-elsewhere risk of having that whole area crumble. I think with the current shape of the feature it should be the other way around, posting Shortforms should be gated by the same kinds of things that remove rate limits or allow stronger voting.
Can you say elaborate on your thinking here? Especially this line:
I think with the current shape of the feature it should be the other way around, posting Shortforms should be gated by the same kinds of things that remove rate limits or allow stronger voting.
If we already grant the premise that broken windows would make Shortform a low-quality place, and that it has less ability to defend itself and recover than other places, then it seems to follow that directing people who might have a greater than usual window-breaking disposition there is the opposite of what you would want to do. In this hypothetical it's appropriate to keep the defense of this feature in mind when making new interventions, like choosing where to direct new users.
(This conclusion would change if the shape of the feature changes, making it more capable of defending itself, or if the premise that makes it out to be unusually defenseless is seen to be wrong. But I don't understand how the conclusion wouldn't follow from the premise.)
I want to distinguish "send people to shortform to write their own shortform posts" from "send people there to comment on other people's shortform posts." I meant the former, it sounds like your comment is expecting the latter, but I'm not sure if I'm missing something.
(Separately, to address one of the points here: I'm imagining also changing shortform to allow options like "only a whitelist of friends can comment on your shortform posts", while making it easier to subscribe to shortform from people you like, which is how FB/Twitter actually work which is what shortform was largely inspired by)
I meant the former, it sounds like your comment is expecting the latter
No, I also mean "their own shortform posts", with the expectation that it might make the list of all shortform posts (top level comments) made on some day a less interesting thing to browse. With how little voting there is currently on shortform posts, it's not as easy to ignore most of them if you don't want to go through all of them, trusting other users to have done a sensible job of evaluating them.
A lot of the shortform posts are just sitting at +1 to +4. A regular post would need to be doing something very wrong to retain such a near-default rating, but good shortform posts remain there, I assume because for most people All Posts is not the bookmarked page and so the shortform posts don't get much more visibility than random comments. Not paying specific attention, I'm not even sure there is any difference between shortform posts and random comments on old posts, that influences their visibility on the main page.
(Btw, the limit=100 part of the All Posts URL that uncollapses lists of posts is a key usability feature for that page for me. I don't remember how I learned about it and I don't see it in the UI. And I don't see a corresponding thing that works for shortforms. So with my obscure custom starting page URL that makes lists of regular posts more convenient than the default, shortforms become by comparison less convenient than lists of regular posts. Not having to go through the inconvenience of clicking on that Load More is important, the default alternative is ignoring everything behind it. Having to click it for regular posts and then separately and additionally for shortform posts makes it so much less likely for the unlucky shortform posts that are behind it to be seen.)
making it easier to subscribe to shortform from people you like
That's the sort of thing I was gesturing at with "shortforms should be gated" (as an example intervention that makes some sense, though not necessarily that good overall, intended as a contrast to the intervention discussed in this thread). But using site-wide metrics, instead of having everyone do the work on their own. The latter is painful, never good enough to make it comfortable to ignore things that don't make it to the list, creating a pressure to do the chore of keeping the list updated. Lots of costs.
An update this made me have: shortform comments should maybe get a self-upvote based on your strongvote strength rather than weakvote, so that it's a bit easier to scan the new content for stuff that's likely to be interesting (and this seems fairly principled since we already do this for posts, which shortform are more like than comments)
(although Ruby noted "maybe we want to build Medium votes (which we've been thinking about anyway) and make shortform use them")
Huh? How does strong-upvote-by-default make it any easier to distinguish the "likely to be interesting"? Oh, with the model that old-timers are more interesting always, I guess I get it, but I don't actually agree.
I'd drop strong voting entirely, perhaps replacing it with medium-voting, but definitely keep it intentional and optional. I really hate seeing comments (even mine) with karma in the teens or twenties and just a few votes.
No, I also mean "their own shortform posts", with the expectation that it might make the list of all shortform posts (top level comments) made on some day a less interesting thing to browse.
Ah. My impression was that ship had kinda already sailed.
Ah. My impression was that ship had kinda already sailed.
You can always make the problem worse.
My feeling is that yes, you could probably find ways to make low quality users posting to Shortform not dilute better Shortforms from better users, but (1) that's a bunch of work we haven't yet done and has opportunity cost, (2) once I'm doing that work, I might as well do it for posts? To the extent low quality users want to write posts (and I think they do), it's better for author and readers if they're a post (e.g. you get ToC, etc)
You can always make the problem worse.
Without necessarily disagreeing with the rest of the comment, I'm not sure I buy this? The mechanism by which I see shortform is either by seeing it in recent discussion (in which case posts, comments or shortforms contribute to about the same rate of dilution, there's nothing special about shortform), or I subscribe to the shortform of a particular author I like (which I can find via clicking on the user profiles of authors with otherwise highly upvoted comments)
I'm not saying there's anything special about it, but just as I don't want to make the SNR of the others worse, I don't want that for Shortform either. I think discoverability for Shortform is a problem to be solved. Question is how do you know which authors have shortforms they regularly post to? Seems like not the best flow for the only way to find them is via manually checking profiles. (Not that this isn't solvable, but it's a thing to do, and if you give up on Shortform SNR, you lose option value and things that maybe could be made to work, etc.
I’m just one datapoint, but just an hour ago I found an interesting shortform via allPosts and left a comment on it.
Yeah I guess I’m just not modeling allPosts page, which is maybe more on the borderline of usable
Oh, that disconnect has caused odd decisions before. I never look at the homepage, and just have a bookmark to allPosts. I've missed out on Petrov Day announcements (and impact), and am often confused about discussions on "discoverability" which assume homepage starting point for everyone.
I suspect that blocking content feels like a much harsher action than redirecting it elsewhere, so I wonder if you could write a custom feature for redirecting a post to the current beginner question thread?
As can be inferred from the title, it's been a bit more than a week when I wrote most of this. I've had many more thoughts in that time, but seems very much still worth sharing this.
The LessWrong team is currently in the midst of rethinking/redesigning/upgrading our moderation policies and principles (announcement). In order to get feedback throughout this process rather than just at the end, here's a dump of things currently on my mind. Many of these are hypotheses and open questions, so feedback is very welcome.
This is a bit of a jumble of thoughts, questions, frames, etc, vaguely ordered.
The enforced standard is a function of how many new people are showing up
With the large influx of new users recently, we realized that we have to enforce a higher standard or risk having the site's quality greatly diluted: Latest Posts filled with poorly written posts about trivial points, comment threads full of incoherent or very 101 questions, etc.
The question then is "how much are we raising it?" where I do think there's quite a range. Here's a suggestive scale I didn't put a tonne of thought into, not every reasonable level is included:
To date, I'd say the LessWrong moderators have approved content that's Level 2 or above, even if we thought it was meh or even bad ("approve and downvote" was a common move). Fearing that this would be inadequate with a large influx of low-quality users, it seemed that the LessWrong mod team should outright reject a lot more content. But if so, where are we raising the bar to?
I've felt that any of Levels 3-5 are plausible, with different tradeoffs. What I'm realizing is the standard that makes sense to enforce depends on how many new people you're getting.
Moderation: a game of tradeoffs
Let's pretend moderation is just binary classification task where the moderator just has to make an "accept" or "reject" calls on posts/comments/users. For a fixed level of discernment (you can get better over time, but approach any decision, you're only so good), you can alter your thresholds for acceptance/rejection, and thereby trade false positives against false negatives. See RoC/AUC for binary classifiers.
Moderation is that. I both don't want to accidentally reject good stuff but also don't want to accept bad stuff that'll erode site quality.
My judgment is only so good, so I am going to make mistakes. For sure. I've made them so far and fortunately others caught them (though I must guess I've made some that have not been caught). That seems unavoidable.
But given I'm going to make mistakes, should I err on the side of too many acceptances or too many rejections?
Depends on the volume
My feeling, as above, is the answer depends on volume. If we don't get that many new users, there's not a risk of them swarming the site and destroying good conversation. I think this has been the situation until recently so it's been okay having a lower standard. In a hypothetical world where we get 10x as many new users as we currently do, we might need to risk having many false negatives to avoid getting too many false positives, because the site couldn't absorb that many bad new users.
My current thinking[1]
My feeling is we will be requiring new contributions to be between Level 2.5 and Level 4 right now, but keeping an eye on volume of new users. If the storm comes (and it might be bursty), we might want to raise the standard higher.
Question: where to filter?
New users generally write lower-quality content than established ones. Some of that content is still valuable, just less valuable than other stuff around, but some of it is probably isn't worth anyone's time to read.
It's been a persistent question for me of how to assess which content should be displayed on the site, and how, and who should do the assessing.
Where and how content gets filtered is one the major knobs/dials the moderation team has in our toolkit.
Some Options
My current thinking[1]
I think we will be rejecting content to Level 2.5-4 as above for now (but providing feedback and being transparent), and then beyond that approving and downvoting if we're less confident something is bad but personally don't like it. This is a mix of 1 and 2. I'm not excited about different major site sections right now, that said maybe a little bit of that for 101 AI content.
Transparency & Accountability
I think many things go better when one (or a few people) in charge and can just make decisions. This can avoid many failure modes of excessive democratization. But also things go better there if there's some accountability for those few people, and accountability requires knowing what's actually going on, i.e., transparency.
As the judgment calls get closer to a blurry line, I don't feel great about being the one to decide whether or not a person's post or comment or self belongs on LessWrong. I will make mistakes. But also – tradeoffs – I don't want LessWrong to get massively diluted because I wasn't willing to reject enough people.
A solution for this that I think is very promising is that the LessWrong team will make relatively more decisions on our own to reject posts/comments/users, raising it from what we did before, but we'll make it possible to audit us. Rejected content will be publicly visible on a Rejected Content section of the site, and rejected items will usually include some indication of why we rejected it. I predict some people will be motivated to look over the reject pile, and call out places where we screwed up. Or at the very least, people can look and see that we're not just rejecting things because we disagreed with them. What I hope is that people look over the rejected stuff and go "overwhelmingly, seems good that this stuff is not on LessWrong".
But the transparency and accountability isn't about just that feature, it's a principle I'm generally inclined towards trying to achieve.
Tensions in Transparency
A hard thing about trying to be transparent about our moderation decisions and actions is that this also requires publicly calling out a user or their content. So you get more transparency but also more embarrassment. I don't have any good solution to this.
Ideally, we'd use all public comments to moderate rather than Private Messages, but yeah, can make people feel called out rather than taking them aside quietly. The other challenge here is that moderating in public requires figuring out how to not communicate just to one person, but many, and that takes more time when the team is already short on it.
The Expanding Toolbox
I think of the coolest things about the LessWrong team (and something core to the LessWrong/Lightcone philosophy) is that we wear all the hats, we are all both moderators and software engineers, allowing us to build and iterate on our tools in a way I don't think most forum moderators get to.
So we get to experiment a lot our moderation stack, if we want.
Rate Limits
Some months ago, we introduced rate limits as a tool. Instead of a time-bound ban (1-month, 3-months, etc) that's pretty heavy and rarely done[2], we can see users who are pulling down quality/behaving badly and simultaneously warn them and slow them down, hopefully encouraging them to put more effort into their contributions.
An informal poll showed that many people liked the idea of being rate limited more than having the visibility of their content reduced.
I do like have a gradation of options, and think this is pretty great to have in addition bans.
My fears about Rate Limiting
I do fear that done wrong, rate limiting could do a lot of harm and I'm very keen to avoid that as we experiment with it.
I do think some rate limits (the ones we currently have set up?) are 80% of a ban without feeling to us moderators like that. Also historically with bans there was a public explanation and a set duration, and we're not currently communicating that well about them (we can and will fix this though – we really do have to though).
Separating limits for post and comments (and comments on your own post vs other people's) is a basic next step. Also more granularity for the limits. I wonder if 10 comments/a week might be better than 1/day, because you could spend them all on a conversation on one day.
Why focus on moderation now
The proximate cause for focusing here is the increase in new users who find their way to the site because the topic of AI is becoming much higher profile. LessWrong both does well in search results and also gets linked from my many places. Also some older users are now posting and commenting for the first time since the topic feels urgent.
Goal: Maintaining the Signal-to-Noise Ratio
Or more accurately, maintaining or improving the shape of the distribution of quality that users experience when they visit the site. If you're reader, when you visit the homepage, do you find it easier or harder to find posts that you want to read? If you're a poster, do you find the comments more or less valuable to read?
I care a lot that as the site grows, the experience of quality only goes up. The question I want to ask with new users (or existing users even), is given the current site setup, how is the signal-to-noise ratio changed?
Frame: The Wall & The Spire
When launching the LessWrong team into a big stretch of time focusing on moderation, I said that I want us to build both walls surrounding our site (with gates and chambers and perhaps inner and outer walls that allow a progression inwards) and also a spire to elevate the discourse unto the heavens.
The imagery is supposed to emphasise that although the proximate cause for focusing on moderation has been a dramatic increase in new poster and commenters, I don't want our moderation efforts just focused on keeping site quality "as good as it has been to date", we should aim to raise the quality level too.
This works well because (a) many new moderation tools work for new and established users, (b) clarifying site values and standards is useful for new and established users.
Added: How to relate to 101 content
I've been thinking about this topic but forgot to write it up here. Doing so in response to a comment.
I am unsure what to do with and how to relate 101 users/101 content. Here's a mix of thoughts.
For a while (particularly when we were experiencing a large burst of new accounts), I was thinking that we should basically reject 101 submissions. By 101, I mean questions and comments that you wouldn't make if you'd read Superintelligence, new about orthogonality thesis, existing discussion of Tool AI, and other topics that are generally treated as very basic background knowledge. (As above, I think LessWrong could absorb this and still be suitable for serious progress, but it's not a trivial design problem and there are opportunity costs for the team's time.)
Mulling over the policy of approx no 101 AI content for a while, it doesn't feel good to shut our doors to it completely. I want people coming for that to have some kind of home, and it seems plausible good if LessWrong serves as a general beacon of sanity around the topic of 101 AI, standard questions and objections, etc.
In the near-term, solutions I like include:
I'd previously considered redirecting new users to stampy.ai, but at present, that UI doesn't seem enough. I think people should get do the equivalent of make a question post (similar to a basic StackOverflow post), whereas Stampy seems much more to want you to have a sentence-length question, and match to an existing one (taking about three layers before you can propose your own sentence level question).
Another idea of interest to me is "Argument Wiki". This would be useful for a range of users at a range of knowledge. E.g., I'd be interested in it as a place that compiles all the arguments for and against Sharp Left Turn, but it could also have the answers to more elementary questions such as "wouldn't a really smart AI realize it shouldn't be evil?" or whatever.
I'll add that a lot of the team's current work is trying to get ahead of the problems here, even before they're obviously really bad, because I worry that once the Great Influx is upon us, we'll take a lot of damage if we're not already prepared to handle it.
This means what I'm thinking today and maybe tomorrow, it's suggestive but not definitive of what we'll settle on by the end of the current period of refiguring out moderation stuff.
It feels like a Big Deal to ban a user even temporarily, so I think only having this as a tool has caused us to historically under-moderate.