All of Neil 's Comments + Replies

Neil 62

Can confirm. Half the LessWrong posts I've read in my life were read in the shower.

Neil 150

no one's getting a million dollars and an invitation the the beisutsukai 

3Sniffnoy
Oh, nice! I guess this is basically saying the same thing as that, I'll add a link.
Neil 710

I like object-level posts that also aren't about AI. They're a minority on LW now, so they feel like high signals in a sea of noise. (That doesn't mean they're necessarily more signally, just that the rarity makes it seem that way to me.)

3Arjun Panickssery
More object-level content: Why Have Sentence Lengths Decreased? — LessWrong
Neil 4914

It felt odd to read that and think "this isn't directed toward me, I could skip if I wanted to". Like I don't know how to articulate the feeling, but it's an odd "woah text-not-for-humans is going to become more common isn't it". Just feels strange to be left behind. 

Neil 3111

Thank you for this. I feel like a general policy of "please at least disclose" would make me feel significantly less insane when reading certain posts. 

Neil 41

Have you tried iterating on this? Like, the "I don't care about the word prodrome'" sounds like the kind of thing you could include in your prompt and reiterate until everything you don't like about the LLM's responses is solved or you run out of ideas. 

Also fyi ChatGPT Deep Research uses the "o3" model, not 4o, even if it says 4o at the top left (you can try running Deep Research with any of the models selected in the top left and it will output the same kind of thing).

o3 was RLed (!) into being particularly good at web search (and tangential skills ... (read more)

Neil 10

My highlight link didn't work but in the second example, this is the particular passage that drove me crazy:

The punchline works precisely because we recognize that slightly sheepish feeling of being reflexively nice to inanimate objects. It transforms our "irrational" politeness into accidental foresight.

The joke hints at an important truth, even if it gets the mechanism wrong: our conversations with current artificial intelligences may not be as consequence-free as they seem.

1Nicholas Andresen
Thanks for articulating this – it's genuinely helpful. You've pinpointed a section I found particularly difficult to write. Specifically, the paragraph explaining the comic's punchline went through maybe ten drafts. I knew why the punchline worked, but kept fumbling the articulation. I ended up in a long back-and-forth with Claude trying to refine the phrasing to be precise and concise, and that sentence is the endpoint. I can see that the process seems to have sanded off the human feel. As for the "hints at an important truth" line... that phrasing feels generic in retrospect. I suspect you're right – after the prior paragraph I probably just grabbed the first functional connector I could find (a direct Claude suggestion I didn't think about too much) just to move the essay forward. It does seem like the type of cliché I was trying to avoid. Point taken that leveraging LLM assistance without falling into the uncanny valley feel is tricky, and I didn't quite nail it there. Appreciate the pointer. My general workflow involves writing the outline and main content myself (this essay actually took several weeks, though I'm hoping to get faster with practice!) and then using LLMs as a grammar/syntax checker, to help with sign-posting and logical flow, or to help rephrase awkward or run-on sentences. Primarily I'm trying to make my writing more information dense and clear.
Neil 20

That's fair, I think I was being overconfident and frustrated, such that these don't express my real preferences.

But I did make it clear these were preferences unrelated to my call, which was "you should warn people" not "you should avoid direct LLM output entirely". I wouldn't want such a policy, and wouldn't know how to enforce it anyway. 

I think I'm allowed to have an unreasonable opinion like "I will read no LLM output I don't prompt myself, please stop shoving it into my face" and not get called on epistemic grounds except in the context of "wait... (read more)

2the gears to ascension
Sounds interesting, I talk to LLMs quite a bit as well, I'm interested in any tricks you've picked up. I put quite a lot of effort into pushing them to be concise and grounded. eg, I think an LLM bot designed by me would only get banned for being an LLM, despite consistently having useful things to say when writing comments - which, relatedly, would probably not happen super often, despite the AI reading a lot of posts and comments - it would be mostly showing up in threads where someone said something that seemed to need a specific kind of asking them for clarification, and I'd be doing prompt design for the goal of making the AI itself be evaluating its few and very short comments against a high bar of postability. I also think a very well designed summarizer prompt would be useful to build directly into the site, mostly because otherwise it's a bunch of work to summarize each post before reading it - I often am frustrated that there isn't a built-in overview of a post, ideally one line on the homepage, a few lines at the top of each post. Posts where the author writes a title which accurately describes post contents and an overview at the top are great but rare(r than I'd prefer they be); the issue is that pasting a post and asking for an overview typically gets awful results. My favorite trick for asking for overviews is "Very heavily prefer direct quotes any time possible." also, call it compression, not summarization, for a few reasons - unsure how long those concepts will be different, but usually what I want is more like the former, in places where the concepts differ. However, given the culture on the site, I currently feel like I'm going to get disapproval for even suggesting this. Eg, There are circumstances where I don't think this is accurate, in ways beyond just "that's a lot of asking, though!" - I would typically want to ask an LLM to help me enumerate a bunch of ways to put something, and then I'd pick the ones that seem promising. I would only
2No77e
Sorry for not replying in more detail, but in the meantime it'd be quite interesting to know whether the authors of these posts confirm that at least some parts of them are copy-pasted from LLM output. I don't want to call them out (and I wouldn't have much against it), but I feel like knowing it would be pretty important for this discussion. @Alexander Gietelink Oldenziel, @Nicholas Andresen you've written the posts linked in the quote. What do you say?  (not sure whether the authors are going to get a notification with the tag, but I guess trying doesn't hurt)
1Neil
My highlight link didn't work but in the second example, this is the particular passage that drove me crazy:
Neil 62

If it doesn't clutter the UI too much, I think an explicit message near the submit button saying "please disclose if part of your post is copy-pasted from an LLM" would go a long way! 

If this is the way the LW garden-keepers feel about LLM output, then why not make that stance more explicit? Can't find a policy for this in the FAQ either!

I think some users here think LLM output can be high value reading and they don't think a warning is necessary—that they're acting in good faith and would follow a prompt to insert a warning if given. 

Neil 60

Touching. Thank you for this. 

When I was 11 I cut off some of my very-much-alive cat's fur to ensure future cloning would be possible, and put it in a little plastic bag I hid from my parents. He died when I was 15, and the bag is still somewhere in my Trunk of Everything. 

I don't imagine there's much genetic content left but also I have a vague intuition that we severely underestimate how much information a superintelligence could extract from reality—so I'll keep onto a lingering hope.

My past self would have wanted me to keep tabs on how the te... (read more)

3Metacelsus
I'm sorry to break this to you, but cloning requires live cells, not just DNA. This is one of the reasons why it's so hard to bring back the woolly mammoth. (The other reason is that it's really hard to do IVF on elephants.) So if you want to make a clone, you'll need to do something like what I did (take cells and preserve them in liquid nitrogen).
Neil 2510

Can we have a LessWrong official stance to LLM writing? 

The last 2 posts I read contained what I'm ~95% sure is LLM writing, and both times I felt betrayed, annoyed, and desirous to skip ahead.

I would feel saner if there were a "this post was partially AI written" tag authors could add to as a warning. I think an informal standard of courteously warning people could work too, but that requires slow coordination-by-osmosis. 



Unrelatedly to my call, and as a personal opinion, I don't think you're adding any value to me if you include even a single p... (read more)

6No77e
You seem overconfident to me. Some things that kinda raised epistemic red flags from both comments above: It's really hard to believe this and seems like a bad exaggeration. Both models sometimes output good things, and someone who copy-pastes their paragraphs on LW could have gone through a bunch of rounds of selection. You might already have read and liked a bunch of LLM-generated content, but you only recognize it if you don't like it! Unfortunately, there are people who have a similar kind of washed-out writing style, and if I don't see the posts, it's hard for me to just trust your judgment here. Was the info content good or not? If it wasn't, why were you "desirous of skipping ahead" and not just stopping to read? Like, it seems like you still wanted to read the posts for some reason, but if that's the case then you were getting some value from LLM-generated content, no? This is almost the most obvious ChatGPT-ese possible. Is this the kind of thing you're talking about? There's plenty of LLM-generated text that just doesn't sound like that and maybe you just dislike a subset of LLM-generated content that sounds like that.
8habryka
Yeah, our policy is to reject anything that looks like it was written or heavily edited with LLMs from new users, and I tend to downvote LLM-written content from approved users, but it is getting harder and harder to detect the difference on a quick skim, so content moderation has been getting harder. 
3Raemon
LW moderators have a policy of generally rejecting LLM stuff, but some things slip through cracks. (I think maybe LLM writing got a bit better recently and some of the cues I used are less reliable now, so I may have been missing some)
8Seth Herd
I agree that it would be useful to have an official position. There is no official position AFAIK but individuals in management have expressed the opinion that uncredited AI writing on LW is bad because it pollutes the epistemic commons (my phrase and interpretation). I agree with this statement. I don't care if an AI did the writing as long as a human is vouching for the ideas making sense. If no human is actively vouching for the ideas and claims being plausibly correct and useful, I don't want to see it. There are more useful ideas here than I have time to take in. That applies even if the authorship was entirely human. Human slop pollutes the epistemic commons just as much as AI slop. If AI is used to improve the writing, and the human is vouching for the claims and ideas, I think it can be substantially useful. Having writing help can get more things from draft to post/comment, and better writing can reduce epistemic pollution. So I'm happy to read LLM-aided but not LLM-created content on LW. I strongly believe that authorship should be clearly stated. It's considered an ethical violation in academia to publish others' ideas as your own, and that standard seems like it should include LLM-generated ideas. It is not necessary or customary in academia to clearly state editing/writing assistance IF it's very clear those assistants provided absolutely no ideas. I think that's the right standard on LW, too.
Neil 10

My calendar reminder didn't go off, are submissions closed-closed? 

1Henry Prowbell
I'm afraid so. Sorry. We hope to run more in the future!
Neil 10

Oh yeah no problem with writing with LLMs, only doing it without disclosing it. Though I guess this wasn't the case here, sry for flagging this. 

I'm not sure I want to change my approach next time though, bc I do feel like I should be on my toes. Beware of drifting too much toward the LLM's stylebook I guess. 

Neil 20

Maybe I'm going crazy, but the frequent use of qualifiers for almost every noun in your writing screams of "LLM" to me. Did you use LLM assistance? I don't get that same feel from your comments, so I'm learning toward an AI having written only the Shortform itself.

If you did use AI, I'd be in favor of you disclosing that so that people like me don't feel like they're gradually going insane. 

If not, then I'm sorry and retract this. (Though not sure what to tell you—I think this writing style feels too formal and filled with fluff like "crucial" or "invaluable", and I bet you'll increasingly be taken for an AI in other contexts.)

3Alexander Gietelink Oldenziel
Yes I use LLMs in my writing [not this comment] and I strongly encourage others to do so too.  This the age of Cyborgism. Jumping on making use of the new capabilities opening up will likely be key to getting alignment right. AI is coming, whether you like it or not.  There is also a mundane reason: I have an order of magnitude more ideas than I can write down. Using LLMs allows me to write an essay in 30 min which otherwise would take half a day. 

The original post, the actual bet, and the short scuffle in the comments is exactly the kind of epistemic virtue, basic respect, and straight-talking object-level discussion that I like about LessWrong. 

I'm surprised and saddened that there aren't more posts like this one around (prediction markets are one thing; loud, public bets on carefully written LW posts are another).

Having something like this occur every ~month seems important from the standpoint of keeping the garden on its toes and remind everyone that beliefs must pay rent, possibly in the form of PayPal cash transfers. 

I wrote this after watching Oppenheimer and noticing with horror that I wanted to emulate the protagonist in ways entirely unrelated to his merits. Not just unrelated but antithetical: cargo-culting the flaws of competent/great/interesting people is actively harmful to my goals! Why would I do this!? The pattern generalized, so I wrote a rant against myself, then figured it'd be good for LessWrong, and posted it here with minimal edits. 

I think the post is crude and messily written, but does the job. 

Meta comment: I notice I'm surprised that out ... (read more)

Neil 92

I think you're right, but I rarely hear this take. Probably because "good at both coding and LLMs" is a light tail end of the distribution, and most of the relative value of LLMs in code is located at the other, much heavier end of "not good at coding" or even "good at neither coding nor LLMs". 

(Speaking as someone who didn't even code until LLMs made it trivially easy, I probably got more relative value than even you.) 

Neil 10

need any help on post drafts? whatever we can do to reduce those trivial inconveniences 

Neil 32

I'm very pro- this kind of post. Whatever this is, I think it's important for ensuring LW doesn't get "frozen" in a state where specific objects are given higher respect than systems. Strong upvoted.

Neil 60

I think you could get a lot out of adding a temporary golden dollar sign with amount donated next to our LW names! Upon proof of donation receipt or whatever.

Seems like the lowest hanging fruit for monetizing vanity— benches being usually somewhat of a last resort!

(The benches seem still underpriced to me, given expected amount raised and average donation size in the foreseeable future).

Neil 10

I've been at Sciences Po for a few months now. Do you have any general advice? I seem to have trouble taking the subjects seriously enough to any real effort in them, which you seem to point out as a failure mode you skirted. Asking as many people I can for this, as I'm going through a minor existential crisis. Thanks!

Neil 83

Yeah that'd go into some "aesthetic flaws" category which presumably has no risk of messing with your rationality. I agree these exist. And I too am picky.

Neil 52

I agree about the punchline. Chef's kiss post

Good list!

I personally really like Scott Alexander's Presidential Platform, it hits the hilarious-but-also-almost-works spot so perfectly. He also has many Bay Area house party stories in addition to the one you link (you can find a bunch (all?) linked at the top of this post). He also has this one from a long time ago, which has one of the best punchlines I've read.

Neil 10

Can I piggy-back off your conclusions so far? Any news you find okay?

3AnthonyC
The first few that came to mind have, it turns out, already retired since I last talked to them. The next few are basically all bloggers with a tighter focus that I learned about either here on LW or through recommendations that ultimately chain back to SSC/ASX.  There are a lot of good sources of data in the world, and very few good sources of analysis, and those that exist have very little relationship to popularity or price or prestige. Beyond that, it really is "buyer beware," and learning to know your own limited and improve your own speed at sorting the nonsense and spotting the bad assumptions and wrong inferences. That's why I'm not being more specific - without knowing your own habits of thought, it's hard to guess whose habitual mistakes and quirks will be transparent to you, and which will mislead you or just not paint the intended picture for you. Probably my advice is, if something you read seems worth understanding, try to spot and discount (what Zvi calls) the Obvious Nonsense. Back in 2008 my aunt sent me a NYT article that (seemingly) claimed having granite countertops was like smoking a pack of cigarettes a day. Obvious nonsense. It was basically pretending the cancer risk from cigarettes was radiological and not chemical, but unless you knew enough about biology and physics it was easy to miss. I was primed for that, because I was a physics undergrad and because I'd recently been reading about how coal plants release more radiation than nuclear plants even if you ignore all other chemical pollution. Eventually this kind of thinking became habitual for me, and it became easier to learn useful things from inadequate sources.  It's also been educational to go through a few healthcare situations where you need to find doctors that are Actually Good and not just cargo-cult-style going through the motions. There's a vibe, a style, that just shines through regardless of specific topic, related to curiosity and excitement about something new and diff
Neil 10

Well then, I can update a little more in the direction not to trust this stuff.

3AnthonyC
To be fair, there are exceptions, and some reporters or publications consistently do better at oarticular kinds of reporting. It just takea a lot of work to reliably figure out which are which.
Neil 10

Ah right, the decades part--I had written about the 1930 revolution, commune, and bourbon destitution, then checked the dates online and stupidly thought "ah, it must be just 1815 then" and only talked about that. Thanks

Neil 10

"second" laughcries in french

Neil 5-2

Ahem, as one of LW's few resident Frenchmen, I must interpose to say that yes, this was not the Big Famous Guillotine French revolution everyone talks about, but one of the ~ 2,456^2 other revolutions that went on in our otherwise very calm history. 

Specifically, we refer to the Les Mis revolution as "Les barricades" mostly because the people of Paris stuck barricades everywhere and fought against authority because they didn't like the king the other powers of Europe put into place after Napoleon's defeat. They failed that time, but succeeded 15 years... (read more)

7David Matolcsi
That's not quite true. Les Mis starts in 1815, but the book spans decades and the revolution is in 1832, a short-lived uprising against the king who got in power two years before, in the 1830 revolution against the dynasty the other European powers restored after Napoleon's defeat in 1815.
Neil 90

Do we know what side we're on? Because I opted in and don't know whether I'm East or West, it just feels Wrong. I guess I stand a non-trivial chance of losing 50 karma ahem please think of the daisy girl and also my precious internet points.

Neil 30

Anti-moderative action will be taken in response if you stand in the way of justice, perhaps by contacting those hackers and giving them creative ideas. Be forewarned.

Neil 10

Fun fact: it's thanks to Lucie that I ended up stumbling onto PauseAI in the first place. Small world + thanks Lucie.

Neil 71

Update everyone: the hard right did not end up gaining a parliamentary majority, which, as Lucie mentioned, could have been the worse outcome wrt AI safety.

Looking ahead, it seems that France will end up being fairly confused and gridlocked as it becomes forced to deal with an evenly-split parliament by playing German-style coalition negociation games. Not sure what that means for AI, except that unilateral action is harder.

For reference, I'm an ex-high school student who just got to vote for the first 3 times in his life because of French political turmoi... (read more)

1Neil
Fun fact: it's thanks to Lucie that I ended up stumbling onto PauseAI in the first place. Small world + thanks Lucie.
Neil 10

I'm working on a non-trivial.org project meant to assess the risk of genome sequences by comparing them to a public list of the most dangerous pathogens we know of. This would be used to assess the risk from both experimental results in e.g. BSL-4 labs and the output of e.g. protein folding models. The benchmarking would be carried out by an in-house ML model of ours. Two questions to LessWrong: 

1. Is there any other project of this kind out there? Do BSL-4 labs/AlphaFold already have models for this? 

2. "Training a model on the most dangerous pa... (read more)

Neil 10

I'm taking this post down, it was to set up an archive.org link as requested by Bostrom, and no longer serves that purpose. Sorry, this was meant to be discreet.

Neil 40

Poetry and practicality

I was staring up at the moon a few days ago and thought about how deeply I loved my family, and wished to one day start my own (I'm just over 18 now). It was a nice moment.

Then, I whipped out my laptop and felt constrained to get back to work; i.e. read papers for my AI governance course, write up LW posts, and trade emails with EA France. (These I believe to be my best shots at increasing everyone's odds of survival).

It felt almost like sacrilege to wrench myself away from the moon and my wonder. Like I was ruining a moment of poetr... (read more)

Neil 54

Too obvious imo, though I didn't downnvote. This also might not be an actual rationalist failure mode; in my experience at least, rationalists have about the same intuition all the other humans have about when something should be taken literally or not.

As for why the comment section has gone berserk, no idea, but it's hilarious and we can all use some fun.

Neil 2-13

Can we have a black banner for the FHI? Not a person, still seems appropriate imo.

Neil 86

FHI at Oxford
by Nick Bostrom (recently turned into song):

the big creaky wheel
a thousand years to turn

thousand meetings, thousand emails, thousand rules
to keep things from changing
and heaven forbid
the setting of a precedent

yet in this magisterial inefficiency
there are spaces and hiding places
for fragile weeds to bloom
and maybe bear some singular fruit

like the FHI, a misfit prodigy
daytime a tweedy don
at dark a superhero
flying off into the night
cape a-fluttering
to intercept villains and stop catastrophes

and why not base it here?
our spandex costumes
blend in wi... (read more)

Neil 10

I've come to think that isn't actually the case. E.g. while I disagree with Being nicer than clippy, it quite precisely nails how consequentialism isn't essentially flawless:

I haven't read that post, but I broadly agree with the excerpt. On green did a good job imo in showing how weirdly imprecise optimal human values are. 

It's true that when you stare at something with enough focus, it often loses that bit of "sacredness" which I attribute to green. As in, you might zoom in enough on the human emotion of love and discover that it's just an endless ti... (read more)

Neil 0-2

Interesting! Seems like you put a lot of effort into that 9,000-word post. May I suggest you publish it in little chunks instead of one giant post? You only got 3 karma for it, so I assume that those who started reading it didn't find it worth the effort to read the whole thing. The problem is, that's not useful feedback for you, because you don't know which of those 9,000 words are presumably wrong. If I were building a version of utilitarianism, I would publish it in little bursts of 2-minute posts. You could do that right now with a single section of your original post. Clearly you have tons of ideas. Good luck! 

Neil 21

You know, I considered "Bob embezzled the funds to buy malaria nets" because I KNEW someone in the comments would complain about the orphanage. Please don't change. 

Actually, the orphanage being a cached thought is precisely why I used it. The writer-pov lesson that comes with "don't fight the hypothetical" is "don't make your hypothetical needlessly distracting". But maybe I miscalculated and malaria nets would be less distracting to LWers. 

Anyway, I'm of course not endorsing fund-embezzling, and I think Bob is stupid. You're right in that failu... (read more)

8tailcalled
I think your position here is approximately-optimal within the framework of consequentialism. It's just that I worry that consequentialism itself is the reason we have problems like AI x-risk, in the sense that the thing that drives x-risk scenarios may be the theory of agency that is shared with consequentialism. I've been working on a post - actually I'm going to temporarily add you as a co-author so you can see the draft and add comments if you're interested - where I discuss the flaws and how I think one should approach it differently. One of the major inspirations is Against responsibility, but I've sort of taken inspiration from multiple places, including critics of EA and critics of economics. I've come to think that isn't actually the case. E.g. while I disagree with Being nicer than clippy, it quite precisely nails how consequentialism isn't essentially flawless: Unbounded utility maximization aspires to optimize the entire world. This is pretty funky for just about any optimization criterion people can come up with, even if people are perfectly flawless in how well they follow it. There's a bunch of attempts to patch this, but none have really worked so far, and it doesn't seem like any will ever work.
3faul_sname
Fixed, thanks
Neil 30

Re: sociology. I found a meme you might enjoy, which would certainly drive your teacher through the roof: https://twitter.com/captgouda24/status/1777013044976980114 

Neil 10

Yeah, that's an excellent idea. I often spot typos in posts, but refrain from writing a comment unless I collect like three. Thanks for sharing!

Neil 6356

A functionality I'd like to see on LessWrong: the ability to give quick feedback for a post in the same way you can react to comments (click for image). When you strong-upvote or strong-downvote a post, a little popup menu appears offering you some basic feedback options. The feedback is private and can only be seen by the author. 

I've often found myself drowning in downvotes or upvotes without knowing why. Karma is a one-dimensional measure, and writing public comments is a trivial inconvience: this is an attempt at middle ground, and I expect it to ... (read more)

6papetoast
Relatedly, in-line private feedback. I saw a really good design for alerting typos here.
Yoav Ravid*1912

I suggested something similar a few months back as a requirement for casting strong votes.

Strong upvote, but I won't tell you why.

Load More