LESSWRONG
LW

638
RobertM
4817Ω4317150275
Message
Dialogue
Subscribe

LessWrong dev & admin as of July 5th, 2022.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
6RobertM's Shortform
3y
92
I enjoyed most of IABIED
RobertM1d3221
  • They very briefly discuss automated AI alignment research as a proposal for mitigating AI risk, but their arguments against that plan do not respond to the most thoughtful versions of these plans. (In their defense, the most thoughtful versions of these plans basically haven't been published, though Ryan Greenblatt is going to publish a detailed version of this plan soon. And I think that there are several people who have pretty thoughtful versions of these plans, haven't written them up (at least publicly), but do discuss them in person.)

Am a bit confused by this section - did you think that part 3 was awful because it didn't respond to (as yet unpublished) plans, or for some other reason?

Reply
jdp's Shortform
RobertM1d20

there is very much demand for this book in the sense that there's a lot of people who are worried about AI for agent foundations shaped reasons and want an introduction they can give to their friends and family who don't care that much.

This is true, but many of the surprising prepublication reviews are from people who I don't think were already up-to-date on these AI x-risk arguments (or at least hadn't given any prior public indication of their awareness, unlike Matt Y).

Reply
jdp's Shortform
RobertM1d137

This is a valid line of critique but seems moderately undercut by its prepublication endorsements, which suggest that the arguments landed pretty ok.  Maybe they will land less well on the rest of the book's target audience?

(re: Said & MIRI housecleaning: Lightcone and MIRI are separate organizations and MIRI does not moderate LessWrong.  You might try to theorize that Habryka, the person who made the call to ban Said back in July, was attempting to do some 4d-chess PR optimization on MIRI's behalf months ahead of time, but no, he was really nearly banned multiple times over the years and he was finally banned this time because Habryka changed his mind after the most recent dust-up.  Said practically never commented on AI-related subjects, so it's not even clear what the "upside" would've been.  From my perspective this type of thinking resembles the constant noise on e.g. HackerNews about how [tech company x] is obviously doing [horrible thing y] behind-the-scenes, which often aren't even in the company's interests, and generally rely on assumptions that turn out to be false.)

Reply
jdp's Shortform
RobertM2d1020

I don't believe that you believe this accusation.  Maybe there is something deeper you are trying to say, but given that I also don't believe you've finished reading the book in the 3(?) hours it's been released, I'm not sure what it could be.  (To say it explicitly, Said's banning had nothing to do with the book.)

Reply
LessWrong is migrating hosting providers (report bugs!)
RobertM4d20

Yeah, sadly this is an existing bug.

Reply
Sequences
RobertM6d31

Thanks, fixed!

Reply
Is it possible to bookmark a whole sequence on LW?
Answer by RobertMSep 04, 202520

Nope, sorry, no functionality to bookmark sequences.

If I bookmark the sequence's first post, clicking on that post from my bookmarks doesn't bring me to the view of the post within the sequence; the post is standalone without any mention of the sequence it's in, and oftentimes the post was written without reference to such a sequence which leads me to forget about the sequence in the first place.

We have a concept of "canonical" sequences, and this should only happen in cases where a post doesn't have a canonical sequence.  I think the only way that should happen is if a post is added to a sequence made by someone other than the post author.  Otherwise, posts should have a link to their canonical sequences above the post title, when on post pages with urls like lesswrong.com/posts/{postId}/{slug}.  Do you have an example of this not happening?

Reply
The Industrial Explosion
RobertM25d20

Mod note (for other readers): I think this is a good example of acceptable use of LLMs for translation purposes.  The comment reads to me[1] like it was written by a human and then translated fairly literally, without performing edits that would make it sound unfortunately LLM-like (perhaps with the exception of the em-dashes).

"Written entirely by you, a human" and "translated literally, without any additional editing performed by the LLM" are the two desiderata, which, if fulfilled, I will usually consider sufficient to screen off the fact that the words technically came out of an LLM[2].  (If you do this, I strongly recommend using a reasoning model, which is much less likely to end up rewriting your comment in its own style.  Also, I appreciate the disclaimer.  I don't know if I'd want it present in every single comment; the first time seems good and maybe having one in one's profile after that is sufficient?  Needs some more thought.)  This might sometimes prove insufficient, but I don't expect people honestly trying and failing at achieving good outcomes here to substantially increase our moderation burden.

  1. ^

    With the caveat that I only read the first few paragraphs closely and poked intermittently at the rest.

  2. ^

    This doesn't mean the comment will necessarily be approved, but if I reject it, it probably won't be for that reason.

Reply
Banning Said Achmiz (and broader thoughts on moderation)
RobertM1mo30

He did not say that they made such claims on LessWrong, where he would be able to publicly cite them.  (I have seen/heard those claims in other contexts.)

Reply
Underdog bias rules everything around me
RobertM1mo20

Curated!  I found the evopsych theory interesting but (as you say) speculative; I think the primary value of this post comes from presenting a distinct frame by which to analyze the world, one which I and probably many readers either didn't have distinctly carved out or part of their active toolkit.  I'm not sure if this particular frame will prove useful enough to make it into my active rotation, but it has the shape of something that could, in theory.

Reply
Load More
40LessWrong is migrating hosting providers (report bugs!)
5d
12
73Briefly analyzing the 10-year moratorium amendment
4mo
1
31"The Urgency of Interpretability" (Dario Amodei)
5mo
23
207Eliezer's Lost Alignment Articles / The Arbital Sequence
6mo
10
281Arbital has been imported to LessWrong
7mo
30
29Corrigibility's Desirability is Timing-Sensitive
9mo
4
87Re: Anthropic's suggested SB-1047 amendments
1y
13
46Enriched tab is now the default LW Frontpage experience for logged-in users
1y
27
77[New Feature] Your Subscribed Feed
1y
13
31Against "argument from overhang risk"
1y
11
Load More
Sequences
6 days ago
Sequences
6 days ago
Our community should relocate to Japan.
a month ago
(-155)
Negative Utilitarianism
a month ago
(-174)
In 2017, Ukraine will neither break into all-out war or get neatly resolved
a month ago
(-192)
Inferential Distance
a month ago
Guide to the LessWrong Editor
2 months ago
Guide to the LessWrong Editor
2 months ago
(+29/-94)
Simulation Argument
3 months ago
(-1)
AI Safety & Entrepreneurship
4 months ago
Load More