LESSWRONG
LW

869
Ben Pace
36339Ω10882784738201
Message
Dialogue
Subscribe

I'm an admin of LessWrong. Here are a few things about me.

  • I generally feel more hopeful about a situation when I understand it better.
  • I have signed no contracts nor made any agreements whose existence I cannot mention.
  • I believe it is good take responsibility for accurately and honestly informing people of what you believe in all conversations; and also good to cultivate an active recklessness for the social consequences of doing so.
  • It is wrong to directly cause the end of the world. Even if you are fatalistic about what is going to happen.

Randomly: If you ever want to talk to me about anything you like for an hour, I am happy to be paid $1k for an hour of doing that.

(Longer bio.)

Sequences

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
AI Alignment Writing Day 2019
Transcript of Eric Weinstein / Peter Thiel Conversation
AI Alignment Writing Day 2018
Share Models, Not Beliefs
23Benito's Shortform Feed
Ω
7y
Ω
333
Musings from a Lawyer turned AI Safety researcher (ShortForm)
Ben Pace10h143

I thought this meant she endorsed the title as her beliefs, so in case anyone else was confused, she has endorsed the book as worth reading, but not (to my knowledge) endorsed its thesis as "what I believe".

Here's what she writes:

Although, in general, I disagree with catastrophic framings of AI risk (which have been exploited by AI CEOs to increase interest in their products, as I recently wrote in my newsletter), the AI safety debate is an important one, and it concerns all of us.

There are differing opinions on the current path of AI development and its possible futures. There are also various gray zones and unanswered questions on possible ways to mitigate risk and avoid harm.

Yudkowsky has been researching AI alignment for over 20 years, and together with Soares, he has built a strong argument for why AI safety concerns are urgent and why action is needed now. Whether you agree with their tone or not, their book is worth reading.

Seems that she finds Eliezer+Nate credible in their concerns because they are not AI company CEOs but have been working on the problem for 2 decades.

Reply1
The Relationship Between Social Punishment and Shared Maps
Ben Pace1d62

I admit I am a bit confused about the thesis here... I get that accurate behavioral accounting is sometimes tightly related to social punishment such that the attempt to give or defend oneself from punishment provides incentive to lie about the behavior (and attempts to describe the behavior have direct implications for punishment). 

But are you further claiming that that all social punishment is identical[1] to truth-claims about other things (i.e. "reasons for the punishment")? This seems like an ideal that I aspire to, but not how most people relate to social punishment, where social ostracism can sometimes simply be a matter of fashion or personal preference. 

Personally I use phrases like "X is lame" or "X isn't cool" to intentionally and explicitly set the status of things. I endeavor to always have good reasons for why and to provide them (or at least to have them ready if requested), but the move itself does not require justification in order to successfully communicate that something is having its status lowered or is something that I oppose. People would often happily just accept the status-claims without reasons, similar to learning what is currently 'in fashion'.

  1. ^

    On reflection I don't quite mean identical to, but something more like "Is a deterministic function of truth-claims about good/bad behavior, taking that-and-only-that as input".

Reply
Buck's Shortform
Ben Pace2d60

I think of "It seems" as coming from the view from nowhere. I say "It seems to me" to own it, unless I mean to make the active claim that I believe it will seem this way to everyone.

"Arguably" feels super weak to me. I would only use it if I actively wanted to distance myself from a view. Almost anything is 'arguable' in principle.

Reply
Buck's Shortform
Ben Pace3d3019

Probably this isn't the exclusive reason, but typically I use "I think" whenever I want to rule out the interpretation that I am implying we all agree on my claim. If I say "It was a mistake for you to paint this room yellow" this is more natural if you agree with me; if I say "I think it was a mistake for you to paint this room yellow" this is more natural if I'm informing you of my opinion but I expect you to disagree.

This is not a universal rule, and fwiw I do think there's something good about clear and simple writing that cuts out all the probably-unnecessary qualifiers, but I think this is a common case where I find it worth adding it in.

Reply63
Cole Wyeth's Shortform
Ben Pace5d40

Datapoint: I'm currently setting up a recording studio at Lighthaven, and I am using them all the time to get guides for things like "how to change a setting on this camera" or "how to use this microphone" or "how to use this recording software".

Yes, they confabulate menus and things a lot, but as long as I keep uploading photos of what I actually see, they know the basics much better than me (e.g. what bit rate to set the video vs the audio, where to look to kill the random white noise input I'm getting, etc).

I'd say they confabulate like 50% of the time but that they're still a much more effective search engine for me than google, and can read the manual much faster than me. My guess is I simply couldn't do some of the projects I'm doing without them.

Reply1
Benito's Shortform Feed
Ben Pace6d20

Awesome, thanks so much gustaf! 

I've re-arranged the order a little bit to have some of the more common ones first.

Yeah, I have been thinking about linking it from the reacts menu. Will think about that next time I ship some changes. (I'll be more inclined to do that if there are more explanations and links and culture and stuff.)

Reply11
Notes on fatalities from AI takeover
Ben Pace7d40

Curated. This is a topic that is getting a ton of attention at the minute, and the post does a decent job at laying out relevant considerations and summarizing some prior debate.

(I disagree with the conclusion but it seems positive to me to curate something with a different conclusion than the recent book on the issue.)

Reply
Notes on fatalities from AI takeover
Ben Pace7d90

Link to Nate's comments.

Reply11
Mikhail Samin's Shortform
Ben Pace9d20

Further follow-up: Guido Reichstadter wraps up after 30 days. Impressively long! And a bit longer than I'd guessed.

Reply
A Reply to MacAskill on "If Anyone Builds It, Everyone Dies"
[+]Ben Pace9d-1417
Load More
134The Inkhaven Residency
2mo
35
37LessOnline 2025: Early Bird Tickets On Sale
7mo
5
20Open Thread Spring 2025
7mo
50
281Arbital has been imported to LessWrong
8mo
30
141The Failed Strategy of Artificial Intelligence Doomers
8mo
77
109Thread for Sense-Making on Recent Murders and How to Sanely Respond
8mo
146
83What are the good rationality films?
Q
11mo
Q
54
942024 Petrov Day Retrospective
1y
25
136[Completed] The 2024 Petrov Day Scenario
1y
114
55Thiel on AI & Racing with China
1y
10
Load More
LessWrong Reacts
6 days ago
LessWrong Reacts
6 days ago
LessWrong Reacts
6 days ago
LessWrong Reacts
6 days ago
(+3354/-3236)
LessWrong Reacts
15 days ago
LessWrong Reacts
16 days ago
(+638/-6)
LessWrong Reacts
16 days ago
(+92)
LessWrong Reacts
18 days ago
(+248)
Adversarial Collaboration (Dispute Protocol)
9 months ago