LESSWRONG
LW

653
Zach Stein-Perlman
10655Ω3928667912
Message
Dialogue
Subscribe

AI strategy & governance. ailabwatch.org. ailabwatch.substack.com. 

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
4Zach Stein-Perlman's Shortform
Ω
4y
Ω
299
kave's Shortform
Zach Stein-Perlman4d20

I observe that https://www.lesswrong.com/posts/BqwXYFtpetFxqkxip/mikhail-samin-s-shortform?commentId=dtmeRXPYkqfDGpaBj isn't frontpage-y but remains on the homepage even after many mods have seen it. This suggests that the mods were just patching the hack. (But I don't know what other shortforms they've hidden, besides the political ones, if any.)

Reply
Mikhail Samin's Shortform
Zach Stein-Perlman4d90

fwiw I agree with most but not all details, and I agree that Anthropic's commitments and policy advocacy have a bad track record, but I think that Anthropic capabilities is nevertheless net positive, because Anthropic has way more capacity and propensity to do safety stuff than other frontier AI companies.

I wonder what you believe about Anthropic's likelihood of noticing risks from misalignment relative to other companies, or of someday spending >25% of internal compute on (automated) safety work.

Reply1
Zach Stein-Perlman's Shortform
Zach Stein-Perlman8d455

I think "Overton window" is a pretty load-bearing concept for many LW users and AI people — it's their main model of policy change. Unfortunately there's lots of other models of policy change. I don't think "Overton window" is particularly helpful or likely-to-cause-you-to-notice-relevant-stuff-and-make-accurate-predictions. (And separately people around here sometimes incorrectly use "expand the Overton window" to just mean with "advance AI safety ideas in government.") I don't have time to write this up; maybe someone else should (or maybe there already exists a good intro to the study of why some policies happen and persist while others don't[1]).

Some terms: policy windows (and "multiple streams"), punctuated equilibrium, policy entrepreneurs, path dependence and feedback (yes this is a real concept in political science, e.g. policies that cause interest groups to depend on them are less likely to be reversed), gradual institutional change, framing/narrative/agenda-setting.

Related point: https://forum.effectivealtruism.org/posts/SrNDFF28xKakMukvz/tlevin-s-quick-takes?commentId=aGSpWHBKWAaFzubba.

  1. ^

    I liked the book Policy Paradox in college. (Example claim: perceived policy problems are strategically constructed through political processes; how issues are framed—e.g. individual vs collective responsibility—determines which solutions seem appropriate.) I asked Claude for suggestions on a shorter intro and I didn't find the suggestions helpful.

    I guess I think if you work on government stuff and you [don't have poli sci background / aren't familiar with concepts like "multiple streams"] you should read Policy Paradox (although the book isn't about that particular concept).

Reply
kave's Shortform
Zach Stein-Perlman10d20

I guess I'll write non-frontpage-y quick takes as posts instead then :(

Reply1
kave's Shortform
Zach Stein-Perlman11d229

I'd like to be able to see such quick takes on the homepage, like how I can see personal blogposts on the homepage (even though logged-out users can't).

Are you hiding them from everyone? Can I opt into seeing them?

Reply
LW Reacts pack for Discord/Slack/etc
Zach Stein-Perlman11d20

I failed to find a way to import to Slack without doing it one by one.

Reply1
Consider donating to Alex Bores, author of the RAISE Act
Zach Stein-Perlman11d20

Bores knows, at least for people who donate via certain links. For example, the link in this post is https://secure.actblue.com/donate/boresai?refcode=lw rather than https://secure.actblue.com/donate/boresweb.

Reply
Which side of the AI safety community are you in?
Zach Stein-Perlman12d5418

I'm annoyed that Tegmark and others don't seem to understand my position: you should try for great global coordination but also invest in safety in more rushed worlds, and a relatively responsible developer shouldn't unilaterally stop.

(I'm also annoyed by this post's framing for reasons similar to Ray.)

Reply
Zach Stein-Perlman's Shortform
Zach Stein-Perlman12d*30

Part is thinking about donation opportunities, like Bores. Hopefully I'll have more to say publicly at some point!

Reply
Zach Stein-Perlman's Shortform
Zach Stein-Perlman12d505

Recently I've been spending much less than half of my time on projects like AI Lab Watch. Instead I've been thinking about projects in the "strategy/meta" and "politics" domains. I'm not sure what I'll work on in the future but sometimes people incorrectly assume I'm on top of lab-watching stuff; I want people to know I'm not owning the lab-watching ball. I think lab-watching work is better than AI-governance-think-tank work for the right people on current margins and at least one more person should do it full-time; DM me if you're interested.

Reply4
Load More
44AI companies' policy advocacy (Sep 2025)
1mo
0
104xAI's new safety framework is dreadful
2mo
6
52AI companies have started saying safeguards are load-bearing
Ω
2mo
Ω
2
15ChatGPT Agent: evals and safeguards
3mo
0
33Epoch: What is Epoch?
4mo
1
15AI companies aren't planning to secure critical model weights
4mo
0
207AI companies' eval reports mostly don't support their claims
Ω
5mo
Ω
13
58New website analyzing AI companies' model evals
5mo
0
72New scorecard evaluating AI companies on safety
5mo
8
71Claude 4
5mo
24
Load More
Ontology
2 years ago
(+45)
Ontology
2 years ago
(-5)
Ontology
2 years ago
Ontology
2 years ago
(+64/-64)
Ontology
2 years ago
(+45/-12)
Ontology
2 years ago
(+64)
Ontology
2 years ago
(+66/-8)
Ontology
2 years ago
(+117/-23)
Ontology
2 years ago
(+58/-21)
Ontology
2 years ago
(+41)
Load More