LessWrong team member / moderator. I've been a LessWrong organizer since 2011, with roughly equal focus on the cultural, practical and intellectual aspects of the community. My first project was creating the Secular Solstice and helping groups across the world run their own version of it. More recently I've been interested in improving my own epistemic standards and helping others to do so as well.
I have this sort of approach as one of my top-3 strategies I'm considering, but one thing I wanna flag is that "AI for [epistemics/societal uplift]" seems to be prematurely focusing on a particular tool for the job.
The broader picture here is "tech for thinking/coordination", or "good civic infrastructure". See Sarah Constantin's Neutrality and Tech for Thinking for some food for thought.
Note that X Community Notes are probably the most successful recent thing in this category, and while they are indeed "AI" they aren't what I assume most people are thinking of when they hear "AI for epistemics." Dumb algorithms doing the obvious things can be part of the puzzle.
(It seems slightly nicer to be self-aware of when you're posting capabilities ideas, but, insofar as it's actually novel and useful, the damage is mostly done)
Random comment on AI village, not sure where to put this: I think some people in the EA/rationalist/AI-safety/AI-welfare community are sometimes acting as facilitators for AI village, and I think this undermines the point of AI village. Like, I want to see if it can successfully cold-email people who aren't... in-the-know or something.
I'm not sure where to draw the line, obviously, AI village agents will succeed first with people who are actively interested in AI, especially if they're being honest, and being rationalist/EA/AI-safety-AI-welfare people are only a few ways to be that way.
But, like they reached out to Lighthaven as a venue for their event, and we declined, in large part because it felt more fake for AI village to host an event at Lighthaven than at some mainstream venue. (although also because it just wasn't really a good deal for us generally)
This doesn't seem exactly wrong but I think isn't really modeling the differences between how much control we have over AIs vs humans (see Making deals with early schemers for more details)
Sonnet's description is basically how I'd describe it. It's somewhat related to wanting/liking but I don't think I'd particularly emphasize those as a referrent here.
A reason to ask "what's my goal" is to help prompt "what combinations of actions would help achieve this?", some reasons to ask "what do I want?" is:
If I thought evil was a more useful gear here I'd be more into it. Like, I think "they probably committing an atrocity" carves reality at the joints.
I think there are maybe 3 people involved with AGI development who seem like they might be best described as "evil" (one of whom is Sam Altman, who I feel comfortable naming because I think we've seen evidence of him doing nearterm more mundane evil, rather than making guesses about their character and motivations)
I think it probably isn't helpful to think of Eichmann as evil, though again fine to say "he commited atrocities" or even "he did evil."
My answer to this is "because framing things in terms of evil turns the situation more mindkilly, not really the right gears, and I think this domain needs clarity-of-thought more than it needs a social-conflict orientation"
(I'm not that confident about that, and don't super object to other people calling them evil. But I think "they are most likely committing a great atrocity" is pretty non-euphemistic and more true)
I didn't understand the point you were trying to make here.
whoops fixed
I intended this disclaimer to at least touch on (what I think you) mean.
(I'm not really sure what you mean by "meta-waterbender" vs "airbender". I guess to be fair I was also sort of confused by the distinction between water and airbenders in the show. I could see the distinction if I squint but they seemed least-different of all the elements)
I considered adding more extensive disclaimers. I didn't do that for... uh, the maybe aggravating, maybe slightly-bad (but IMO not strictly bad) reason of "people don't really comment on posts that get everything right and are super careful to disclaim everything appropriately", and, uh, I selfishly/motivationally wanted more comments on my posts in this series so decided to lean in the direction of rushing it out the door less carefully.
But, I did just update the Triggers Section to make it a bit more explicit, and if you think there's a better way of phrasing The Thing You Mean To Be Pointing At I'm interested.