I have to look for a while before finding any non-AI posts. Seems LW is mainly an AI / alignment discussion forum at this point.
It seems more informative to just look at top (inflation adjusted) karma for 2022 (similar to what habryka noted in the sibling). AI posts in bold.
I count 18/29 about AI. A few AI posts are technically more general. A few non-AI posts seem to indirectly be about AI.
I think the AI posts are definitely substantially more interlinked than the non-AI posts, so I think specific metric oversamples AI posts.
I just updated the spreadsheet to include All Time posts. Lists of Lethalities is still the winner by Total Pingback Karma, although not by Pingback Count (and this seems at least partially explained by karma inflation)
The list would look pretty different if self-cites were excluded. E.g. my posts would probably all be gone 😂
Yeah if I have time today I'll make an "exclude self-cites" column, although fwiw I think the "total pingback karma" is fairly legit even if including self-cites. If your followup work got a lot of karma, I think that's a useful signal about your original post even if you quote yourself liberally.
For the past couple years I've wished LessWrong had a "sort posts by number of pingbacks, or, ideally, by total karma of pingbacks". I particularly wished for this during the Annual Review, where "which posts got cited the most?" seemed like a useful thing to track for potential hidden gems.
We still haven't built a full-fledged feature for this, but I just ran a query against the database, and made it into a spreadsheet, which you can view here:
LessWrong 2022 Posts by Pingbacks
Here are the top 100 posts, sorted by Total Pingback Karma
Title/Link
Post Karma
Pingback Count
Total Pingback Karma
Avg Pingback Karma
AGI Ruin: A List of Lethalities
MIRI announces new "Death With Dignity" strategy
A central AI alignment problem: capabilities generalization, and the sharp left turn
Simulators
Without specific countermeasures, the easiest path to transformative AI likely leads to AI takeover
Reward is not the optimization target
A Mechanistic Interpretability Analysis of Grokking
How To Go From Interpretability To Alignment: Just Retarget The Search
On how various plans miss the hard bits of the alignment challenge
[Intro to brain-like-AGI safety] 3. Two subsystems: Learning & Steering
How likely is deceptive alignment?
The shard theory of human values
Mysteries of mode collapse
[Intro to brain-like-AGI safety] 2. “Learning from scratch” in the brain
Why Agent Foundations? An Overly Abstract Explanation
A Longlist of Theories of Impact for Interpretability
How might we align transformative AI if it’s developed very soon?
A transparency and interpretability tech tree
Discovering Language Model Behaviors with Model-Written Evaluations
A note about differential technological development
Causal Scrubbing: a method for rigorously testing interpretability hypotheses [Redwood Research]
Supervise Process, not Outcomes
Shard Theory: An Overview
Epistemological Vigilance for Alignment
A shot at the diamond-alignment problem
Where I agree and disagree with Eliezer
Brain Efficiency: Much More than You Wanted to Know
Refine: An Incubator for Conceptual Alignment Research Bets
Externalized reasoning oversight: a research direction for language model alignment
Humans provide an untapped wealth of evidence about alignment
Six Dimensions of Operational Adequacy in AGI Projects
How "Discovering Latent Knowledge in Language Models Without Supervision" Fits Into a Broader Alignment Scheme
Godzilla Strategies
(My understanding of) What Everyone in Technical Alignment is Doing and Why
Two-year update on my personal AI timelines
[Intro to brain-like-AGI safety] 15. Conclusion: Open problems, how to help, AMA
[Intro to brain-like-AGI safety] 6. Big picture of motivation, decision-making, and RL
Human values & biases are inaccessible to the genome
You Are Not Measuring What You Think You Are Measuring
Open Problems in AI X-Risk [PAIS #5]
[Intro to brain-like-AGI safety] 1. What's the problem & Why work on it now?
Conditioning Generative Models
Conjecture: Internal Infohazard Policy
A challenge for AGI organizations, and a challenge for readers
Superintelligent AI is necessary for an amazing future, but far from sufficient
Optimality is the tiger, and agents are its teeth
Let’s think about slowing down AI
Niceness is unnatural
Announcing the Alignment of Complex Systems Research Group
[Intro to brain-like-AGI safety] 13. Symbol grounding & human social instincts
ELK prize results
Abstractions as Redundant Information
[Link] A minimal viable product for alignment
Acceptability Verification: A Research Agenda
What an actually pessimistic containment strategy looks like
Let's See You Write That Corrigibility Tag
chinchilla's wild implications
Worlds Where Iterative Design Fails
why assume AGIs will optimize for fixed goals?
Gradient hacking: definitions and examples
Contra shard theory, in the context of the diamond maximizer problem
We Are Conjecture, A New Alignment Research Startup
Circumventing interpretability: How to defeat mind-readers
Evolution is a bad analogy for AGI: inner alignment
Refining the Sharp Left Turn threat model, part 1: claims and mechanisms
MATS Models
Common misconceptions about OpenAI
Prizes for ELK proposals
Current themes in mechanistic interpretability research
Discovering Agents
[Intro to brain-like-AGI safety] 12. Two paths forward: “Controlled AGI” and “Social-instinct AGI”
What's General-Purpose Search, And Why Might We Expect To See It In Trained ML Systems?
Inner and outer alignment decompose one hard problem into two extremely hard problems
Threat Model Literature Review
Language models seem to be much better than humans at next-token prediction
Will Capabilities Generalise More?
Pivotal outcomes and pivotal processes
Conditioning Generative Models for Alignment
Training goals for large language models
It’s Probably Not Lithium
Latent Adversarial Training
“Pivotal Act” Intentions: Negative Consequences and Fallacious Arguments
Conditioning Generative Models with Restrictions
The alignment problem from a deep learning perspective
Instead of technical research, more people should focus on buying time
By Default, GPTs Think In Plain Sight
[Intro to brain-like-AGI safety] 4. The “short-term predictor”
Don't leave your fingerprints on the future
Strategy For Conditioning Generative Models
Call For Distillers
Thoughts on AGI organizations and capabilities work
Optimization at a Distance
[Intro to brain-like-AGI safety] 5. The “long-term predictor”, and TD learning
What does it take to defend the world against out-of-control AGIs?
Monitoring for deceptive alignment
Late 2021 MIRI Conversations: AMA / Discussion
How to Diversify Conceptual Alignment: the Model Behind Refine
wrapper-minds are the enemy
But is it really in Rome? An investigation of the ROME model editing technique
An Open Agency Architecture for Safe Transformative AI