This website requires javascript to properly function. Consider activating javascript to get access to all site functionality.
LESSWRONG
is fundraising!
Tags
LW
$
Login
AI Safety Camp
•
Applied to
Nobody Asks the Monkey: Why Human Agency Matters in the AI Age
by
Miloš Borenović
15d
ago
•
Applied to
Formalize the Hashiness Model of AGI Uncontainability
by
Remmelt
1mo
ago
•
Applied to
Agency overhang as a proxy for Sharp left turn
by
Eris
1mo
ago
•
Applied to
AI Safety Camp 10
by
Remmelt
2mo
ago
•
Applied to
Embracing complexity when developing and evaluating AI responsibly
by
Aliya Amirova
2mo
ago
•
Applied to
Invitation to lead a project at AI Safety Camp (Virtual Edition, 2025)
by
Linda Linsefors
4mo
ago
•
Applied to
Some reasons to start a project to stop harmful AI
by
Remmelt
4mo
ago
•
Applied to
Thinking About Propensity Evaluations
by
Maxime Riché
4mo
ago
•
Applied to
A Taxonomy Of AI System Evaluations
by
Maxime Riché
4mo
ago
•
Applied to
Self-Other Overlap: A Neglected Approach to AI Alignment
by
Remmelt
4mo
ago
•
Applied to
Whirlwind Tour of Chain of Thought Literature Relevant to Automating Alignment Research.
by
sevdeawesome
6mo
ago
•
Applied to
Immunization against harmful fine-tuning attacks
by
domenicrosati
6mo
ago
•
Applied to
Training-time domain authorization could be helpful for safety
by
domenicrosati
7mo
ago
•
Applied to
Launching applications for AI Safety Careers Course India 2024
by
Axiom_Futures
8mo
ago
•
Applied to
"Open Source AI" is a lie, but it doesn't have to be
by
jacobhaimes
8mo
ago
•
Applied to
AISC9 has ended and there will be an AISC10
by
Linda Linsefors
8mo
ago
•
Applied to
Towards a formalization of the agent structure problem
by
Alex_Altair
8mo
ago
•
Applied to
[Aspiration-based designs] 1. Informal introduction
by
Jobst Heitzig
9mo
ago
•
Applied to
Podcast interview series featuring Dr. Peter Park
by
jacobhaimes
9mo
ago