This website requires javascript to properly function. Consider activating javascript to get access to all site functionality.
LESSWRONG
Wikitags
LW
Login
Research Agendas
Settings
Applied to
Give Neo a Chance
by
ank
12d
ago
Applied to
Share AI Safety Ideas: Both Crazy and Not
by
ank
16d
ago
Applied to
The Theoretical Reward Learning Research Agenda: Introduction and Motivation
by
Joar Skalse
17d
ago
Applied to
How to Contribute to Theoretical Reward Learning Research
by
Joar Skalse
17d
ago
Applied to
Unaligned AGI & Brief History of Inequality
by
ank
23d
ago
Applied to
Intelligence–Agency Equivalence ≈ Mass–Energy Equivalence: On Static Nature of Intelligence & Physicalization of Ethics
by
ank
24d
ago
Applied to
Notes on notes on virtues
by
David Gross
25d
ago
Applied to
Human-AI Relationality is Already Here
by
bridgebot
26d
ago
Applied to
Rational Utopia & Narrow Way There: Multiversal AI Alignment, Non-Agentic Static Place AI, New Ethics... (V. 4)
by
ank
1mo
ago
Applied to
The Road to Evil Is Paved with Good Objectives: Framework to Classify and Fix Misalignments.
by
Shivam
2mo
ago
Applied to
False Positives in Entity-Level Hallucination Detection: A Technical Challenge
by
MaxKamachee
2mo
ago
Applied to
You should delay engineering-heavy research in light of R&D automation
by
Daniel Paleka
2mo
ago
Applied to
My AGI safety research—2024 review, ’25 plans
by
Steven Byrnes
3mo
ago
Applied to
Shallow review of technical AI safety, 2024
by
Dakara
3mo
ago
Applied to
Shallow review of live agendas in alignment & safety
by
Dakara
3mo
ago
Applied to
Retrospective: PIBBSS Fellowship 2024
by
DusanDNesic
3mo
ago
Applied to
Agency overhang as a proxy for Sharp left turn
by
Eris
4mo
ago
Applied to
Seeking Collaborators
by
Steven Byrnes
4mo
ago
Applied to
Self-prediction acts as an emergent regularizer
by
Cameron Berg
5mo
ago
Applied to
NAO Updates, Fall 2024
by
ProgramCrafter
5mo
ago