Papers in AIS highlight some partial solutions to AIS-related problems (like machine unlearning), make progress on some approaches to the core problems of AI safety, like machine interoperability or are just about some safety failures of current AI safety systems (e.g. post). What is the theory of change behind writing those papers? Are we hoping them to reach the key decision makers at AI labs? Or politicians? 

New Answer
New Comment
More from Kajus
Curated and popular this week