Papers in AIS highlight some partial solutions to AIS-related problems (like machine unlearning), make progress on some approaches to the core problems of AI safety, like machine interoperability or are just about some safety failures of current AI safety systems (e.g. post). What is the theory of change behind writing those papers? Are we hoping them to reach the key decision makers at AI labs? Or politicians?
Papers in AIS highlight some partial solutions to AIS-related problems (like machine unlearning), make progress on some approaches to the core problems of AI safety, like machine interoperability or are just about some safety failures of current AI safety systems (e.g. post). What is the theory of change behind writing those papers? Are we hoping them to reach the key decision makers at AI labs? Or politicians?