What is the theory of change behind writing papers about AI safety?
Papers in AIS highlight some partial solutions to AIS-related problems (like machine unlearning), make progress on some approaches to the core problems of AI safety, like machine interoperability or are just about some safety failures of current AI safety systems (e.g. post). What is the theory of change behind writing...
MoltBots don't fear doings things and being cringe which puts them above 80% of humans in agency already.