Intelligence enhancement as existential risk mitigation — LessWrong