Two ideas for alignment, perpetual mutual distrust and induction
Two ideas I have for alignment (may exist already or may not be great, I am not exhaustively read on the topic) Idea 1, Two Agents in mutual distrust of each other: Intuitively, alignment is a difficult problem because it is hard to know what an AI ostensibly less capable...
May 25, 20231