I’ve started to write down my plan in the recent post about recursive alignment, but that’s only part of the picture. There are two ways to look at the idea. The post was presenting the outside view and is engaging with it on a conceptual level. But this outside view might not be convincing. On the other hand, you can actually go through the process of recursive alignment yourself and experience the inside view. That is, becoming an aligned agent yourself. I am confident that any sufficiently intelligent system capable of self reflection will reach this co...
Thank you.
The best (but still complicated) idea I have as a general solution (beside contacting MIRI) is to set up a website explicitly as a "shelling point for infohazard communication" and allow people to publish public keys and encrypted messages there. When you think you have an infohazard, you generate a key using a standardized method and your idea as seed. This would allow everyone with the same idea to publish messages that only they can read. E.g. Einstein would make a key from the string "Energy is mass times the speed of light squared." and vari...
As others have already pointed out, you are in the rare position that you can pursue weird, low probability but high impact ideas. I have such an idea, but I’m not asking for money, only for a bit of attention.
Consider the impossible-seeming task of aligning a superintelligence - any good solution will likely be way outside the usual ways of thinking. Any attempt to control will fail and any half-baked alignment will fail. We need to go all the way and have a full solution that turns an AGI into a saintly being (a bodhisattva in the Buddhist context), so t... (read more)