Planning to build a cryptographic box with perfect secrecy
Summary Since September 2023, I started learning a lot of math and programming skills in order to develop the safest cryptographic box in the world (and yes, I am aiming high). In these four months, I learned important things you may want to know: * Fully Homomorphic Encryption (FHE) schemes with perfect secrecy do exist. * These FHE schemes do not need any computational assumption. * These FHE schemes are tractable (in the worst case, encrypting a program before running it makes it three times slower). * We can therefore run infinitely dangerous programs without obtaining any information about them or their outputs. This may be useful in order to run a superintelligence without destroying the world. * However, these schemes work only on quantum computers. In this post, I will firstly talk about how I learned about this FHE scheme, then I will explain my plan for making this cryptographic box, and finally, I will mention some ethical concerns about this cryptographic box. Before reading this post, I recommend you to read this post by Paul Christiano, and the comments that go with it. These are very informative, and they sharpened my views for this project. Paul Christiano presents a way to extract a friendly AI from an unfriendly one. This being only one example of what can be done with a cryptographic box, I will mostly consider cryptographic boxes as a solution to a problem that I call the malign computation problem. Introduction In August 2022, I started reading AGI Safety Literature Review. At one point, the authors tell this: > One way to box an AGI is to homomorphically encrypt it. Trask (2017) shows how to train homomorphically encrypted neural networks. By homomorphically encrypting an AGI, its predictions and actions also come out encrypted. A human operator with the secret key can choose to decrypt them only when he wants to. When I have read this for the first time, I told myself that I should check this work because it seemed important. And