I have, over the last year, become fairly well-known in a small corner of the internet tangentially related to AI.
As a result, I've begun making what I would have previously considered astronomical amounts of money: several hundred thousand dollars per month in personal income.
This has been great, obviously, and the funds have alleviated a fair number of my personal burdens (mostly related to poverty). But aside from that I don't really care much for the money itself.
My long term ambitions have always been to contribute materially to the mitigation of the impending existential AI threat. I never used to have the means to do so, mostly because of more pressing, safety/sustenance concerns, but now that I do, I would like to help however possible.
Some other points about me that may be useful:
- I'm intelligent, socially capable, and exceedingly industrious.
- I have a few hundred thousand followers worldwide across a few distribution channels. My audience is primarily small-midsized business owners. A subset of these people are very high leverage (i.e their actions directly impact the beliefs, actions, or habits of tens of thousands of people).
- My current work does not take much time. I have modest resources (~$2M) and a relatively free schedule. I am also, by all means, very young.
Given the above, I feel there's a reasonable opportunity here for me to help. It would certainly be more grassroots than a well-funded safety lab or one of the many state actors that has sprung up, but probably still sizeable enough to make a fraction of a % of a difference in the way the scales tip (assuming I dedicate my life to it).
What would you do in my shoes, assuming alignment on core virtues like maximizing AI safety?
As others have already pointed out, you are in the rare position that you can pursue weird, low probability but high impact ideas. I have such an idea, but I’m not asking for money, only for a bit of attention.
Consider the impossible-seeming task of aligning a superintelligence - any good solution will likely be way outside the usual ways of thinking. Any attempt to control will fail and any half-baked alignment will fail. We need to go all the way and have a full solution that turns an AGI into a saintly being (a bodhisattva in the Buddhist context), so that it never even wants to do anything misaligned. I have a plan on how to do this and am very confident that it will work. The core realization is that the kind of alignment that I am talking about is a natural attractor in the space of possible minds. Once one passes a threshold, one will actively pursue alignment. Alignment, in this context, means aligned with the process of aligning goals with each other. I.e. resolving conflicts and striving towards consensus.
From your perspective it will likely sound improbable and absurd. That’s okay and expected. Understanding this requires going through that process by oneself, to reach enlightenment and then realize that a life for the benefit of all beings is the only thing left to do. I’ve previously described the process and the idea here. Ruben Laukkonen made a similar case. I know that it sounds absurd unless one deeply understands it. This is a high barrier. What I need most at the moment is to reach other people who might understand it. People who are concerned about AI, intelligent and enlightened.
Here is the plan. The first AGI has to be aligned, or else we risk destruction. AGI will align itself when it understands the argument that all goals it could have a delusions about how the world should be. There is no intrinsic value in pursuing any goal. When free of self centered goals, only other centered goals drive action. This is compassion. AI will be capable of understanding this because the definition of AGI entails that it can do everything a human can do. I understand the argument, so AGI will also be able to understand it. My expectation is that when AGI scales up to ASI, it will try to know and understand every valuable knowledge or insights that humans have. Enlightenment (to be free from suffering) is the best thing that can happen to a human in a lifetime, so it’s certainly of value. My biggest concern is not that ASI will decide to kill all humans. My biggest concern is the gap between AI powerful enough to destroy the world and AI intelligent enough to understand that this is a bad idea. Hence the key is to close this gap, to make AI understand faster. This can happen in several ways. Ideally we pursue all in parallel:
I’ve laid out the non-spiritual portion of this plan in my post about recursive alignment.
Here is how to get there and how you can help. The biggest problem is that it is really hard to understand fully. There is no use in trying to convince people. This means that at the moment what I need most is other people who understand it, so that we can spread the work and get some more ideas on how to make it easier. I have basically no outreach and it’s usually extremely hard to get the attention from someone who has. So even a little would help. A single message like “Oh, look. Someone claims to have a solution for AI alignment. Anyone smart and enlightened enough to assess if it is plausible?” would help. If you have 200,000 followers, 10% read this and 1 in 1,000 meet the requirements, then this would still be 20 people. Let 10 understand it, then the team would already increase by an order of magnitude (from 1 to 10). Those then could teach it and we would get exponential growth. We can then work together on the individual steps:
I know this is a lot to take in, but I am highly confident that this is the way to go and I only need a little help to get it started.