Intelligence allocation from a Mean Field Game Theory perspective
Each of us now faces a choice every minute of every day: Should I really use my brain power for this, or shoud I let ChatGPT do it? Using your own brain has its merits. You might get more intelligent and resilient in the long run. But ChatGPT already does most things much better and quicker, so why not let it spare you stirring your own grey goo? Then again, supporting ChatGPT may increase the likelihood of an extinction event. In this post, we’re gonna look at this from a mean field game theory perspective. First, let’s state the assumptions: There are N individuals, each of them having a level of intelligence described by the stochastic process X_i(t) for i=1,...,N. Each individual can choose at each time t a control α_i(t) in [0,1], where α_i(t) = 0 means the individual uses their own intelligence to solve the task and α_i(t) = 1 means they use ChatGPT to solve the task. The intelligence of an individual and the AGI evolve as a result of these decisions. The payoff for each individual is the cumulative tasks solved up to time T, but if the average intelligence of ChatGPT surpasses a certain threshold before time T, everyone is extinct and the payoff is zero. Individual Behavior (SDE) The dynamics of the individual's intelligence could be modeled as the following stochastic differential equation: dX_i(t) = [α_i(t) b_1(X_i(t), m(t)) + (1-α_i(t)) b_2(X_i(t), m(t))]dt + σ W_i(t) Here, the terms b_1 and b_2 represent the drifts under the two options available to the agents - use their own intelligence or use ChatGPT to solve tasks. Here's a possible way to define them: 1. b_1 could be a function that represents the rate of increase in the player's intelligence when they solve tasks on their own. For instance, we could model this as proportional to the current intelligence of the player, but with diminishing returns. This might look like: b_1(X_i(t), m(t)) = a_1 * X_i(t) / (1 + d_1 * X_i(t)) Here, a_1 is a positive constant that represents the maximum p
I agree on both counts. You're right that I should model the alignment of the system as well as its intelligence. I guess the alignment could be thought of as minimizing the distance of high dimensional vectors representing the players' and the AI's values. So each user (and the AI, too) could have a value vector associated with it, and the cost functions of the user could then incorporate how much they care about their own alignment (to the rest of the users), and the cost function of the AI needs to be tuned so that it is enough aligned when it reaches a critical threshold of intelligence. That way, you could express how important it is that the AI is aligned, as a function of its intelligence.