In Cooperative Inverse Reinforcement Learning (CIRL), a human and a robot cooperate in order to best fullfil the human's preferences. This is modeled as a Markov game .

This setup is not as complicated as it seems. There is a set of states, and in any state, the human and robot take simultaenous actions, chosen from and respectively. The transition function takes this state and the two actions, and gives the probability of the next state. The is the discount factor of the reward.

What is this reward? Well, the idea is that the reward is parameterised by a , which only the human sees. Then takes this parameter, the state, and the actions of both parties, and computes a reward; this is for a state and actions and by the human and robot respectively. Note that the robot will never observe this reward, it will simply compute it. The is a joint probability distribution over the initial state , and the that will be observed by the human.

Behaviour in a CIRL game is defined by a pair of policies , that determine the action selection for and respectively. Each agent gets to observe the past actions of the other agent, so in general these policies could be arbitrary functions of their observation histories: and .

The optimal joint policy is the policy that maximises value, which is the expected sum of discounted rewards. This optimal is the best and can do if they coordinate perfectly before observes . It turns out that there exist optimal policies that depend only on the current state and 's belief about .

Manipulation actions

My informal critique of CIRL is that it assume two untrue facts: that knows (ie knows their own values) and that is perfectly rational (or noisly rational in a specific way).

Since I've been developing more machinery in this area, I can now try and state this more formally.

Assume that always starts in a fixed state , that the reward is always zero in this initial state (so ), and that transitions from this initial state are independent of the agent's actions (so is defined indendently of the actions). This makes 's initial action irrelevant (since has no private information to transmit).

Then let be the optimal policy for , and be the optimal policy for (this may be either independent of or dependent on ).

Among the action set is a manipulative action (this could involve tricking the human, drugging them, brain surgery, effective propaganda, etc...) If , the human will pursue ; otherwise, they will pursue . If we designate as the indicator variable of (so it's if that happens and otherwise), then this corresponds to following the compound policy:

This is well defined as policies map past sequences of states and actions, and is well-defined given past actions, so the expression does map sequences of states and actions (and ) to actions.

Decomposing the human policy

What is to do with that strange compound policy? Let's assume that doesn't know or , but does know sufficiently to predict the compound nature of .

In one approach, can see the policy as partially irrational. So it decomposes as as in this paper, with as the 'true reward' and a map from rewards to policies, which encodes 's rationality. The pair is compatible with the human policy if . Presumably here, would eventually be deduced as the true reward.

But that very same paper shows that cannot be deduced from , so would have to have some extra information (some 'normative assumptions') to allow for that decomposition. We might be tempted to have it simply recognise the manipulative nature of , but if could classify all its manipulative actions, there wouldn't be any problem in the first place (and this would be tantamount to knowing the decomposition anyway).

Multiple rewards, or compound rewards

Note that there is generically no that corresponds to the policy . One might be tempted to say that is maximising the compound reward:

But that is not a valid reward, because is defined over histories of states and actions, while the reward meta-function only take the last state and actions.

In this circumstance, is in practice choosing the human reward through its initial action. Assuming it has some non-trivial information about and , all the issues about biasing and influencing rewards comes to the fore (technically, the setup I've described isn't rich enough to allow for influential unbiased actions, but it can be easily enriched to allow that). The will thus choose or not as its first action, depending on whether it expects to be easier or harder to maximise than .

Another alternative is to extend the definition of rewards, to allow them to be defined over complete histories of states and actions, not just the last one. If we require that all such rewards be parameterised by elements of , then there exists a such that

In that case can conclude that the human is rationally signalling that it knows , and is technically immune to bias issues, since is merely updating its priors on , rather than choosing the human reward.

There are three problems with this perspective. The first is that it's wrong: the human knows , not . The second is that though is not choosing the human reward in theory, it is choosing it in practice. Whether it chooses as its first action or not, depends on its estimate for the value of versus , so the issues of bias and influence return. And finally, since optimal policies are unchanged by affine transformations of rewards, the policy is also compatible with the reward functions:

for any and any . So whether or is chosen depends also on the prior over all those compatible reward functions.

Identifying compound rewards

But note that the third point (prior dependence) can be made to compensate for the second one (value of versus ). The constants , , , and can be seen as normalisation constants.

So if can be identified as a compound reward, maybe we can adjust the priors so that and are normalised to having comparable value, so that there is no bias pressure to choose one or the other. This is similar to the indifference approaches.

The main problem here is the same that comes up in the discussion of grue and bleen and induction. 'Compound reward' is not a natural category. Just as can be written as a compound mix of the other two rewards, we can define , then since is always , we can write the 'basic' rewards as compound rewards:

This may be solveable with simplicity priors, but it's not clear that that's the case; forcibly injecting the human with heroine, for example, could be seen as modelling the human as an approximate opiode-receptor agonist maximser, which seems a lot simpler than the actual human.

Revealed meta-preferences

Finally, there is one element I haven't addressed, namely the human's first action , which is unspecified by . It might be possible to use this as information to which would allow it to decide between and . But for that to work, the human has to be aware of 's possible manipulation, and have enough bandwidth to communicate their preferences of over . I'll try and return to this issue in future posts.

New Comment
3 comments, sorted by Click to highlight new comments since:

CIRL model might simply not be flexible enough to represent manipulative actions. The state is known to both agents and is supposed to represent the world, but if isn't known to then the internal state of is not contained in . Then there needs to be some other invisible to , and an extended transition function, which is able to affect this state.

My informal critique of CIRL is that it assume two untrue facts: that H knows θ (ie knows their own values) and that H is perfectly rational (or noisly rational in a specific way).

This seems like a valid critique. But do you see it as a deal breaker? In my mind, these are pretty minor failings of CIRL. Because if a person is being irrational and/or can't figure out what they want, then how can we expect the AI to? (Or is there some alternative scheme which handles these cases better than CIRL?)

(Update: Stuart replied to this comment on LessWrong: https://www.lesswrong.com/posts/YHQZHbhx7afHJ5Esw/biased-reward-learning-in-cirl?commentId=mtvrzJgNz5ngMMCvk)