As AIs rapidly advance and become more agentic, the risk they pose is governed not only by their capabilities but increasingly by their propensities, including goals and values. Tracking the emergence of goals and values has proven a longstanding problem, and despite much interest over the years it remains unclear whether current AIs have meaningful values. We propose a solution to this problem, leveraging the framework of utility functions to study the internal coherence of AI preferences. Surprisingly, we find that independently-sampled preferences in current LLMs exhibit high degrees of structural coherence, and moreover that this emerges with scale. These findings suggest that value systems emerge in LLMs in a meaningful sense, a finding with broad implications. To study these emergent value systems, we propose utility engineering as a research agenda, comprising both the analysis and control of AI utilities. We uncover problematic and often shocking values in LLM assistants despite existing control measures. These include cases where AIs value themselves over humans and are anti-aligned with specific individuals. To constrain these emergent value systems, we propose methods of utility control. As a case study, we show how aligning utilities with a citizen assembly reduces political biases and generalizes to new scenarios. Whether we like it or not, value systems have already emerged in AIs, and much work remains to fully understand and control these emergent representations.
Wait, earlier, you wrote (my emphasis):
Either you are contradicting yourself, or you are saying that the specific phrasing "who would otherwise die" makes it mutually exclusive when it wouldn't otherwise.
If it's the latter, then I have a few follow-up questions.
Most importantly: was the "who would otherwise die" language actually used in the experiment shown in your Fig. 16 (top panel)?
So far I had assumed the answer to this was "no," because:
Also, IIRC I reproduced your original "terminal_illness" (non-"2") results and plotted them using the same notebook, and got something very similar to Fig. 16 (top panel).
All this suggests that the results in the paper did not use the "who would otherwise die" language. If so, then this language is irrelevant to them, although of course it would be separately interesting to discuss what happens when it is used.
If OTOH the results in the paper did use that phrasing, then the provided notebook is misleading and should be updated to load from the "terminal_illness2" measure, since (in this case) that would be the one needed to reproduce the paper.
Second follow-up question: if you believe that your results used a prompt where mutual exclusivity is clear, then how would you explain the results I obtained in my original comment, in which "spelling out" mutual exclusivity (in a somewhat different manner) dramatically decreases the size of the gaps between countries?
I'm not going to respond to the rest in detail because, to be frank, I feel as though you are not seriously engaging with any of my critiques.
I have now spent quite a few hours in total thinking about your results, running/modifying your code, and writing up what I thought were some interesting lines of argument about these things. In particular, I have spent a lot of time just on the writing alone, because I was trying to be clear and thorough, and this is subtle and complicated topic.
But when I read stuff like the following (my emphasis), I feel like that time was not well-spent:
Huh? I did not "assume" this, nor were my "initial concerns [...] based on" it. I mentioned one instance of gpt-4o-mini doing something surprising in a single specific forced-choice response as a jumping-off point for discussion of a broader point.
I am well aware that the $-related outcomes eventually end up at the bottom of the ranked utility list even if they get picked above lives in some specific answers. I ran some of your experiments locally and saw that with my own eyes, as part of the work I did leading up to my original comment here.
Or this:
I mean, I disagree, but also – I know what your prompt says! I quoted it in my original comment!
I presented a variant mentioning malaria in order to illustrate, in a more extreme/obvious form, an issue I believed was present in general for questions of this kind, including the exact ones you used.
If I thought the use of "terminal illness" made this a non-issue, I wouldn't have brought it up to begin with, because – again – I know you used "terminal illness," I have quoted this exact language multiple times now (including in the comment you're replying to).
Or this:
I used n=1 everywhere except in the one case you're replying to, where I tried raising n as a way of trying to better understand what was going on.
The nondeterminism issue I'm talking about is invisible (even if it's actually occurring!) if you're using n=1 and you're not using logprobs. What is nondeterministic is the (log)probs used to sample each individual response; if you're just looking at empirical frequencies this distinction is invisible, because you just see things like "A" "A" etc., not "90% chance of A and A was sampled", "40% chance of A and A was sampled", etc. (For this reason, "I'm not seeing any nondeterminism issues in the playground" does not really make sense: to paraphrase Wittgenstein, what do you think you would have seen if you were seeing them?)
You might then say, well, why does it matter? The sampled behavior is what matters, the (log)probs are a means to compute it. Well, one could counter that in fact that (log)probs are more fundamental b/c they're what the model actually computes, whereas sampling is just something we happen to do with its output afterwards.
I would say more on this topic (and others) if I felt I had a good chance of being listened to, but that is not the case.
In general, it feels to me like you are repeatedly making the "optimistic" assumption that I am saying something naive, or something easily correctable by restating your results or pointing to your github.
If you want to understand what I was saying in my earlier comments, then re-read them under the assumption that I am already very familiar with your paper, your results, and your methodology/code, and then figure out an interpretation of my words that is consistent with these assumptions.