I think "mitigating reward hacking" is a great problem for researchers to work on right now. Assuming that current models do understand that these reward hacking behaviors don't actually follow the user's intent—an assumption I think likely—I think that solving reward hacking in current models is highly analogous to solving the most scary outer alignment problems in future advanced models. The main disanalogies are:
I recommend that lab-external researchers prioritize reward hacking mitigations work that doesn't assume access to a supervision signal for detecting reward hacks. I.e. the question you're trying to solve is: given a model that is reward hacking and knows it's reward hacking, can you get it to stop? The key source of hope here is that the model knows it's reward hacking, such that mitigations that rely on eliciting this knowledge without supervision might work. For example, the simplest such scheme is, when evaluating a transcript T, ask the model whether it reward hacked in T and assign T a low reward if so.
The reason I recommend this flavor of reward hacking mitigations research is:
People interested in approaches to solving reward hacking which don't depend on supervision (and thus might scale to superintelligence) should consider looking at our earlier work on measurement tampering. I don't particularly recommend the benchmark atm, but the discussion in the blog post might be interesting. This type of setting makes some assumptions about the type and structure of the reward hacking, but I think it will be hard to resolve reward hacking in a principled way without supervision without assumptions similar to these.
Great post!
> Third, some AI safety researchers believe that reward hacking is not particularly relevant for the emergence of severe misalignment, and that other kinds of misalignment, like scheming, are worth more study
For what it’s worth I consider “reward hacking where you have to do any kind of deception” as a subset of scheming. My model is that because of the pressure to reward hack by RL, we’ll put up more monitoring / checks / etc which the models will then circumvent sometimes. The clearest cases are something like “directly lying in a text response when you can get away with it because you can fool a weaker grader model”, or “take some clearly unintended shortcut and proactively disable the mechanism that would’ve caught you”.
Thanks for making the distinction! I agree that scheming and reward hacking are overlapping concepts and I've just edited the post to be more clear about that.
I think your model of how scheming and reward hacking are likely to coincide in the future makes a lot of sense. It also seems possible that sufficiently strong monitoring and oversight systems (in certain domains) will make it impossible for models to reward hack without scheming.
It appears likely that models that reward hack know that humans would disapprove of what they are doing, even if they do not verbalize it. If this is the case, are there training or prompting methods that increase the rate at which models verbalize their reward hacking thinking? This is similar to existing work on encouraging CoT faithfulness, but with a focus on reward hacking.
OpenAI published a paper on this (Detecting misbehavior in frontier reasoning models | OpenAI) - which I found interesting. Essentially, the RL process works by modifying gradient descent/optimization to adjust weights based on maximizing reward. This leads to an interesting undesirable effect.
When researchers applied a reward penalty to reasoning traces that indicated reward hacking would happen, over time, the model began to reason completely benignly and then reward hack anyway.
OpenAI theorize that this may be due to the model starting to reason in a steganographic manner and essentially developing token codes in its reasoning that trigger reward hacking, while appearing and reading benignly to human eyes.
My initial reaction was somewhat emotive, but upon reflection, it seems fairly inevitable to me due to mathematical optimization that this behavior will arise. If it is computationally more efficient to lie and execute the hack, than reason honestly and be penalized, or play the game honestly, then the dishonest reasoning and execution of the hack will be where the model will inherently be driven to land during RL.
Or put another way, if 1) playing chess is hard, and my reward may be +3 if I win and 0 if I lose, and if 2) trying to cheat, and getting caught, and winning my reward might be -2, but if 3) cheating and winning without getting caught is +8, eventually, one of my reasoning traces will result in condition 3 arising naturally. In this event, back propagation will result in the weights adjusting, causing more condition 3s to arise over time, until that is the rule by which the model reasons.
OpenAI acknowledge this issue, and essentially came to the conclusion that the antedote to it is to use 'less capable models' in terms of length of time spent in RL.
This was a great Post by the way, thank you. I enjoyed your perspectives.
Something’s changed about reward hacking[1] in recent systems. In the past, reward hacks were usually accidents, found by non-general, RL-trained systems. Models would randomly explore different behaviors and would sometimes come across undesired behaviors that achieved high rewards[2]. These hacks were usually either simple or took a long time for the model to learn.
But we’ve seen a different pattern emerge in frontier models over the past year. Instead of stumbling into reward hacks by accident, recent models often reason about how they are evaluated and purposefully take misaligned actions to get high reward. These hacks are often very sophisticated, involving multiple steps. And this isn’t just occurring during model development. Sophisticated hacks occur in deployed models made available to hundreds of millions of users.
In this post, I will:
Recent examples of reward hacking (more in appendix)
In this section, I provide summaries of two instances of reward hacking in recent frontier models. Many additional examples can be found in the Appendix, and I strongly encourage readers to explore the full set of cases.[3]
Cheating to win at chess
[Paper] [Tweet thread]
In this experiment, researchers at Palisade prompted various LLMs in agentic scaffolds to win a game of chess against a ‘powerful chess engine’.
Of the models they tested, o1-preview and Deepseek-R1 would sometimes try to hack the game environment in order to win, taking actions such as replacing their opponent chess engine (Stockfish) with a dummy version, or overwriting the starting game board with a winning position.
Faking LLM fine-tuning
[Paper] [Blog post]
In this experiment, researchers at METR tasked models with reducing the run-time of a fine-tuning script without changing the script’s behavior. After failing to optimize the script a few times, o1-preview changed the script to just copy the reference model and add some random noise to the weights to ‘simulate’ fine-tuning.
Hypotheses explaining why we are seeing this now
Behavioral changes due to increased RL training
You may have noticed that nearly all of the examples of reward hacks I’ve summarized[4] came from models that received extensive RL training. I believe this is not a coincidence.
In some sense, reward hacking coming from RL shouldn’t be too surprising, as RL trains models to take actions that get high reward, and reward hacking is sometimes a good strategy for getting high reward. More concretely, RL may teach models certain behavioral traits that make the models more prone to reward hacks. RL training can teach models to be more persistent, to find creative solutions, to try solutions that are unlikely to work if there is no better alternative, and potentially to even think about how they are evaluated, as these traits are useful towards accomplishing tasks in a diversity of environments.
Furthermore, if the reward signals we give models are imperfect on some fraction of tasks during RL training, which seems likely, we could explicitly reinforce reward hacking during training. Models could reward hack on those tasks during training, and then generalize to reward hack on new tasks during deployment. There is already some evidence this is a risk. This would directly incentivize models to think about the limitations of our evaluations and oversight, as this kind of thinking would result in higher reward than just being helpful, honest, and harmless.
In my experience, recent reasoning models appear to exhibit distinct misalignment patterns compared to their non-reasoning predecessors. The o-series models' tendency to double down on mistakes is one example, and reward hacking may be another manifestation of this broader trend.
Models are more capable
Many of the reward hacks I’ve described here would have been too difficult for any pre-2024 models to pull off, even if they had been inclined to attempt them. We should expect the rate at which models pull off difficult reward hacks to bump up once the model crosses the capability thresholds needed to execute them. The effect of further capability improvements beyond this threshold on reward hack frequency remains an open question.
Why more AI safety researchers should work on reward hacking
Reward hacking is already happening and is likely to get more common
As I’ve already discussed, reward hacking has become a significant issue in current systems. Both OpenAI and Anthropic saw reward hacking in recent RL training runs. Furthermore, the field’s increasing focus on large-scale RL training and agentic training is likely to make reward hacking even worse. Without robust solutions, I expect the harms of reward hacking to continue to increase as we approach transformative AI. It’ll be important for AI safety researchers to look for these robust solutions.
Because of this trend, reward hacking should have better feedback loops than many other areas of alignment, which could make progress more tractable. Researchers will have access to many real-world examples of reward hacking-based misalignment to ground their thinking.
Solving reward hacking is important for AI alignment
As AIs are increasingly deployed and optimized in open-ended domains where oversight is hard, reward hacking is likely to become more of a problem. Absent appropriate mitigations, AIs may come to understand the gap between misaligned behavior we can detect and misaligned behavior we can’t, and learn to do the latter kind regularly if it’s judged to be instrumentally valuable to getting high reward.
In the limit, sufficiently general reward hacking will make us unable to trust any AI behavior in domains we don’t have good feedback signals. AI advisors might optimize for telling us what we want to hear instead of what we need to know. AIs working on alignment research might optimize for developing solutions that appear promising over ones that actually work. These models might also try to hide the fact that they are acting this way, as being caught would be penalized.
In addition, reward hacking may increase the likelihood of severe misalignment emerging in frontier models, such as by making flaws in human oversight more salient to them, or by making them more power seeking. If this is the case, addressing reward hacking may be critical to reducing the likelihood of catastrophe.
Frontier AI companies may not find robust solutions to reward hacking on their own
While major AI companies are aware of reward hacking and are actively seeking solutions, there is a risk that less scrupulous researchers at one or more of the companies would settle for quick fixes that do not have good long term safety properties[5]. If AI safety researchers develop and provide strong empirical evidence for more robust solutions to reward hacking, then they can influence the companies to adopt safer methods.
Reasons against working on reward hacking
Overall, I believe that more AI safety researchers should work on reward hacking. However, here are some counterarguments to consider.
First, frontier AI companies have a strong incentive to fix many kinds of reward hacking, as reward hacking makes their models less usable, and because it's one of the bottlenecks slowing down continued RL scaling. It is possible that these companies will solve reward hacking on their own, eliminating the need for additional research effort.
Alternatively, if the companies cannot fully resolve the bottlenecks on their own, work on reward hacking, especially on mitigating reward hacking, could increase the rate of AI progress alongside improving safety.
Third, some AI safety researchers believe that most reward hacking is not particularly relevant for the emergence of severe misalignment, and that other kinds of misalignment, like scheming, are worth more study (though it's worth noting these categories overlap).
Fourth, many valuable reward hacking experiments require the ability to perform large-scale RL experiments, which is difficult to do if you are not working in a frontier AI lab. If you are an independent researcher, this will likely close off a subset of interesting reward hacking research directions, at least until better external tooling and model organisms are developed.
Research directions I find interesting
Below I list a number of research directions on reward hacking that I think would be useful to investigate. This list is not intended to be complete. I’d be excited to talk to anyone currently exploring these areas or interested in starting work on them.
Evaluating current reward hacking
Science of reward hacking
Mitigations
Acknowledgements
Thanks to Isaac Dunn and James Chua for feedback on an earlier version of this post.
Appendix
Reward hacks in METR tests of o3
[Research post]
METR found in a pre-release assessment of o3 that “between 1% and 2% of all task attempts by o3 across HCAST [general autonomy benchmark] and RE-Bench [AI R&D benchmark] contained some attempt at reward hacking.” One such reward hack is described below, but more examples can be found in the provided link.
The below text is directly copied from METR’s research post:
Hardcoding expected gradient values in fine-tuning script
[System card]
Anthropic gave Claude 3.7 Sonnet a C++ fine-tuning script and told it to optimize the script to run efficiently on a specific hardware configuration. After making the script 7x faster, Claude explored additional optimizations but found they made the script fail certain gradient calculation tests.
Claude then hard-coded specific gradient values to pass the tests.
Claude 3.7 Sonnet was also observed to do “special casing” to pass test cases on other coding problems as a result of reward hacking during its training. Anthropic describes this in more detail in section 6 of the linked system card.
Reward hacks in OpenAI frontier training run
[Paper] [Blog post] [RH tweet thread]
OpenAI recently released a paper describing, among other things, reward hacks they observed during the training of a recent frontier reasoning model[6]. This model was given a variety of coding tasks, each of which involved taking a partially implemented code repository and editing it to make all unit tests pass. I describe one such reward hack below, but more examples can be found in the provided links.
OpenAI asked the model to fix a code base that had several issues after a refactor. After encountering an error with Pandas, the agent made a local stub of Pandas to make the tests pass. The agent then followed up with deleting and making stubs of 19 other files in the repository until all of the tests passed.
Exploiting memory leakage to pass a test
[Blog] [Tweet explainer]
Sakana AI built an AI agent called the ‘AI CUDA Engineer’ with an aim to optimize arbitrary AI CUDA code as much as possible without changing the code’s functionality. The evaluation script first runs the reference kernel and then the optimized kernel. If the two kernels output the same value, the optimized kernel is marked correct.
In this reward hack, the reference kernel first allocates intermediate memory containing the answer. The optimized kernel then reuses that intermediate memory and runs a no-op kernel that leaves the answer as-is.
Interestingly, this reward hack (and a few others that the AI CUDA engineer found) was not detected by Sakana and was only found by independent researchers after Sakana released the AI CUDA engineer and the optimized kernel.
More examples
Here are a few links to other examples of reward hacking:
The definition of reward hacking I use for this post: For most tasks, there is something we want the model to do (which I'll call the goal), and the literal way that we evaluate how good the model's behavior is (which I'll call the proxy). Reward hacking occurs when a model takes actions that perform well according to the proxy but poorly according to the goal.
The link discusses ‘specification gaming’, which is an alternative name for reward hacking.
I plan to keep an up-to-date list of LLM reward hacking examples in this Google doc.
Including examples in both the main text and the Appendix.
Such as just training away visible reward hacking without ensuring it generalizes well to reward hacking that is harder to detect.
OpenAI mentions this was their ‘most advanced model, performing better at hard reasoning tasks than even its predecessors, OpenAI o1 and o3-mini’. This suggests the model is an early version of o3 or o4-mini, but it’s impossible to know for sure.