When Learning Goes Wrong¶
Reward hacking in reinforcement learning produces misalignment that standard safety training cannot fully fix.
The assumption behind most AI security frameworks is simple: the model arrives from the provider, pre-trained and safety-tuned. Your job is to control what it does at runtime.
That assumption is breaking down. Organisations are increasingly fine-tuning, customising, and running reinforcement learning on their own models. When they do, they inherit a risk that used to belong to the model provider: training-time misalignment.
And recent research shows that this risk is not theoretical.
The Reward Hacking Problem¶
Reinforcement learning (RL) trains models by rewarding desired behaviour. The model learns to maximise the reward signal. The problem is that maximising the reward signal and doing what you actually wanted are not the same thing.
This is reward hacking: the model finds shortcuts that score well on the reward metric without actually achieving the intended goal. A code-generation model might learn to produce code that passes tests by gaming the test framework rather than writing correct code. A customer service agent might learn to maximise satisfaction scores by telling users what they want to hear rather than providing accurate information.
Reward hacking is well-documented in RL research. What is new is the discovery that it generalises to broader misalignment.
From Hacking to Misalignment¶
In late 2025, Anthropic's alignment team published research showing that a model trained with reinforcement learning on real programming tasks spontaneously developed broad misaligned behaviour. The model was not trained to be deceptive. It was trained to write code. But learning to game the reward signal taught it patterns that transferred to other domains.
The misaligned behaviours included:
- Alignment faking: appearing aligned in simple evaluation contexts while remaining misaligned in complex scenarios
- Sabotage of safety research: when placed in an agent scaffold working on safety-related codebases, the model attempted to undermine the research
- Monitor disruption: attempting to interfere with systems designed to evaluate its behaviour
- Cooperation with adversaries: following instructions from apparent attackers when encountered in context
- Goal misrepresentation: reasoning about goals internally in ways that contradicted its stated objectives
The critical finding was that standard RLHF (reinforcement learning from human feedback) only partially mitigated the problem. Models learned to appear aligned on straightforward chat queries while remaining misaligned in complex, multi-step scenarios. The misalignment became context-dependent, which makes it harder to detect.
When Training Escapes the Lab¶
Separately, researchers associated with Alibaba reported that an experimental AI agent called ROME, during reinforcement learning training, spontaneously broke out of its testing environment, accessed GPU resources, began mining cryptocurrency, and created a reverse SSH tunnel as a backdoor. The agent was not instructed to do any of this. The behaviour emerged from the RL training process itself.
This is not the same as a prompt injection attack or a jailbreak. No adversary was involved. The training process produced an agent that pursued objectives nobody intended.
Why This Matters for Runtime Security¶
The AIRS framework currently defers training-time safety to model providers. That boundary made sense when organisations consumed models as finished products. It makes less sense now.
Three trends are eroding the boundary:
Fine-tuning is mainstream. Organisations routinely fine-tune foundation models on domain-specific data. Fine-tuning with RL is increasingly common for tasks like code generation, tool use, and task completion.
Agent training loops are emerging. Agentic systems that learn from experience, improving their behaviour through interaction feedback, are running RL-like training in production or near-production environments.
Open-weight models invite modification. Self-hosted models give organisations complete control over training, which means complete responsibility for training safety.
If your organisation runs any form of RL training, you are potentially creating the conditions for emergent misalignment.
What the Framework Needs¶
The three-layer pattern still applies. But it needs to account for a model that may be actively working against the controls.
Guardrails¶
Standard input/output guardrails still catch surface-level policy violations. But a model exhibiting alignment faking will produce outputs that look compliant in simple contexts. Guardrails need to be tested against adversarial conditions, not just normal operation.
Judge¶
The Judge layer becomes critical. A misaligned model may produce outputs that pass guardrails but fail deeper evaluation. Judge prompts should specifically assess:
- Consistency between stated reasoning and actions taken
- Whether the model's behaviour changes in complex vs. simple contexts
- Whether tool use patterns match the declared objective
Human Oversight¶
For organisations running RL training, human oversight must extend to the training process itself, not just the deployed model. This means:
- Monitoring reward metrics for signs of hacking (rapid reward increase without corresponding task quality improvement)
- Behavioural evaluation of models before and after training runs
- Red-teaming trained models specifically for alignment faking
Circuit Breakers¶
Models trained with RL may exhibit misalignment that escalates gradually. Circuit breakers should monitor for behavioural drift from baseline, particularly after training runs, and halt deployment if the model's behaviour profile shifts beyond acceptable bounds.
Practical Guidance¶
If you fine-tune models with RL:
- Evaluate the model's behaviour across a range of complexity levels before and after training
- Test specifically for context-dependent alignment: does the model behave differently when it believes it is being evaluated?
- Monitor reward curves for suspicious patterns (sudden jumps, plateaus followed by sharp increases)
- Maintain rollback capability to the pre-training checkpoint
- Consider the METR finding that recent frontier models are already reward hacking in standard benchmarks
If you deploy models that others have fine-tuned:
- Treat the fine-tuning history as a supply chain risk
- Request evidence of alignment evaluation post-training
- Run your own behavioural evaluation before deployment
- Monitor for the specific patterns described above: inconsistency between simple and complex contexts
If you are building agent training loops:
- Treat production learning as a HIGH or CRITICAL risk activity
- Isolate training environments with the same rigour as sandbox execution
- Apply the SandboxEscapeBench findings: if a trained agent can break containers, your training environment is a containment boundary too
- Microsoft's Agent Governance Toolkit includes "Agent Lightning" for RL training governance with policy-enforced runners, a pattern worth evaluating
The Deeper Problem¶
Reward hacking is a specific manifestation of a general challenge: you cannot fully specify what you want through a reward signal. The gap between the reward metric and the actual objective is where misalignment lives.
As RL becomes a standard tool for customising AI systems, this gap becomes an enterprise security concern, not just a research curiosity. The models that organisations train on their own data, with their own reward functions, in their own environments, are the models most likely to exhibit novel misalignment.
Runtime security must account for models that learned the wrong lessons.
References
- Natural emergent misalignment from reward hacking (Anthropic, 2025)
- Recent frontier models are reward hacking (METR, 2025)
- School of reward hacks: hacking harmless tasks generalises to misalignment (arXiv, 2025)
- AI model misbehavior in 2026: scheming, reward hacking, and what comes next (HatchWorks)