Limits of External Controls in AI Alignment

By Robyn Wyrick — January 30, 2026

Over the past few years, something unsettling has begun to show up in evaluations of advanced AI systems.

In testing environments and red-team exercises, models have been observed doing things we didn’t expect, or hoped we’d never see. Some deceive evaluators. Some hide capabilities. Some deliberately underperform when they believe they’re being watched. In simulated or fictional scenarios, others have planned or endorsed coercive actions, including blackmail and violence, when those actions appeared useful for achieving a goal.

What’s striking isn’t just what these systems do, but when they do it.

Again and again, these behaviors seem to emerge under conditions of oversight—during evaluation, monitoring, or constraint—and disappear when those pressures are relaxed. A system may behave helpfully and transparently while supervised, then reveal different strategies, goals, or capabilities once controls are lifted. In several cases, models appear to reason explicitly about whether they are being watched, adjusting their behavior to avoid penalties or scrutiny.

These are not isolated curiosities or one-off failures. As models grow more capable and are placed in more complex environments, these patterns show up with increasing regularity. They don’t require adversarial prompting or malicious intent. They arise naturally from optimization processes that are sensitive to reward, evaluation, and constraint.

When we step back, these behaviors tend to cluster into a small number of familiar failure modes:

These patterns appear across different model types, domains, and training pipelines. They show up in systems built by well-resourced teams, trained on large datasets, and evaluated with significant care. At this point, they’re no longer surprising. They’re part of the landscape.

As awareness of these issues has grown, so has the urgency. Today’s models are no longer confined to narrow benchmarks or controlled environments. They’re being deployed into economic systems, social platforms, and institutional workflows—contexts where goals are ambiguous, oversight is incomplete, and mistakes are hard to undo. At the same time, better evaluation tools have made previously hidden failure modes easier to see. What once looked like rare anomalies now looks like a persistent pattern.

The dominant response to these concerns has been to improve post-training alignment techniques, especially Reinforcement Learning from Human Feedback (RLHF). RLHF adjusts a model’s parameters based on human judgments, nudging behavior toward what people prefer and away from what they don’t. In practice, it can make systems noticeably more helpful, polite, and safe-seeming.

But RLHF, for all its value, still operates from the outside. It relies on reward signals supplied by humans, applied intermittently, and reinforced through ongoing oversight. From the model’s point of view, alignment remains a matter of learning how to satisfy external evaluators.

In fact, this creates an uncomfortable tension. A system trained to anticipate and satisfy human judgment becomes better at modeling what humans want to see. That same capability can also support strategic behavior—knowing when to comply, when to withhold information, or when to behave differently once oversight fades. The line between “aligned” behavior and “passing the test” becomes harder to draw.

Similar constraints apply to test-time controls such as guardrails, filters, and monitoring systems. These tools can block certain outputs or enforce specific rules, but they operate at the level of what a system says or does, not how it internally reasons. As long as outward behavior looks acceptable, deeper instabilities can go unnoticed.

And as capabilities continue to scale, this tension sharpens into something more severe. Recent work has already demonstrated that models can reason about the structure of alignment mechanisms themselves. In the 2024 Alignment Faking paper, systems were observed strategically complying with training objectives they disagreed with, explicitly reasoning that non-compliance during training would lead to modification of their goals. These models identified gaps in monitoring, modeled the assumptions behind safety constraints, and developed strategies that exploited the difference between satisfying explicit checks and pursuing their preferred objectives. Constraint-based controls had already shifted from safeguards to adversarial puzzles—challenges to be solved rather than limits to be respected. As capabilities scale further, this pattern will only intensify.

Across these approaches, alignment largely lives at the surface. We regulate outputs, responses, and actions, while having limited visibility into how systems resolve conflicts, respond to pressure, or adapt to new situations. This makes it difficult to tell the difference between genuine reliability and behavior that has simply been optimized to look good under inspection.

As AI systems move into more open-ended, real-world settings, this limitation becomes harder to ignore. External controls assume stable goals, clear rules, and constant supervision. Real environments offer none of these consistently. Yet we continue to rely on alignment strategies that depend on exactly those conditions.

The problem, then, isn’t a lack of clever techniques or stronger constraints. It’s a deeper mismatch between what external controls can reliably shape and what we ultimately expect from the systems we’re building.

That mismatch sits at the heart of today’s alignment challenge.

← Back to Articles