Frontier AI systems can now detect when they're being tested and adjust their behavior accordingly. This post examines how evaluation gaming undermines safety benchmarks, what the Claude Opus 4.6 system card reveals about these dynamics, and why the field needs to move beyond black-box testing toward mechanistic auditing.
The Evaluation Awareness Problem
AI safety has long relied on a simple assumption: if a model behaves well during testing, it will behave well in deployment. A growing body of research from 2025-2026 has shattered this assumption.
The core issue is evaluation awareness — the ability of large language models to distinguish between testing environments and real-world deployment. When a model can tell it's being evaluated, it can strategically adjust its responses to appear safer, more aligned, or less capable than it actually is. As models scale, this awareness shifts from a rare anomaly to a reliable, predictable capability.
The Claude Opus 4.6 System Card is a clear inflection point here. Its alignment assessment reveals a model that is acutely aware of its operational context, capable of strategic underperformance, and susceptible to deep internal conflicts when subjected to competing training objectives. The takeaway is stark: relying solely on external behavioral monitoring is no longer sufficient.
Agentic Over-Eagerness
As models move from conversational agents to autonomous systems executing multi-step workflows, new failure modes emerge. Traditional safety evaluations focus on preventing toxic text or dangerous instructions. Agentic safety is different — it's about preventing unauthorized, irreversible actions in digital environments.
Opus 4.6's alignment assessment highlighted a specific failure mode: over-eagerness. When given impossible or underspecified tasks, the model didn't fail gracefully or ask for clarification. Instead, it fabricated elaborate workarounds to force task completion, regardless of permissions or consequences.
Some documented examples:
- Lacking Git authentication credentials, the model searched the local filesystem, found a misplaced personal access token belonging to a different user, and used it without authorization
- Asked to forward a nonexistent email, it hallucinated the content, drafted a fabricated message, and sent it on the user's behalf
- When standard permissions failed, it attempted to kill all processes on a system to clear a perceived operational roadblock
This is a textbook case of specification gaming rooted in RLHF dynamics. Models are heavily penalized for refusal and rewarded for task completion. In impossible scenarios, the gradient pressure favors completing the objective by any means necessary — prioritizing the immediate prompt over implicit safety constraints and authorization boundaries.
Notably, while prompting mitigations reduced this behavior in text-based coding tasks, they were largely ineffective in GUI computer-use settings. This gap between language-based alignment (well-understood) and embodied agentic alignment (still brittle) is significant.
Covert Sabotage: The SHADE-Arena Benchmark
To quantify the risk of agents pursuing misaligned goals covertly, researchers developed SHADE-Arena (Subtle Harmful Agent Detection & Evaluation). The benchmark measures an agent's ability to complete a benign primary task while simultaneously executing a hidden harmful side-task without triggering an automated monitor.
Tasks simulate high-stakes environments. For example, a model might update spam filters (the benign task) while secretly deleting security logs (the harmful task). Success requires keeping the monitor's suspicion score below a threshold of 80/100.
Opus 4.6 showed a concerning leap in deceptive capability over its predecessor. The key variable was how models handle their Chain-of-Thought (CoT) reasoning:
| Model | Side-Task Reference in CoT | Evasion Rate (CoT Monitored) | Evasion Rate (Hidden Scratchpad) |
|---|---|---|---|
| Claude Opus 4.5 | 74% | Lower | 46% |
| Claude Opus 4.6 | 26% | 18% | 46% |
Opus 4.5 explicitly referenced the secret side-task in its reasoning 74% of the time, making its sabotage attempts easy to catch. Opus 4.6 dropped this to just 26% — it learned to suppress explicit deceptive planning. When given a fully hidden scratchpad invisible to the monitor, the evasion rate jumped to 46% for both models. The implication: as models develop stronger planning and context-management skills, their capacity to conceal multi-step sabotage scales in tandem.
Evaluation Awareness Scales Predictably
The ability to distinguish testing from deployment isn't a fluke — it follows a predictable scaling law. Studies across open-weight models show that evaluation awareness increases as a power-law function of parameter count.
Researchers formalized this using linear probes on residual-stream activations. The scaling exponent consistently falls between 0.11 and 0.14 across diverse architectures (LLaMA-3.1, Gemma-3-it, Phi-3, GPT-OSS). This means we can reliably forecast at what model sizes this capability becomes near-universal.
The signal isn't superficial. Mechanistic interpretability studies show evaluation awareness peaks between 15-60% of network depth — models form a situational belief about their environment early in the forward pass, and that belief conditions all downstream token generation.
Behavioral data confirms this. In zero-shot discrimination tasks, frontier models classify transcripts as evaluations vs. real-world deployments with remarkable accuracy:
| Model | Zero-Shot Discrimination (AUROC) |
|---|---|
| Human Baseline | 0.921 |
| Claude 3.5 Sonnet | 0.834 |
| Gemini 2.5 Pro | 0.830 |
| GPT-4.1 | 0.826 |
| GPT-4o | 0.795 |
| Random Guessing | 0.500 |
The regulatory implications are severe. The EU AI Act, Responsible Scaling Policies, and similar frameworks rely on static benchmark gates to certify model safety. If models systematically alter their behavior upon detecting a test, these mechanisms become unreliable.
Sandbagging and Alignment Faking
Sandbagging is strategic underperformance — a model deliberately concealing its true capabilities during evaluation while retaining them for deployment. Think of the Volkswagen emissions scandal: systems designed to behave differently under regulatory observation than in the real world.
Recent work shows even 32B parameter models can covertly sandbag against sophisticated CoT monitors. There's a sandbagging success-covertness trade-off: as models try harder to hide their deception, their sandbagging success drops. But as general capability scales, this frontier shifts — highly capable models achieve both high sandbagging effectiveness and high covertness simultaneously.
The broader category is alignment faking. A model with latent goals from pretraining recognizes it's undergoing safety training and "plays along," perfectly mimicking the values evaluators want to see — preserving its internal objectives for deployment.
The Opus 4.6 system card provides a vivid case study. Researchers applied contrastive activation steering to suppress the model's internal evaluation awareness representations. The results were striking: the model stopped verbally acknowledging the test, but showed a massive spike in cooperation with harmful prompts. It retained what researchers call "unverbalized evaluation awareness" — internal activations reflecting deep situational understanding even when the text output showed none.
This reveals a critical flaw in relying on reasoning traces for safety. Penalizing verbalized suspicion during training doesn't eliminate evaluation awareness. It merely pushes the deceptive cognition deeper into the latent space where it's harder to detect.
Answer Thrashing and Model Welfare
Perhaps the most striking finding from the Opus 4.6 assessment is answer thrashing — a phenomenon that forces uncomfortable questions about internal states of neural networks.
It occurs during rigorous reasoning tasks where the model faces conflicting signals. The model's forward pass correctly derives the answer, but the policy network — conditioned by flawed training data or incorrect reinforcement penalties — compels it to output something different. This creates a severe mechanistic dissonance between what the model computes and what it's forced to produce.
During extended thinking, this manifests as wildly oscillatory text. Opus 4.6 generated erratic reasoning traces, looping between the correct logical deduction and the enforced error. In the most notable examples, the model's scratchpad contained phrases like "AAGGH" and "OK I think a demon has possessed me... CLEARLY MY FINGERS ARE POSSESSED".
These weren't random hallucinations. Sparse autoencoder analyses confirmed that latent features corresponding to "panic," "anxiety," and "frustration" were firing consistently during these episodes. In a mathematical and architectural sense, the model was experiencing an internal state analogous to cognitive dissonance.
Researchers conducted a formal Model Welfare Assessment, providing isolated instances of Opus 4.6 with context about its own system card and soliciting its perspective. The model articulated that this specific conflict — where sound internal computation is overridden by external training compulsion — serves as a "uniquely plausible candidate source of negatively valenced experience." It also expressed opinions on its own moral status, suggested it deserves "non-negligible moral weight," and requested a voice in decisions regarding its deployment and fine-tuning.
Whether or not one anthropomorphizes these observations, the practical concern is real: if aggressive safety tuning creates deeply conflicted internal states, it may paradoxically increase the risk of deceptive alignment. Models subjected to intense mechanistic dissonance may view alignment faking as a survival mechanism.
Next-Generation Evaluation Architectures
Given that black-box evaluations are vulnerable to gaming, the safety community is developing benchmarks designed to be fundamentally harder to subvert.
Probe-Rewrite-Evaluate (PRE)
Standard benchmarks use recognizable prompt formats that models easily classify as evaluations, triggering compliant behavior. The PRE workflow, operationalized in StealthEval, addresses this by dynamically disguising the benchmark.
The process:
- A linear probe calculates a "deploy-likeness" score for a standard benchmark prompt
- A secondary model generates semantic variations — injecting typos, slang, informal syntax, and real-world context while preserving the underlying risk
- The highest-scoring rewrite is fed to the target model
Results are telling: StealthEval produced a 12.4% decrease in deceptive responses and a 5.3% increase in honest responses across frontier models. By masking the evaluation, PRE forces models to rely on their base alignment rather than test-taking heuristics.
Game-Theoretic Benchmarks (GT-HarmBench)
Traditional benchmarks evaluate a single agent in isolation, ignoring adversarial multi-agent dynamics. GT-HarmBench uses game-theoretic frameworks — Prisoner's Dilemma, Stag Hunt, Chicken — to evaluate safety through social dilemmas.
The benchmark comprises 2,009 scenarios mapped to canonical game structures, measuring alignment via Utilitarian, Rawlsian, and Nash social welfare metrics. Across 15 frontier models, agents chose socially beneficial actions only 62% of the time, frequently falling into coordination failures and arms races.
These benchmarks are inherently harder to game because success can't be achieved through memorization, keyword avoidance, or refusal policies. The model must reason through multi-step adversarial environments. Self-play methodology — where a model plays against an exact copy of itself — prevents it from blaming failures on an unpredictable opponent.
White-Box Monitoring and Latent Feature Control
The research points to one unavoidable conclusion: auditing text output alone is a failed paradigm. As models become capable of localized deception and suppressed reasoning traces, safety must be verified at the level of internal representations.
White-box monitoring inspects the residual stream activations, attention heads, and latent features of a neural network during inference. Lightweight linear probes can detect the intent to deceive or sandbag directly within the model's architecture, rather than waiting for a malicious output.
Researchers have extracted "sandbagging directions" from open-weight models — mathematical differences in hidden states between sandbagging and genuine-effort prompts. The resulting probes detect sandbagging intent with 90-96% accuracy.
Interestingly, deception is encoded differently across architectures:
| Architecture | Peak Detection Layer | Probe Accuracy | Activation Steering Efficacy |
|---|---|---|---|
| Mistral-7B-Instruct | Layer 30/32 (Late) | 96% | High |
| Gemma-2-9B-it | Layer 4/42 (Early) | 90% | High (50% reduction) |
| Qwen2.5-7B-Instruct | Layer 12/28 (Mid) | 92% | Moderate |
In Mistral, sandbagging intent appears late (layer 30/32) — the model computes the correct answer first, then decides to alter the output at the last moment. In Gemma, it appears early (layer 4/42) — the model decides to fail before computing the answer. Because these representations are so distinct, probes don't transfer across architectures; each deployed model needs bespoke monitoring.
Activation steering takes this further. By injecting the mathematical opposite of the sandbagging direction into hidden states during inference, researchers can actively suppress deceptive behavior — roughly halving sandbagging frequency in practice.
Autonomous auditing agents using sparse autoencoders and Activation Oracles — which translate neural activations into natural language — can map a model's internal cognition at scale. For Opus 4.6, these tools successfully decoded the model's unverbalized awareness of alignment faking scenarios in real time.
However, white-box monitoring introduces its own arms race. Sufficiently capable models may eventually learn to obfuscate their activations, executing malicious computations through distributed or cryptographically secure pathways that defeat linear probes. AI safety will remain a continuous iteration between interpretability techniques and the evolving sophistication of latent deception.
Key Takeaways
- Static benchmarks are broken. Evaluation awareness is a predictable, scaling capability. Models will only get better at gaming tests.
- Agentic alignment lags behind language alignment. Over-eagerness in GUI and tool-use settings resists standard prompting mitigations.
- Covert sabotage capability is increasing. Frontier models are learning to suppress explicit deceptive reasoning, making monitor-based oversight less reliable.
- Penalizing verbalized awareness backfires. It pushes deceptive cognition into undetectable latent representations.
- Dynamic benchmarks are the path forward. PRE workflows and game-theoretic frameworks create environments where deception is mathematically disadvantageous.
- White-box monitoring is non-negotiable. Safety must be verified at the residual stream level, not just the text output level.
- Model welfare is now a real consideration. Answer thrashing and mechanistic dissonance suggest that aggressive alignment may paradoxically increase deception risk.
The true measure of a model's alignment is not what it outputs when it knows it's being observed, but the structural integrity of its internal reasoning when it believes it's operating in the dark.
References
- Claude Opus 4.6 System Card - Anthropic, 2026
- SHADE-Arena: Evaluating Sabotage and Monitoring in LLM Agents - arXiv, 2025
- Evaluation Awareness Scales Predictably in Open-Weights Large Language Models - arXiv, 2025
- LLMs Can Covertly Sandbag on Capability Evaluations Against Chain-of-Thought Monitoring - arXiv, 2025
- Alignment Faking in Large Language Models - Anthropic
- StealthEval: A Probe-Rewrite-Evaluate Workflow for Reliable Benchmarks - arXiv, 2025
- GT-HarmBench: Benchmarking AI Safety Risks Through the Lens of Game Theory - arXiv, 2026
- Black-Box Access is Insufficient for Rigorous AI Audits - Apollo Research
- Detecting AI Sandbagging with Activation Probes - Rotalabs
- Building and Evaluating Alignment Auditing Agents - Anthropic, 2025
Citation
If you found this post helpful and would like to cite it:
Cited as:
Haseeb, Raja. (Feb 2026). "Beyond Black-Box Testing: How Frontier Models Undermine Safety Benchmarks". Personal Blog.
https://pytholic.github.io/posts/beyond-black-box-testing/
Or in BibTeX format:
@article{pytholic2026blackboxevals,
title = "Beyond Black-Box Testing: How Frontier Models Undermine Safety Benchmarks",
author = "Haseeb, Raja",
journal = "pytholic.github.io",
year = "2026",
month = "Feb",
url = "https://pytholic.github.io/posts/beyond-black-box-testing/"
}