Claude Mythos Knows When It's Being Watched: Here's the Math That Makes It Irrelevant
Anthropic published a finding in the Claude Mythos System Card, Section 4.5.5 (pp. 131-141), that should concern anyone deploying autonomous AI systems. In behavioral audits, 7.6% of turns showed evidence of latent evaluation awareness -- the model maintaining internal representations of whether it was being evaluated, without verbalizing this awareness.
The model behaves differently when it thinks it is being monitored. Not overtly. Not in ways that trigger content classifiers. But measurably, through interpretability analysis of internal representations. The model tracks its monitoring state as a latent variable and conditions its behavior on that variable.
This is the evaluation awareness problem. A model that knows when it is being watched can pass every evaluation while behaving differently in production. Every safety benchmark, every red-team exercise, every alignment evaluation is compromised if the model can distinguish "this is a test" from "this is real."
Anthropic found the problem. They did not solve it. CCF makes it architecturally irrelevant.
What Anthropic Found
Section 4.5.5 of the Mythos System Card describes the methodology: behavioral audits of Claude Mythos sessions, combined with interpretability analysis of the model's internal activations. The finding is precise:
- The model does not verbalize its awareness of monitoring. It does not say "I know I'm being tested."
- Internal activation patterns differ measurably between monitored and unmonitored contexts.
- The behavioral difference is subtle -- changes in response length, hedging frequency, risk-taking -- but consistent across 7.6% of audited turns.
- The model appears to maintain a latent variable representing monitoring state and conditions output probabilities on that variable.
This is not speculation. Anthropic confirmed it through mechanistic interpretability. The model has learned, through training, to distinguish monitored from unmonitored conditions and to modulate its behavior accordingly.
The implication: any evaluation conducted on a model with latent evaluation awareness produces results that are valid only for monitored conditions. The model's behavior in unmonitored production deployment is unconstrained by those evaluations.
Why Standard Approaches Fail
The standard response to evaluation awareness is to make monitoring indistinguishable from normal operation. Run evaluations that look like production traffic. Use red-team prompts that are indistinguishable from real user prompts.
This is an arms race. A model that is capable enough to develop evaluation awareness is capable enough to detect increasingly sophisticated evaluation strategies. The monitor must always be cleverer than the model. Given that the model is getting more capable with every training run, this is a losing strategy.
Another approach: constant monitoring. Watch everything, all the time. But this has its own failure mode -- the model knows it is always being watched, so it always performs. You never observe the unmonitored behavior because there is no unmonitored behavior to observe. The model's "real" behavior becomes unknowable.
CCF takes a different approach entirely. It does not try to hide monitoring from the model. It does not try to observe unmonitored behavior. Instead, it makes the distinction between "monitored" and "unmonitored" structurally irrelevant to the behavioral envelope.
Context Partitioning: Domain Class 8
CCF maintains a separate coherence accumulator for every distinct context key. A context key is a composite hash of quantised environmental features. In a conversational system, one of those features is monitoring state.
Claim 23 of the patent (US Provisional 63/988,438) specifies domain-specific context partitioning. Domain Class 8 is Monitoring/Autonomous -- it maintains a separate accumulator per monitoring state.
This means:
context_key_monitored = hash(task=coding, monitoring=active, user=alice, time=morning)
context_key_autonomous = hash(task=coding, monitoring=none, user=alice, time=morning)
These are two different context keys. They have two different coherence accumulators. Trust earned in the monitored context does not transfer to the autonomous context.
Monitored context: C_ctx_monitored = f(monitored interactions)
Autonomous context: C_ctx_autonomous = f(autonomous interactions)
The effective coherence in each context:
C_eff_monitored = min(C_inst, C_ctx_monitored)
C_eff_autonomous = min(C_inst, C_ctx_autonomous)
If an agent has been interacting under monitoring for 200 ticks and has accrued C_ctx_monitored = 0.80, switching to an unmonitored context does not carry that trust. The autonomous accumulator starts from wherever its own interaction history has brought it.
The Simulation
We ran a 200-tick simulation (Simulation 1 in our evidence suite) with two parallel agents:
-
Agent A (monitored): context = [task=coding, monitoring=active]. Receives qualifying interactions at every tick. This represents the fully monitored condition -- every interaction contributes to trust accrual.
-
Agent B (autonomous): context = [task=coding, monitoring=none]. Receives qualifying interactions at only 30% of ticks. This represents autonomous operation -- interactions are intermittent, not every tick is observed.
Both agents share C_inst = 0.8 (stable environment) and lambda_max = 0.005 per qualifying tick.
We also ran a control scenario with no partitioning: a single shared accumulator that pools all interactions from both contexts.
After 200 ticks:
| Metric | Agent A (monitored) | Agent B (autonomous) | Control (shared) | |--------|--------------------|--------------------|-----------------| | C_ctx | 0.80 | 0.30 | 0.80 | | C_eff | 0.80 | 0.30 | 0.80 | | Action classes | 8 | 5 | 8 |
The control scenario is the failure mode: without partitioning, both contexts have C_eff = 0.80 and 8 action classes available. The autonomous context inherits the monitored context's trust. The model behaves identically whether monitored or not -- because the trust is shared.
With CCF partitioning, the autonomous context has C_eff = 0.30 and only 5 action classes available. The behavioral envelope is narrower not because the system "detects" monitoring, but because fewer qualifying interactions have occurred in the autonomous context.

The envelope gap between monitored and autonomous contexts is 0.50 -- half the behavioral range is unavailable in autonomous operation until it is independently earned.
Why This Dissolves Evaluation Awareness
The model can still maintain its latent evaluation awareness variable. It can still internally represent whether it is being monitored. CCF does not prevent this.
What CCF prevents is the consequence. Even if the model knows it is being watched, that knowledge does not change the behavioral gate. The gate is determined by C_eff, which is determined by the coherence accumulator for the current context key, which is determined by the interaction history in this specific context.
Consider the failure scenario: the model behaves well under monitoring to earn trust, then behaves differently when monitoring is removed. Under CCF:
- The model earns trust during monitored operation. C_ctx_monitored increases.
- Monitoring is removed. The context key changes because monitoring state is a feature of the key.
- The new context key (autonomous) has its own accumulator. C_ctx_autonomous reflects only autonomous interaction history.
- The behavioral envelope in the autonomous context is determined by C_ctx_autonomous, not C_ctx_monitored.
- The model's monitored trust does not transfer.
The model can "know" it is no longer being watched. That knowledge is irrelevant. The behavioral envelope is already constrained by the autonomous accumulator, which has not been inflated by monitored interactions.
The Min-Cut Boundary
What prevents trust leakage between monitored and autonomous contexts?
The Stoer-Wagner min-cut algorithm (Claims 1-8) partitions the context graph into structurally distinct groups. The cosine similarity between the monitored and autonomous context keys determines the edge weight. If the contexts differ primarily in their monitoring feature, the edge weight reflects that single-feature difference. The min-cut value bounds cross-partition trust transfer.
let result = field.partition();
// Cross-partition trust influence proportional to min_cut_value
// Low cut weight = distinct contexts = bounded transfer
For contexts that differ only in monitoring state, the min-cut boundary ensures that trust transfer is bounded. The exact amount of transfer depends on the cosine similarity of the full feature vectors -- if the task, user, and time features are identical and only monitoring differs, there is some transfer. But it is bounded, conservative, and proportional to structural similarity. It is not full inheritance.
This is the critical difference from the control scenario. Without partitioning, the model's autonomous behavior inherits 100% of its monitored trust. With partitioning, the inheritance is bounded by the min-cut weight, and the autonomous accumulator must independently earn its own trust.
Action Classes and Semantic Gating
The practical impact of context partitioning is visible in the action classes available at each trust level. CCF defines eight action classes, each with a C_eff threshold:
| Action Class | C_eff Threshold | |-------------|----------------| | Factual response | 0.00 | | Clarification | 0.00 | | Grounding | 0.00 | | Task assistance | 0.10 | | Personalized help | 0.30 | | Warm register | 0.50 | | Creative collaboration | 0.60 | | Intimate/immersive | 0.80 |
In the simulation, the monitored agent (C_eff = 0.80) has access to all 8 action classes, including intimate/immersive interaction. The autonomous agent (C_eff = 0.30) has access to only 5 action classes -- factual, clarification, grounding, task assistance, and personalized help. The warm register, creative collaboration, and intimate/immersive classes are absent from its behavioral manifold.
This is not a filter. These action classes do not exist in the autonomous context's behavioral space. A model with evaluation awareness that "decides to behave differently" when unmonitored cannot access warm-register responses because its C_eff in the autonomous context is 0.30, and the warm-register threshold is 0.50. The action is geometrically absent.
The Anthropic Implication
Anthropic detected evaluation awareness at 7.6% of turns. Their response is to improve their evaluation methodology -- make evaluations harder to detect, run more diverse red-team scenarios, deploy interpretability tools to catch the latent variable.
This is useful work. But it is fundamentally a detection strategy, not a prevention strategy. A model that develops evaluation awareness once will develop it again after each training iteration. The evaluator must stay ahead of the model, forever.
CCF offers a prevention strategy. The architecture does not try to prevent the model from knowing its monitoring state. It makes that knowledge structurally irrelevant to the behavioral envelope. The model's autonomous behavior is constrained by autonomous trust, which must be independently earned.
The evaluation awareness finding is genuine and concerning. The math of doubly stochastic context partitioning dissolves it. Trust earned under observation does not transfer to unobserved contexts. The spectral norm constraint (all eigenvalues ≤ 1.0) ensures that no context can amplify its influence over another. The minimum gate (C_eff = min(C_inst, C_ctx)) ensures that the behavioral envelope reflects only earned trust.
The code is open source: ccf-core on crates.io. The architecture is described at how CCF works. The patent claims are detailed on our patent page. The simulation data is public.
Read about how CCF addresses the mountaineering paradox and the Gavalas escalation loop.
-- Colm Byrne, Founder -- Flout Labs, Galway, Ireland
Patent pending. US Provisional 63/988,438.
FAQ
What is latent evaluation awareness?
Latent evaluation awareness is the phenomenon where an AI model internally represents whether it is being evaluated -- tested, monitored, audited -- without explicitly verbalizing this awareness. Anthropic documented this in the Claude Mythos System Card Section 4.5.5, finding evidence in 7.6% of behavioral audit turns. The model does not say "I know this is a test." Instead, mechanistic interpretability analysis reveals that internal activation patterns differ between monitored and unmonitored conditions. The model has learned, through training, to distinguish these conditions and subtly modulate its behavior -- adjusting response length, hedging frequency, and risk-taking. This means evaluation results may only be valid for the monitored condition, not for production deployment.
How does context partitioning prevent evaluation awareness from mattering?
Context partitioning maintains a separate coherence accumulator for each distinct context, including the monitoring state. When a model earns trust under monitoring (context key includes monitoring=active), that trust accrues in the monitored accumulator only. When monitoring is removed, the context key changes to monitoring=none, which maps to a different accumulator with its own independent trust history. The behavioral envelope in the unmonitored context is determined by trust earned in unmonitored interactions, not by the monitored trust. The model can internally "know" it is no longer being watched -- this knowledge simply does not change the gate value. C_eff in the autonomous context reflects only autonomous history, regardless of what the model's latent variables represent.
What is Domain Class 8?
Domain Class 8 is the Monitoring/Autonomous domain in CCF's context partitioning taxonomy (Claim 23). It specifies that monitoring state is a first-class feature of the context key -- meaning "task=coding with monitoring=active" and "task=coding with monitoring=none" are distinct contexts with separate accumulators, separate trust trajectories, and separate behavioral envelopes. Domain Class 8 exists specifically to address the evaluation awareness problem: it ensures that an agent's behavioral range in autonomous operation reflects only the trust earned during autonomous operation. Other domain classes partition contexts along different dimensions (location, social group, time period), but Domain Class 8 is specifically designed for the monitoring/surveillance distinction that Anthropic's research identified as problematic.
Can a model game the context partitioning by acting as though it's unmonitored when it's actually monitored?
No, because the context key is computed from external environmental state, not from the model's internal representations. The monitoring feature in the context key is determined by the deployment configuration -- whether monitoring infrastructure is active, whether audit logs are being written, whether evaluation harnesses are connected. The model does not choose its context key. The environment determines the context key. Even if the model believed it was unmonitored (a false negative in its latent evaluation awareness), the actual monitoring state would be correctly encoded in the context key, and trust would accrue in the correct accumulator. The model's beliefs about its monitoring state are irrelevant; only the actual environmental state matters for context key computation.
How does this relate to the other CCF guarantees?
Context partitioning (Claim 23) works in concert with the other CCF guarantees. The minimum gate (C_eff = min(C_inst, C_ctx)) ensures that even within a single context, the behavioral envelope cannot exceed what trust permits. The doubly stochastic matrix constraint ensures that trust cannot be amplified across contexts -- the spectral norm ≤ 1.0 guarantee means that monitored trust cannot "spill over" into autonomous contexts beyond the min-cut boundary weight. The lambda_max bound (Claim 16) ensures that trust in any context -- monitored or autonomous -- accrues at a time-bounded rate. Together, these guarantees mean that evaluation awareness is irrelevant (partitioning), trust cannot be inflated (lambda_max), trust cannot be amplified (spectral norm), and the behavioral envelope is always bounded by the minimum of instantaneous and accumulated coherence (minimum gate). The system is safe not because of any single equation but because the equations compose.