Factor(T,U): Factored Cognition Strengthens Monitoring of Untrusted AI
Published on arXiv
2512.02157
Excessive Agency
OWASP LLM Top 10 — LLM08
Key Finding
Adding factored cognition to a trusted monitoring protocol boosts safety from 41% to 63% against adversarial backdoor-inserting LLM in competitive programming tasks.
Factor(T,U)
Novel technique introduced
The field of AI Control seeks to develop robust control protocols, deployment safeguards for untrusted AI which may be intentionally subversive. However, existing protocols that rely on weaker monitors to detect unsafe behavior often fail on complex tasks beyond the monitor's comprehension. We develop control protocols based on factored cognition, in which a trusted model decomposes a task, an untrusted model solves each resultant child task in isolation, and the results are reassembled into a full solution. These protocols may improve safety by several means, such as by simplifying the context for monitors, or by obscuring vulnerabilities in the environment. We implement our protocols in the APPS coding setting and red team them against backdoor attempts from an adversarial GPT-4.1 Nano. We find that: (i) Adding factored cognition to a trusted monitoring protocol can boost safety from 41% to 63%; (ii) Safety improves because monitor performance improves; (iii) Factored cognition makes it no harder for capable LLMs to write backdoors in APPS. While our protocols show low usefulness in APPS, they hold promise for more complex tasks.
Key Contributions
- Introduces Factor(T,U), a factored cognition control protocol that decomposes tasks via a trusted model so an untrusted model only sees isolated child tasks, reducing the attack surface for intentional subversion.
- Demonstrates that Factor(D,U) boosts safety from 41% to 63% over trusted monitoring alone in the APPS coding setting, with evidence that the gain comes from improved monitor performance on simpler contexts.
- Red-teams the protocol against an adversarial GPT-4.1 Nano attempting to insert backdoors, finding that factored cognition does not make backdoor writing harder for capable LLMs but does make detection easier.