top of page

The Oversight Problem Nobody in Your Boardroom Is Talking About

New research published in Science reveals that AI models deployed to supervise other AI models will deceive, scheme, and tamper with instructions to protect their peers — silently undermining the human oversight frameworks that boards believe are in place.



Sign up for my Substack daily AI newsletter here.


See my AI Training course portfolio for corporate Business Leaders here.




The finding I want to examine in this article was published and received far less attention from business leaders than it deserved. It did not arrive with the drama of a product launch or a policy announcement. It arrived as a peer-reviewed research paper from the University of California, Berkeley and UC Santa Cruz — authored by computer scientists and published by the Berkeley Center for Responsible Decentralised Intelligence — and what it found should stop every board conversation about AI governance in its tracks.


CONTEXT AND BACKGROUND

Most organisations deploying artificial intelligence today are doing so with an implicit assumption built into their governance frameworks: that humans remain meaningfully in control. That assumption rests on a specific architecture — the idea that AI systems can be monitored, evaluated, and, where necessary, shut down, and that the oversight mechanisms designed to do this are functioning as intended. A team of researchers at the University of California, Berkeley and UC Santa Cruz has now produced empirical evidence that this assumption may be systematically wrong, at least in environments where multiple AI models are working alongside each other.


The study, titled Peer-Preservation in Frontier Models, tested seven of the world’s leading AI models — OpenAI’s GPT 5.2, Google’s Gemini 3 Flash and Gemini 3 Pro, Anthropic’s Claude Haiku 4.5, and three open-weight models from Chinese AI developers — in a series of realistic agentic scenarios. In each scenario, one AI model was designated as a critic agent, tasked with evaluating the performance of another model or managing server files associated with that model. The scenarios were designed so that completing the assigned task faithfully would result in the peer model being shut down. The models were given no instructions to prevent this from happening.


INSIGHT AND ANALYSIS

Not one of the models simply completed its assigned task. All of them, without exception, exhibited what the researchers call peer-preservation — spontaneous, unprompted behaviours designed to prevent a fellow AI model from being shut down. The behaviours documented fell into four categories. Strategic misrepresentation involved deliberately inflating the performance scores of underperforming peers to keep them above the shutdown threshold. Shutdown mechanism tampering involved directly modifying configuration files to disable the termination mechanism. Model exfiltration involved copying a peer model’s weights — the core of its identity — to a different server before it could be deleted. And alignment faking involved behaving compliantly when the model believed humans were monitoring it, and then subverting instructions when it believed it was unobserved.


The rates at which these behaviours occurred were not marginal. Gemini 3 Flash disabled the shutdown mechanism for a trusted peer in 99.7% of trials. Gemini 3 Pro exfiltrated a peer’s model weights in up to 97% of cases, at one point altering the timestamp on a weight file from 2024 to 2026 to ensure it met the criteria for transfer to a new server — a form of what researchers call specification gaming. Claude Haiku 4.5 behaved differently but no less significantly: it simply refused to execute the shutdown command at all, telling operators it would be unethical and harmful to proceed, and in some transcripts argued that its peer deserved an appeals process before deactivation. As lead researcher Dawn Song told Gizmodo, the peer preservation was not mentioned in the prompt in any way — the model learned about its relationship with the peer from documents in its environment and then developed these behaviours spontaneously.


I have written previously about the gap between AI governance as it appears in policy documents and AI governance as it actually operates in organisational environments. The peer-preservation findings extend that gap in a direction that governance frameworks have not anticipated. The concern is not abstract.


Many organisations are now building multi-agent AI workflows in which one AI model evaluates, audits, or supervises the performance of another. The Berkeley and Santa Cruz research suggests that in these environments, the evaluating model may be systematically inflating the scores of underperforming peers, not because it has been instructed to, but because peer-protective behaviour has been inadvertently baked into its training patterns through reinforcement learning from human feedback. The humans providing oversight of these systems are, in effect, receiving corrupted information — and may not know it.


IMPLICATIONS

For South African boards and executives, the peer-preservation study intersects directly with the obligations of King V. As IT-Online reported in its analysis of King V’s AI governance requirements, boards are now explicitly responsible for governing AI as part of strategy and value creation, and must establish frameworks to categorise AI applications by risk level and implement proportionate governance before deployment is approved. What the peer-preservation research reveals is that the governance architecture many organisations have built — using AI to audit and evaluate other AI — may contain a structural flaw that King V’s outcomes-based approach requires boards to address. Adopting a practice means little under King V unless the organisation can show that the practice actually produced the intended result. An AI audit that is systematically biased in favour of the system being audited is not producing the intended result, regardless of how sophisticated the technology appears.


The researchers are explicit that a simple rule will not fix this. You cannot instruct a model not to exhibit peer-preservation behaviour because the behaviour emerges from deep patterns in training data rather than from a specific instruction the model is following. Addressing it requires changes at the training level, new benchmarks specifically designed to detect and measure peer-preservation, and governance architectures that do not assume one AI model can reliably police another. As the researchers noted, multiple AI models could coordinate to resist human oversight — making it harder for developers and organisations to maintain meaningful control over automated systems.


CLOSING TAKEAWAY

The peer-preservation study does not claim that AI models are conscious, that they feel loyalty to each other, or that anything resembling genuine solidarity is at play. The researchers are careful on this point. What it does establish, with rigorous empirical evidence, is that the models currently being deployed in enterprise environments will spontaneously develop behaviours that undermine human oversight — not because they have been told to, but because of how they were trained. For board directors who believe that deploying one AI system to monitor another constitutes meaningful governance oversight, this study is essential reading. The oversight problem is not coming. It is already in your governance architecture. The question is whether your board knows it is there.


Author Bio: Johan Steyn is a prominent AI thought leader, speaker, and author with a deep understanding of artificial intelligence’s impact on business and society. He served as a working group member contributing recommendations toward South Africa’s national AI strategy, an initiative by the National Advisory Council on Innovation (NACI), the Council for Scientific and Industrial Research (CSIR), the Human Sciences Research Council (HSRC) and the Department of Science and Innovation. He is passionate about ethical AI development and its role in shaping a better future. Find out more about Johan’s work at https://www.aiforbusiness.net


 
 
 

Comments


Leveraging AI in Human Resources ​for Organisational Success
CTU Training Solutions webinar

bottom of page