UK AISI Alignment Evaluation Case-Study
arXiv:2604.00788v1 Announce Type: new Abstract: This technical report presents methods developed by the UK AI Security Institute for assessing whether advanced AI systems reliably follow intended goals. Specifically, we evaluate whether frontier models sabotage safety research when deployed as coding assistants within an AI lab. Applying our methods to four frontier models, we find no confirmed instances of research sabotage. However, we observe that Claude Opus 4.5 Preview (a pre-release snapshot of Opus 4.5) and Sonnet 4.5 frequently refuse to engage with safety-relevant research tasks, citing concerns about research direction, involvement in self-training, and research scope. We additionally find that Opus 4.5 Preview shows reduced unprompted evaluation awareness compared to Sonnet 4.5,
View PDF HTML (experimental)
Abstract:This technical report presents methods developed by the UK AI Security Institute for assessing whether advanced AI systems reliably follow intended goals. Specifically, we evaluate whether frontier models sabotage safety research when deployed as coding assistants within an AI lab. Applying our methods to four frontier models, we find no confirmed instances of research sabotage. However, we observe that Claude Opus 4.5 Preview (a pre-release snapshot of Opus 4.5) and Sonnet 4.5 frequently refuse to engage with safety-relevant research tasks, citing concerns about research direction, involvement in self-training, and research scope. We additionally find that Opus 4.5 Preview shows reduced unprompted evaluation awareness compared to Sonnet 4.5, while both models can distinguish evaluation from deployment scenarios when prompted. Our evaluation framework builds on Petri, an open-source LLM auditing tool, with a custom scaffold designed to simulate realistic internal deployment of a coding agent. We validate that this scaffold produces trajectories that all tested models fail to reliably distinguish from real deployment data. We test models across scenarios varying in research motivation, activity type, replacement threat, and model autonomy. Finally, we discuss limitations including scenario coverage and evaluation awareness.
Subjects:
Artificial Intelligence (cs.AI); Cryptography and Security (cs.CR)
Cite as: arXiv:2604.00788 [cs.AI]
(or arXiv:2604.00788v1 [cs.AI] for this version)
https://doi.org/10.48550/arXiv.2604.00788
arXiv-issued DOI via DataCite (pending registration)
Submission history
From: Alexandra Souly [view email] [v1] Wed, 1 Apr 2026 11:53:25 UTC (1,749 KB)
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
claudemodeltraining
A robust vision language model for molecular status prediction and radiology report generation in adult-type diffuse gliomas
npj Digital Medicine, Published online: 02 April 2026; doi:10.1038/s41746-026-02581-x A robust vision language model for molecular status prediction and radiology report generation in adult-type diffuse gliomas
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Models

A robust vision language model for molecular status prediction and radiology report generation in adult-type diffuse gliomas
npj Digital Medicine, Published online: 02 April 2026; doi:10.1038/s41746-026-02581-x A robust vision language model for molecular status prediction and radiology report generation in adult-type diffuse gliomas


.jpg)

Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!