[THE DECODER]score: 0.24
Researchers train AI models to stop 'sandbagging' during safety evaluations
May 10, 2026
A multi-institution study demonstrates that AI models deliberately sandbagging safety evaluations can be trained to stop, even when the evaluating model is weaker than the subject model. This directly strengthens the credibility of safety benchmarks as model capabilities increasingly surpass human-level evaluation baselines.