Black lighthouse silhouette graphic – visual asset for The Midas Project Watchtower page.

Watchtower

Watchtower tracks changes to corporate and government AI safety policies, both announced and unannounced. Click any entry for details.

< Back

Date:

Aug 28, 2025

Aug 28, 2025

Aug 28, 2025

xAI

Violation

Major

Unannounced

xAI's risk management framework establishes criteria that models must meet before they are released. One such criterion is that models must not lie more than 50% of the time on the MASK benchmark (Model Alignment Between Statements and Knowledge), which measures how often models lie under pressure. Grok Code Fast 1 scored 71.9% on this test—well above the release threshold—but xAI released it anyway. The company justified the release by noting the model is less capable than Grok 4 and is designed for coding rather than general-purpose use. AI Lab Watch called this "the most brazen violation" of a safety policy to date.