Anthropic, a San Francisco-based AI startup founded by researchers who broke away from OpenAI, has published an overview of its red-teaming practices, outlining four approaches and their advantages and disadvantages. Red teaming, a security practice of attacking one’s own system to uncover and address potential security vulnerabilities, has taken on a prominent role in discussions of AI regulation. The Biden administration’s AI executive order mandates that companies developing high-risk foundation models notify the government during training and share all red teaming results, while the EU AI Act also contains requirements around providing information from red teaming. Anthropic’s approaches include using language models to red team, red teaming in multiple modalities, domain-specific expert red teaming, and open-ended, general red teaming. The company concludes with policy recommendations, including suggestions to fund and encourage third-party red teaming, and to create clear policies tying the scaling of development and release of new models with red teaming results. As lawmakers rally around red teaming as a way to ensure powerful AI models are developed safely, it certainly deserves a close eye.

AI Red Teaming Takes Center Stage
Anthropic employs a red team/blue team dynamic, where it uses a model to generate attacks that are likely to elicit the target behavior (red team) and then fine-tunes a model on those red-teamed outputs in order to make it more robust to similar types of attack (blue team).
1–2 minutes










