OpenAI has announced new initiatives to enhance AI safety through red teaming, a process of using people and AI to identify potential risks in new systems. The company is sharing two papers detailing their approach to external red teaming and introducing a new method for automated red teaming. Additionally, OpenAI is launching a Red Teaming Network to formally recruit domain experts from diverse backgrounds to collaborate on evaluating and improving the safety of their AI models throughout the development lifecycle. AI
Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →
RANK_REASON OpenAI published two papers and launched a network for external experts to conduct red teaming, which is a research and safety evaluation activity.