The Model Evaluation & Threat Research (METR) initiative has released an example protocol for assessing AI models' potential for autonomy-related risks. This protocol focuses on systems capable of executing harmful tasks end-to-end without human bottlenecks, including those that autonomously procure human assistance. METR aims for the evaluation to be practical, cost-effective, and completed by a small team within a month, with a budget of a few million dollars. The goal is to provide a continuous metric of dangerous capabilities to inform mitigation strategies and allow for societal oversight. AI
Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →
IMPACT Provides a framework for evaluating AI autonomy risks, potentially guiding safety investments and development.
RANK_REASON This is a research paper detailing a new protocol for evaluating AI safety risks.