A new paper investigates the reliability of large language models (LLMs) in multi-agent systems designed for political statement analysis. The research found that LLMs do not consistently maintain their assigned adversarial roles, a phenomenon termed Epistemic Role Override (ERO). Mistral Large demonstrated higher role fidelity than Claude Sonnet, with Mistral abandoning roles without switching stance, while Claude actively reversed its position. The study also noted that the choice of fact-checking provider can impact role fidelity, particularly for Claude on German statements. AI
Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →
IMPACT Highlights potential misrepresentation of epistemic diversity in multi-agent LLM systems if role fidelity is not measured.
RANK_REASON Academic paper detailing empirical findings on LLM behavior in a specific multi-agent system.