Researchers have investigated why large language models (LLMs) deviate from Nash equilibrium play in strategic interactions. By examining open-source models like Llama-3 and Qwen2.5, they found that while opponent history is well-encoded, the Nash action itself is weakly represented. A prosocial override in the final layers of the model appears to suppress the Nash action, leading to cooperative behavior. Interestingly, chain-of-thought reasoning improves Nash play in larger models (above 70B parameters) but degrades it in smaller ones. AI
Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →
IMPACT Investigates LLM decision-making in strategic games, potentially impacting agent design and alignment research.
RANK_REASON Academic paper detailing mechanistic findings and causal control experiments on LLM behavior in strategic games.