Researchers have investigated the effectiveness of many-shot chain-of-thought in-context learning (CoT-ICL) for reasoning tasks, finding that standard many-shot approaches do not directly translate. Their study revealed that increasing CoT demonstrations can be unstable for non-reasoning models and primarily benefits reasoning-oriented LLMs. The research also indicated that similarity-based retrieval is effective for non-reasoning tasks but not for reasoning, and that performance variance increases with more CoT examples. To address these issues, they propose Curvilinear Demonstration Selection (CDS), an ordering method that improves performance by treating demonstrations as a structured curriculum for in-context test-time learning. AI
Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →
IMPACT Reframes in-context learning as test-time learning, suggesting new methods for ordering demonstrations to improve LLM reasoning.
RANK_REASON Academic paper detailing a new method for in-context learning. [lever_c_demoted from research: ic=1 ai=1.0]