Researchers have developed a novel technique called Neural Garbage Collection (NGC) that enables language models to learn how to manage their own memory while reasoning. This method allows the model to decide which parts of its memory (KV cache) to evict during the reasoning process, optimizing efficiency without human intervention. NGC was trained solely on task outcome rewards and demonstrated significant KV cache compression, maintaining high accuracy on complex reasoning tasks like Countdown and AIME. AI
Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →
RANK_REASON The cluster describes a new academic paper detailing a novel technique for language models.