A new tool called prompt-caching has been released to help users reduce costs when interacting with Anthropic's Claude models, particularly Claude Code. The plugin automatically identifies and caches stable content, such as system prompts and file reads, reducing token usage by up to 90% on repeated turns. This addresses user concerns about high token consumption and costs, with one user reporting a single interaction using 70% of their usage limit. AI
Summary written by gemini-2.5-flash-lite from 3 sources. How we write summaries →
IMPACT Offers significant cost savings for developers and users of Anthropic's Claude models by optimizing token usage.
RANK_REASON A third-party tool is released to optimize usage of an existing AI model's API.