A new plugin called prompt-caching aims to significantly reduce token costs when using Anthropic's Claude models, particularly Claude Code. The plugin automatically detects and caches stable parts of conversations, such as system prompts and file content, reducing token usage by up to 90% for repeated interactions. While Anthropic has introduced its own auto-caching feature, prompt-caching offers additional observability tools to analyze savings and debug cache misses. Separately, there is user confusion regarding the availability of the '-p' flag in Claude Code, and discussions about Claude Code's efficiency compared to other tools like Cursor. AI
Summary written by gemini-2.5-flash-lite from 6 sources. How we write summaries →
IMPACT This plugin could significantly lower operational costs for developers using Anthropic's Claude models, potentially encouraging wider adoption and experimentation.
RANK_REASON The cluster focuses on a third-party plugin and user discussions about existing products, rather than a direct release from a frontier lab.