A new plugin called prompt-caching has been released that significantly reduces token costs when using Anthropic's Claude models, particularly for developers. The plugin automatically identifies and caches stable content like system prompts and file reads, lowering costs by up to 90% on repeated interactions. While Anthropic has introduced its own auto-caching feature, prompt-caching offers enhanced observability and can be applied to custom applications built with the Anthropic SDK, addressing a different layer of cost optimization. AI
Summary written by gemini-2.5-flash-lite from 5 sources. How we write summaries →
IMPACT Developers can significantly reduce their Claude API costs by using this plugin for applications and agents.
RANK_REASON This is a third-party tool designed to optimize the use of an existing AI model's API, rather than a release of a new model or core technology.