Anthropic, a company specializing in developing AI models and apps, has launched prompt caching for its Claude API in a public beta.
The feature allows developers to cache frequently used context between API calls for Claude 3.5 Sonnet and Claude 3 Haiku models, with support for Claude 3 Opus coming soon. Moreover, the pricing for cached prompts is based on the number of input tokens cached and usage frequency.
Prompt caching enables developers to provide Claude with more background knowledge and example outputs. Furthermore, it is claimed to be effective for conversational agents, coding assistants, large document processing, detailed instruction sets, agentic search and tool use, and interacting with long-form content.
By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.