Response caching allows you to cache model responses, which can significantly improve response times and reduce API costs during development and testing.Documentation Index
Fetch the complete documentation index at: https://agno-v2-shaloo-ai-support-link.mintlify.app/llms.txt
Use this file to discover all available pages before exploring further.
For a detailed overview of response caching, see Response Caching.
This is different from Anthropic’s prompt caching feature. Response caching caches the entire model response, while prompt caching caches the system prompt to reduce processing time.
Basic Usage
Enable caching by settingcache_response=True when initializing the model. The first call will hit the API and cache the response, while subsequent identical calls will return the cached result.
cache_model_response.py