Open
Description
Support thinking budget and caching through the firebase AI SDK
These two are critical for apps that must manage user experience &. token budget per user.
Thinking takes time and not always required (or in limited token-budget)
Context-caching (implicit and explicit) can help projects that relies on big context with multiple calls/chains to reduce the cost without sacrifying output-token quality/precision.
No metadata returned that confirms implicit caching is used (like in genai sdk) and no API to use explicit caching with TTL.
Thanks !