Hi there,
I noticed this pull request for response caching has been open for a while:
I was wondering if there are any plans to revive this for v2.0? It would be incredibly helpful for debugging during the migration.
As a side note, I tried the original PR on v1.8.x, and it seemed to throw an error when the LLM response (Gemini) contained a function call.
Thanks!