Pre Tokenization for Agent and Team

Is there any functionality available to cache the tokens of agent’s role and prompt so instead of passing all this thing along with the query and then it will converted to tokens so I can do like only pass the context , query and any dynamic details to the agent and it should saved cached tokens of system prompt , instruction and role.

So basically its for reducing the TPM to the LLM.
If there is any solution or functionality available then please let me know.

Hey @dp27, thanks for reaching out and supporting Agno. I’ve shared this with the team, we’re working through all requests one by one and will get back to you soon.If it’s urgent, please let us know. We appreciate your patience!

Hello @Monali Its been a while still I don’t got the response from your team can you give me something that I can refer to and have at least clarity to the functionality developed or not.

Hey @dp27, sorry for the delay. This ticket must have slipped through the cracks. We are really very sorry. @Ruan will be here to help you asap

In the meantime, can you pls help us with which model you are using?

Hey @dp27

Apologies for the delay here. The team has been overwhelmed with support requests.

Claude is the only model that has first class prompt caching support in our API.
You can see it in action here: