I've noticed something strange. I've been testing the new OpenAI Prompt Caching on its o1 mini models, using Portkey.
Every few requests, one of the requests has zero cache tokens. Even though it should have cache tokens because the message history is the same.
The previous message would have the cache, and the next one would have the cache too.
I'm wondering if this is an issue with OpenAI, or Portkey? Could it be that Portkey is sending the request from a different IP or something else that causes the cache to not work?
OpenAI doesn't give detailed enough logs for me to check what happened there :/