Control your Generative AI costs with the Vertex API’s context caching
If you were looking to optimize costs when using generative models on Vertex AI please, the post that I wrote with my colleague Nim Jayawardena, can be helpful. The post explores how to use Vertex AI’s context caching to reduce the cost of using Gemini models with large, repeated contexts. It shows criteria for using context caching together with the code samples. Since the platform does not support posting content with multiple authors, you can find the post in leoy.blog or in the Nim’s Medium blog.