Gemini caching is now live!
75% off prompt token prices, and we made the API extremely simple:
Just set *cache_control* on a message. Exactly the same as Anthropic, and others to come👇
Context caching with Gemini is so good!
Here I am caching the entire Gemini Cookbook (around 400k tokens) as an insanely long prompt to create the best Gemini app developer on the planet.
Watch Gemini answer any coding questions related to its own APIs. https://t.co/Azo82GqZIp
Today we are making URL Context, my favorite Gemini API tool, ready for scaled production use 🔗
The model can now visit webpages, PDF's, images, and more when you provide the direct URL, and you simply pay for the tokens it processes, no additional tool cost! https://t.co/ukuev45pJg