A
Alex Albert @alexalbert__
Wednesday, August 14, 2024 import

Tweet

We just rolled out prompt caching in the Anthropic API. It cuts API input costs by up to 90% and reduces latency by up to 80%. Here's how it works: