Prompt caching: 10x cheaper LLM tokens, but how? | ngrok blog
A far more detailed explanation of prompt caching than anyone asked for.
A far more detailed explanation of prompt caching than anyone asked for.

How I'm learning to build with better software architecture design principles (while 'moving fast and breaking things').