Prompt caching: 10x cheaper LLM tokens, but how? | ngrok blog
A far more detailed explanation of prompt caching than anyone asked for.
blogs.nvidia.com
How to Fine-Tune an LLM on NVIDIA GPUs With Unsloth
Fine-tune popular AI models faster in Unsloth with NVIDIA RTX AI PCs and DGX Spark to build personalized assistants for studying, work, creative proje...
Related Articles
Attempting to Design a Back-end with Cleaner Architecture Rules and Boundaries
How I'm learning to build with better software architecture design principles (while 'moving fast and breaking things').
How to Secure Environment Variables for LLMs, MCPs, and AI Tools Using 1Password or Doppler
Stop hardcoding API keys in MCP configs and AI tool settings. Learn how to use 1Password CLI or Doppler to inject secrets just-in-time for Claude, Cur...
Adding a GitHub Contribution Graph to Next.js
How to add a GitHub contribution graph to your Next.js site using GitHub's GraphQL API, with server-side caching.
Related Projects
Book Finder (findmybook.net)
Book search and recommendation engine with OpenAI integration
Related Books
Advanced Algorithms and Data Structures
Marcello La Rocca
React Quickly, Second Edition
Azat Mardan, Morten Barklund
Learn the skills you need to build React applications quickly! This fully revised second edition introduces JSX, functional components, React hooks, e...
Just Use Postgres!
Denis Magda
You probably don’t need a collection of specialty databases. Just use Postgres instead! Need a fast, reliable SQL-compliant RDBMS? Just use Postgres! ...