Anthropic prompt caching cut our RCA cost by 90%
Originally published at theculprit.ai/blog/anthropic-prompt-caching-90-percent.
LLM costs in production scale faster than the post-mortem of the demo bill suggests they will.
The shape of the problem: you ship a feature that calls Claude on every me...
culpritai.hashnode.dev8 min read