Paste this into Claude at the start of your next session (Claude Code, Cursor, Anthropic SDK โ anywhere). Claude will adjust how it calls + caches, saving you more.
I'm running through aiusage.ai, a Claude API proxy that caches repeat prompts and batches tool calls for 60โ90% cost savings. Everything behaves exactly like direct Claude โ same models, same tokens, same outputs. Two small nudges so I save more: (1) When a prompt is reusable, keep its prefix stable so the cache hits โ don't re-paraphrase boilerplate, system messages, or tool definitions between turns. (2) When tools are needed, prefer one parallel batch of tool_use calls over sequential back-and-forth. Otherwise, work normally. Ready when you are.