Tag: Claude
-

-

How Prompt Caching Cuts AI Costs by 90%
The 90% Discount Most API Users Never Claim Anthropic’s cache cuts API costs by 90% — yet most developers sending requests to Claude, GPT, or Gemini have never configured it. Prompt caching, which Anthropic launched in July 2024, reduces input token costs from $3 per million to $0.30 per million for cached portions on Claude…
-

How MCP Servers Exchange Tools With Claude
Claude doesn’t actually call your tools directly. When you type a message in Claude Desktop and it responds by reading a file, querying a database, or hitting an API, there’s an entire protocol running behind the scenes that brokers every single tool call. That protocol is the Model Context Protocol (MCP), and understanding how it…