About
A lightweight MCP server that stores frequently accessed or computed data in memory, reducing repeated token consumption for language model interactions. It works seamlessly with any MCP client and offers configurable cache size, TTL, and cleanup intervals.
Capabilities

The Ibproduct Ib Mcp Cache Server is a lightweight, token‑saving MCP service that sits between your AI client and any language model. By caching data locally, it eliminates the need to repeatedly transmit large payloads—such as file contents or computed results—across the network, thereby cutting token usage and improving response times. This is especially valuable for developers who frequently query the same resources or perform repetitive calculations in their AI workflows.
At its core, the server intercepts data requests from an MCP client and stores them in an in‑memory cache keyed by the request parameters. When the same request is issued again, the server returns the cached result instead of forwarding the original payload to the model. Because the token count is tied directly to the amount of data sent, this mechanism can reduce token consumption by orders of magnitude for workloads that involve repeated file reads or heavy computations.
Key capabilities include:
- Configurable cache limits – Set maximum entries, memory usage, and default TTL to balance performance with resource constraints.
- Automatic eviction policies – Least‑recently‑used items are purged when limits are exceeded, and expired entries are cleaned up on a configurable interval.
- Transparent integration – Works with any MCP client and any model that tokenizes inputs; no changes to the client code are required.
- Real‑time statistics – Periodic updates of hit/miss rates help developers monitor cache effectiveness and fine‑tune settings.
Typical use cases span from data‑driven research (repeatedly querying the same dataset), to interactive debugging tools that re‑evaluate code snippets, and even chatbot backends that reuse common knowledge bases. In each scenario, the cache removes redundant data transfer, lowers latency, and keeps token budgets within limits.
By embedding this cache server into an AI workflow, developers gain a powerful, low‑overhead optimization that seamlessly scales with the complexity of their models and the frequency of their data accesses.
Related Servers
Data Exploration MCP Server
Turn CSVs into insights with AI-driven exploration
BloodHound-MCP
AI‑powered natural language queries for Active Directory analysis
Google Ads MCP
Chat with Claude to analyze and optimize Google Ads campaigns
Bazi MCP
AI‑powered Bazi calculator for accurate destiny insights
Smart Tree
Fast AI-friendly directory visualization with spicy terminal UI
Google Search Console MCP Server for SEOs
Chat‑powered SEO insights from Google Search Console
Weekly Views
Server Health
Information
Explore More Servers
LangChain MCP Client Streamlit App
Interactive LLM playground with multi‑provider, tool‑enabled, file‑aware chat
Unusual Whales MCP Server
MCP interface for Unusual Whales stock data API
Mcp Server Indexnow
Bridge MCP clients to IndexNow URL indexing
Backlog MCP Server
Integrate Backlog data into Claude with ease
Fireflies MCP Server
Unlock meeting insights with Fireflies transcript tools
Memory-Plus
Local RAG memory store for MCP agents