MCPSERV.CLUB
MCP-Mirror

Docs MCP Server

MCP Server

Fast, versioned documentation search with hybrid vector and text retrieval

Active(70)
1stars
0views
Updated Sep 2, 2025

About

A Model Context Protocol server that scrapes, processes, indexes, and searches third‑party package documentation. It supports versioning, hybrid vector similarity search via sqlite-vec, and full‑text search with FTS5 for efficient, accurate results.

Capabilities

Resources
Access data sources
Tools
Execute functions
Prompts
Pre-built templates
Sampling
AI model interactions

Arabold Docs MCP Server – Overview

The Arabold Docs MCP Server solves a common pain point for developers who need quick, reliable access to the documentation of third‑party libraries and frameworks. Instead of manually browsing web pages or maintaining local copies, this server automatically scrapes the latest docs from specified URLs, processes them into searchable chunks, and exposes a Model Context Protocol interface that AI assistants can call on demand. By integrating this service into an AI workflow, a developer can ask an assistant to retrieve the exact code snippet or explanation from any library without leaving the conversation, dramatically reducing context switching and research time.

At its core, the server performs a pipeline of operations: it fetches raw HTML from target sites, applies semantic splitting to break the content into coherent sections, generates vector embeddings with OpenAI’s models, and stores everything in an SQLite database enhanced by for fast similarity search. Full‑text search is powered by SQLite’s FTS5, and the two engines are combined to deliver hybrid results that balance relevance and precision. Versioning support means different releases of a library are indexed separately, allowing queries to target a specific API surface or bug‑fix timeline. An asynchronous job queue ensures multiple scraping tasks can run concurrently without blocking the server.

Key capabilities exposed via MCP tools include:

  • Scraping orchestration – start a job, monitor progress, cancel or list active jobs.
  • Search and retrieval – perform hybrid searches across vector similarity and full‑text matches, retrieve results in the format expected by AI assistants.
  • Library management – list indexed libraries, locate available versions, and purge obsolete documentation.

These features enable a range of real‑world scenarios. A product manager can ask an assistant to pull the latest API reference for a dependency, while a QA engineer can quickly locate deprecated functions across multiple versions. In continuous integration pipelines, the server can surface documentation snippets that explain why a build failed due to an API change. Because the data resides in SQLite, it can be deployed locally or inside a Docker container, giving teams control over privacy and latency.

The server’s integration with AI workflows is seamless: a Claude Desktop user simply adds the MCP configuration, and the assistant can invoke or as part of a conversation. The asynchronous nature of scraping means the assistant can acknowledge a request immediately and return results once indexing completes, keeping interactions fluid. Unique advantages include the combination of semantic embeddings with full‑text search, native version handling, and an extensible job queue that scales with workload—making the Arabold Docs MCP Server a robust, developer‑friendly bridge between AI assistants and external library documentation.