mcp-server-lancedb

MCP.Pizza Chef: kyryl-opens-ml

mcp-server-lancedb is an MCP server that integrates LanceDB, an embedded vector database, to provide a semantic memory layer for LLM applications. It enables storing text as vector embeddings and retrieving them later, facilitating efficient memory management and context enrichment for AI workflows. This server supports tools for adding and querying memories, making it ideal for AI-powered IDEs, chat interfaces, and custom workflows requiring persistent, semantic memory storage.

Use This MCP server To

Store text memories as vector embeddings for later retrieval Retrieve relevant memories based on semantic similarity queries Integrate LanceDB vector storage with LLM-powered applications Enable semantic search over stored text data Build AI workflows with persistent, context-rich memory layers

README

mcp-lance-db: A LanceDB MCP server

The Model Context Protocol (MCP) is an open protocol that enables seamless integration between LLM applications and external data sources and tools. Whether you're building an AI-powered IDE, enhancing a chat interface, or creating custom AI workflows, MCP provides a standardized way to connect LLMs with the context they need.

This repository is an example of how to create a MCP server for LanceDB, an embedded vector database.

Overview

A basic Model Context Protocol server for storing and retrieving memories in the LanceDB vector database. It acts as a semantic memory layer that allows storing text with vector embeddings for later retrieval.

Components

Tools

The server implements two tools:

  • add-memory: Adds a new memory to the vector database

    • Takes "content" as a required string argument
    • Stores the text with vector embeddings for later retrieval
  • search-memories: Retrieves semantically similar memories

    • Takes "query" as a required string argument
    • Optional "limit" parameter to control number of results (default: 5)
    • Returns memories ranked by semantic similarity to the query
    • Updates server state and notifies clients of resource changes

Configuration

The server uses the following configuration:

  • Database path: "./lancedb"
  • Collection name: "memories"
  • Embedding provider: "sentence-transformers"
  • Model: "BAAI/bge-small-en-v1.5"
  • Device: "cpu"
  • Similarity threshold: 0.7 (upper bound for distance range)

Quickstart

Claude Desktop

On MacOS: ~/Library/Application\ Support/Claude/claude_desktop_config.json On Windows: %APPDATA%/Claude/claude_desktop_config.json

{
  "lancedb": {
    "command": "uvx",
    "args": [
      "mcp-lance-db"
    ]
  }
}

Development

Building and Publishing

To prepare the package for distribution:

  1. Sync dependencies and update lockfile:
uv sync
  1. Build package distributions:
uv build

This will create source and wheel distributions in the dist/ directory.

  1. Publish to PyPI:
uv publish

Note: You'll need to set PyPI credentials via environment variables or command flags:

  • Token: --token or UV_PUBLISH_TOKEN
  • Or username/password: --username/UV_PUBLISH_USERNAME and --password/UV_PUBLISH_PASSWORD

Debugging

Since MCP servers run over stdio, debugging can be challenging. For the best debugging experience, we strongly recommend using the MCP Inspector.

You can launch the MCP Inspector via npm with this command:

npx @modelcontextprotocol/inspector uv --directory $(PWD) run mcp-lance-db

Upon launching, the Inspector will display a URL that you can access in your browser to begin debugging.

mcp-server-lancedb FAQ

How does mcp-server-lancedb store data?
It stores text as vector embeddings in the LanceDB embedded vector database for semantic retrieval.
What tools does this MCP server provide?
It provides tools like add-memory to store text and query-memory to retrieve relevant memories.
Can mcp-server-lancedb be used with multiple LLM providers?
Yes, it is compatible with any LLM supporting MCP, including OpenAI, Claude, and Gemini.
Is LanceDB embedded or external in this MCP server?
LanceDB is embedded within the server, enabling fast local vector storage and retrieval.
How does this server enhance AI workflows?
By providing a semantic memory layer, it allows LLMs to access and recall relevant past information efficiently.
What kind of data is best suited for storage in mcp-server-lancedb?
Textual data that benefits from semantic search and retrieval, such as chat logs, notes, or knowledge bases.
How do I add new memories to the LanceDB via this MCP server?
Use the add-memory tool with a text content argument to store new vector-embedded memories.
Can this MCP server handle real-time memory updates?
Yes, it supports adding and querying memories dynamically during runtime.