Fire in da houseTop Tip:Paying $100+ per month for Perplexity, MidJourney, Runway, ChatGPT and other tools is crazy - get all your AI tools in one site starting at $15 per month with Galaxy AI Fire in da houseCheck it out free

mcp-qdrant-memory

MCP.Pizza Chef: delorenj

The mcp-qdrant-memory is an MCP server that implements a knowledge graph with semantic search powered by the Qdrant vector database. It supports graph-based knowledge representation with entities and relations, file-based persistence, and semantic similarity search using OpenAI embeddings. It is deployable via Docker and supports HTTPS with reverse proxy compatibility, making it a robust solution for real-time, context-rich memory in AI workflows.

Use This MCP server To

Implement semantic search over knowledge graphs Persist and query AI memory with vector similarity Integrate knowledge graph with LLMs for context Deploy scalable memory server with Docker Use OpenAI embeddings for semantic similarity Support secure HTTPS access with reverse proxy

README

MCP Memory Server with Qdrant Persistence

smithery badge

This MCP server provides a knowledge graph implementation with semantic search capabilities powered by Qdrant vector database.

Features

  • Graph-based knowledge representation with entities and relations
  • File-based persistence (memory.json)
  • Semantic search using Qdrant vector database
  • OpenAI embeddings for semantic similarity
  • HTTPS support with reverse proxy compatibility
  • Docker support for easy deployment

Environment Variables

The following environment variables are required:

# OpenAI API key for generating embeddings
OPENAI_API_KEY=your-openai-api-key

# Qdrant server URL (supports both HTTP and HTTPS)
QDRANT_URL=https://your-qdrant-server

# Qdrant API key (if authentication is enabled)
QDRANT_API_KEY=your-qdrant-api-key

# Name of the Qdrant collection to use
QDRANT_COLLECTION_NAME=your-collection-name

Setup

Local Setup

  1. Install dependencies:
npm install
  1. Build the server:
npm run build

Docker Setup

  1. Build the Docker image:
docker build -t mcp-qdrant-memory .
  1. Run the Docker container with required environment variables:
docker run -d \
  -e OPENAI_API_KEY=your-openai-api-key \
  -e QDRANT_URL=http://your-qdrant-server:6333 \
  -e QDRANT_COLLECTION_NAME=your-collection-name \
  -e QDRANT_API_KEY=your-qdrant-api-key \
  --name mcp-qdrant-memory \
  mcp-qdrant-memory

Add to MCP settings:

{
  "mcpServers": {
    "memory": {
      "command": "/bin/zsh",
      "args": ["-c", "cd /path/to/server && node dist/index.js"],
      "env": {
        "OPENAI_API_KEY": "your-openai-api-key",
        "QDRANT_API_KEY": "your-qdrant-api-key",
        "QDRANT_URL": "http://your-qdrant-server:6333",
        "QDRANT_COLLECTION_NAME": "your-collection-name"
      },
      "alwaysAllow": [
        "create_entities",
        "create_relations",
        "add_observations",
        "delete_entities",
        "delete_observations",
        "delete_relations",
        "read_graph",
        "search_similar"
      ]
    }
  }
}

Tools

Entity Management

  • create_entities: Create multiple new entities
  • create_relations: Create relations between entities
  • add_observations: Add observations to entities
  • delete_entities: Delete entities and their relations
  • delete_observations: Delete specific observations
  • delete_relations: Delete specific relations
  • read_graph: Get the full knowledge graph

Semantic Search

  • search_similar: Search for semantically similar entities and relations
    interface SearchParams {
      query: string;     // Search query text
      limit?: number;    // Max results (default: 10)
    }

Implementation Details

The server maintains two forms of persistence:

  1. File-based (memory.json):

    • Complete knowledge graph structure
    • Fast access to full graph
    • Used for graph operations
  2. Qdrant Vector DB:

    • Semantic embeddings of entities and relations
    • Enables similarity search
    • Automatically synchronized with file storage

Synchronization

When entities or relations are modified:

  1. Changes are written to memory.json
  2. Embeddings are generated using OpenAI
  3. Vectors are stored in Qdrant
  4. Both storage systems remain consistent

Search Process

When searching:

  1. Query text is converted to embedding
  2. Qdrant performs similarity search
  3. Results include both entities and relations
  4. Results are ranked by semantic similarity

Example Usage

// Create entities
await client.callTool("create_entities", {
  entities: [{
    name: "Project",
    entityType: "Task",
    observations: ["A new development project"]
  }]
});

// Search similar concepts
const results = await client.callTool("search_similar", {
  query: "development tasks",
  limit: 5
});

HTTPS and Reverse Proxy Configuration

The server supports connecting to Qdrant through HTTPS and reverse proxies. This is particularly useful when:

  • Running Qdrant behind a reverse proxy like Nginx or Apache
  • Using self-signed certificates
  • Requiring custom SSL/TLS configurations

Setting up with a Reverse Proxy

  1. Configure your reverse proxy (example using Nginx):
server {
    listen 443 ssl;
    server_name qdrant.yourdomain.com;

    ssl_certificate /path/to/cert.pem;
    ssl_certificate_key /path/to/key.pem;

    location / {
        proxy_pass http://localhost:6333;
        proxy_set_header Host $host;
        proxy_set_header X-Real-IP $remote_addr;
    }
}
  1. Update your environment variables:
QDRANT_URL=https://qdrant.yourdomain.com

Security Considerations

The server implements robust HTTPS handling with:

  • Custom SSL/TLS configuration
  • Proper certificate verification options
  • Connection pooling and keepalive
  • Automatic retry with exponential backoff
  • Configurable timeouts

Troubleshooting HTTPS Connections

If you experience connection issues:

  1. Verify your certificates:
openssl s_client -connect qdrant.yourdomain.com:443
  1. Test direct connectivity:
curl -v https://qdrant.yourdomain.com/collections
  1. Check for any proxy settings:
env | grep -i proxy

Contributing

  1. Fork the repository
  2. Create a feature branch
  3. Make your changes
  4. Submit a pull request

License

MIT

mcp-qdrant-memory FAQ

How do I configure the mcp-qdrant-memory server?
Configure using environment variables for OpenAI API key, Qdrant URL, API key, and collection name as documented.
Can I deploy mcp-qdrant-memory with Docker?
Yes, it includes Docker support for easy deployment and scalability.
How does mcp-qdrant-memory perform semantic search?
It uses OpenAI embeddings combined with Qdrant's vector database for efficient semantic similarity search.
Is HTTPS supported by mcp-qdrant-memory?
Yes, it supports HTTPS and is compatible with reverse proxy setups for secure communication.
What kind of data persistence does mcp-qdrant-memory use?
It uses file-based persistence with a memory.json file alongside Qdrant vector storage.
Can mcp-qdrant-memory represent complex knowledge structures?
Yes, it supports graph-based knowledge representation with entities and relations.
What is required to authenticate with Qdrant?
You need to provide the Qdrant API key if authentication is enabled on your Qdrant server.
Which embeddings provider does mcp-qdrant-memory use?
It uses OpenAI embeddings but can be adapted to other providers like Claude or Gemini.