Fire in da houseTop Tip:Most people pay up to $340 per month for Perplexity, MidJourney, Runway, ChatGPT, and more - but you can get them all your AI tools for $15 with Galaxy. It's free to test!Fire in da houseCheck it out

deep-research-mcp

MCP.Pizza Chef: Ozamatash

Deep Research MCP Server is an AI-driven research assistant that integrates search engines, web scraping, and AI to perform deep, iterative exploration of topics. It generates detailed, comprehensive reports by continuously refining queries and aggregating information. Available as an MCP server or standalone CLI, it enables seamless integration with LLM hosts like Claude Desktop for enhanced research workflows.

Use This MCP server To

Generate comprehensive research reports on complex topics Automate iterative web searches and data aggregation Integrate with LLM hosts for real-time research context Scrape and summarize web content for detailed insights Support academic or market research with AI-driven exploration Provide structured research data for downstream AI tasks Run research queries via CLI for quick information retrieval

README

Open Deep Research MCP Server

An AI-powered research assistant that performs deep, iterative research on any topic. It combines search engines, web scraping, and AI to explore topics in depth and generate comprehensive reports. Available as a Model Context Protocol (MCP) tool or standalone CLI. Look at exampleout.md to see what a report might look like.

Quick Start

  1. Clone and install:
git clone https://github.com/Ozamatash/deep-research
cd deep-research
npm install
  1. Set up environment in .env.local:
# Copy the example environment file
cp .env.example .env.local
  1. Build:
# Build the server
npm run build
  1. Run the cli version:
npm run start "Your research query here"
  1. Test MCP Server with Claude Desktop:
    Follow the guide thats at the bottom of server quickstart to add the server to Claude Desktop:
    https://modelcontextprotocol.io/quickstart/server

Features

  • Performs deep, iterative research by generating targeted search queries
  • Controls research scope with depth (how deep) and breadth (how wide) parameters
  • Evaluates source reliability with detailed scoring (0-1) and reasoning
  • Prioritizes high-reliability sources (≥0.7) and verifies less reliable information
  • Generates follow-up questions to better understand research needs
  • Produces detailed markdown reports with findings, sources, and reliability assessments
  • Available as a Model Context Protocol (MCP) tool for AI agents
  • For now MCP version doesn't ask follow up questions

How It Works

flowchart TB
    subgraph Input
        Q[User Query]
        B[Breadth Parameter]
        D[Depth Parameter]
        FQ[Feedback Questions]
    end

    subgraph Research[Deep Research]
        direction TB
        SQ[Generate SERP Queries]
        SR[Search]
        RE[Source Reliability Evaluation]
        PR[Process Results]
    end

    subgraph Results[Research Output]
        direction TB
        L((Learnings with
        Reliability Scores))
        SM((Source Metadata))
        ND((Next Directions:
        Prior Goals,
        New Questions))
    end

    %% Main Flow
    Q & FQ --> CQ[Combined Query]
    CQ & B & D --> SQ
    SQ --> SR
    SR --> RE
    RE --> PR

    %% Results Flow
    PR --> L
    PR --> SM
    PR --> ND

    %% Depth Decision and Recursion
    L & ND --> DP{depth > 0?}
    DP -->|Yes| SQ
    
    %% Final Output
    DP -->|No| MR[Markdown Report]

    %% Styling
    classDef input fill:#7bed9f,stroke:#2ed573,color:black
    classDef process fill:#70a1ff,stroke:#1e90ff,color:black
    classDef output fill:#ff4757,stroke:#ff6b81,color:black
    classDef results fill:#a8e6cf,stroke:#3b7a57,color:black,width:150px,height:150px

    class Q,B,D,FQ input
    class SQ,SR,RE,PR process
    class MR output
    class L,SM,ND results
Loading

Advanced Setup

Using Local Firecrawl (Free Option)

Instead of using the Firecrawl API, you can run a local instance. You can use the official repo or my fork which uses searXNG as the search backend to avoid using a searchapi key:

  1. Set up local Firecrawl:
git clone https://github.com/Ozamatash/localfirecrawl
cd localfirecrawl
# Follow setup in localfirecrawl README
  1. Update .env.local:
FIRECRAWL_BASE_URL="http://localhost:3002"

Optional: Observability

Add observability to track research flows, queries, and results using Langfuse:

# Add to .env.local
LANGFUSE_PUBLIC_KEY="your_langfuse_public_key"
LANGFUSE_SECRET_KEY="your_langfuse_secret_key"

The app works normally without observability if no Langfuse keys are provided.

License

MIT License

deep-research-mcp FAQ

How do I install the Deep Research MCP Server?
Clone the GitHub repo, install dependencies with npm, set up your .env.local file, build the server, and run the CLI or integrate with an MCP host.
Can I use the Deep Research MCP Server without an MCP host?
Yes, it can be used as a standalone CLI tool for research queries.
How does the server perform deep research?
It combines search engines, web scraping, and AI to iteratively explore and refine research topics.
Is the Deep Research MCP Server compatible with multiple LLM providers?
Yes, it works with various LLM hosts including Claude, OpenAI, and Gemini via the MCP protocol.
What kind of reports does the server generate?
It produces comprehensive, structured reports that summarize findings from multiple sources.
How do I integrate this server with Claude Desktop?
Follow the guide in the MCP quickstart documentation to add the server to Claude Desktop.
Can I customize the research queries or scraping parameters?
Yes, the server supports configuration through environment variables and code customization.
Does the server support real-time updates during research?
Yes, it iteratively updates research results to provide the most current information.