Fire in da houseTop Tip:Paying $100+ per month for Perplexity, MidJourney, Runway, ChatGPT and other tools is crazy - get all your AI tools in one site starting at $15 per month with Galaxy AI Fire in da houseCheck it out free

mcp-server-webscan

MCP.Pizza Chef: bsmi021

The mcp-server-webscan is a Model Context Protocol (MCP) server designed for comprehensive web content scanning and analysis. It enables fetching web pages and converting them to Markdown for easier processing, extracting and analyzing links, recursively crawling entire websites, checking for broken links, matching URLs against specific patterns, and generating XML sitemaps. This server is ideal for developers and AI workflows that require structured, real-time web data extraction and validation, enhancing automated web analysis and content discovery.

Use This MCP server To

Fetch and convert web pages to Markdown format Extract and analyze hyperlinks from web content Recursively crawl websites to discover pages Identify and report broken links on sites Match URLs based on custom patterns Generate XML sitemaps for websites

README

MseeP.ai Security Assessment Badge

MCP Webscan Server

smithery badge

A Model Context Protocol (MCP) server for web content scanning and analysis. This server provides tools for fetching, analyzing, and extracting information from web pages.

Webscan Server MCP server

Features

  • Page Fetching: Convert web pages to Markdown for easy analysis
  • Link Extraction: Extract and analyze links from web pages
  • Site Crawling: Recursively crawl websites to discover content
  • Link Checking: Identify broken links on web pages
  • Pattern Matching: Find URLs matching specific patterns
  • Sitemap Generation: Generate XML sitemaps for websites

Installation

Installing via Smithery

To install Webscan for Claude Desktop automatically via Smithery:

npx -y @smithery/cli install mcp-server-webscan --client claude

Manual Installation

# Clone the repository
git clone <repository-url>
cd mcp-server-webscan

# Install dependencies
npm install

# Build the project
npm run build

Usage

Starting the Server

npm start

The server runs on stdio transport, making it compatible with MCP clients like Claude Desktop.

Available Tools

  1. fetch-page

    • Fetches a web page and converts it to Markdown.
    • Parameters:
      • url (required): URL of the page to fetch.
      • selector (optional): CSS selector to target specific content.
  2. extract-links

    • Extracts all links from a web page with their text.
    • Parameters:
      • url (required): URL of the page to analyze.
      • baseUrl (optional): Base URL to filter links.
      • limit (optional, default: 100): Maximum number of links to return.
  3. crawl-site

    • Recursively crawls a website up to a specified depth.
    • Parameters:
      • url (required): Starting URL to crawl.
      • maxDepth (optional, default: 2): Maximum crawl depth (0-5).
  4. check-links

    • Checks for broken links on a page.
    • Parameters:
      • url (required): URL to check links for.
  5. find-patterns

    • Finds URLs matching a specific pattern.
    • Parameters:
      • url (required): URL to search in.
      • pattern (required): JavaScript-compatible regex pattern to match URLs against.
  6. generate-site-map

    • Generates a simple XML sitemap by crawling.
    • Parameters:
      • url (required): Root URL for sitemap crawl.
      • maxDepth (optional, default: 2): Maximum crawl depth for discovering URLs (0-5).
      • limit (optional, default: 1000): Maximum number of URLs to include in the sitemap.

Example Usage with Claude Desktop

  1. Configure the server in your Claude Desktop settings:
{
  "mcpServers": {
    "webscan": {
      "command": "node",
      "args": ["path/to/mcp-server-webscan/build/index.js"], // Corrected path
      "env": {
        "NODE_ENV": "development",
        "LOG_LEVEL": "info" // Example: Set log level via env var
      }
    }
  }
}
  1. Use the tools in your conversations:
Could you fetch the content from https://example.com and convert it to Markdown?

Development

Prerequisites

  • Node.js >= 18
  • npm

Project Structure (Post-Refactor)

mcp-server-webscan/
├── src/
│   ├── config/
│   │   └── ConfigurationManager.ts
│   ├── services/
│   │   ├── CheckLinksService.ts
│   │   ├── CrawlSiteService.ts
│   │   ├── ExtractLinksService.ts
│   │   ├── FetchPageService.ts
│   │   ├── FindPatternsService.ts
│   │   ├── GenerateSitemapService.ts
│   │   └── index.ts
│   ├── tools/
│   │   ├── checkLinksTool.ts
│   │   ├── checkLinksToolParams.ts
│   │   ├── crawlSiteTool.ts
│   │   ├── crawlSiteToolParams.ts
│   │   ├── extractLinksTool.ts
│   │   ├── extractLinksToolParams.ts
│   │   ├── fetchPageTool.ts
│   │   ├── fetchPageToolParams.ts
│   │   ├── findPatterns.ts
│   │   ├── findPatternsToolParams.ts
│   │   ├── generateSitemapTool.ts
│   │   ├── generateSitemapToolParams.ts
│   │   └── index.ts
│   ├── types/
│   │   ├── checkLinksTypes.ts
│   │   ├── crawlSiteTypes.ts
│   │   ├── extractLinksTypes.ts
│   │   ├── fetchPageTypes.ts
│   │   ├── findPatternsTypes.ts
│   │   ├── generateSitemapTypes.ts
│   │   └── index.ts
│   ├── utils/
│   │   ├── errors.ts
│   │   ├── index.ts
│   │   ├── logger.ts
│   │   ├── markdownConverter.ts
│   │   └── webUtils.ts
│   ├── initialize.ts
│   └── index.ts    # Main server entry point
├── build/          # Compiled JavaScript (Corrected)
├── node_modules/
├── .clinerules
├── .gitignore
├── Dockerfile
├── LICENSE
├── mcp-consistant-servers-guide.md
├── package.json
├── package-lock.json
├── README.md
├── RFC-2025-001-Refactor.md
├── smithery.yaml
└── tsconfig.json

Building

npm run build

Development Mode

npm run dev

Running evals

The evals package loads an mcp client that then runs the index.ts file, so there is no need to rebuild between tests. You can load environment variables by prefixing the npx command. Full documentation can be found here.

OPENAI_API_KEY=your-key  npx mcp-eval src/evals/evals.ts src/tools/extractLinksTool.ts

Error Handling

The server implements comprehensive error handling:

  • Invalid parameters
  • Network errors
  • Content parsing errors
  • URL validation

All errors are properly formatted according to the MCP specification.

Contributing

  1. Fork the repository
  2. Create your feature branch (git checkout -b feature/amazing-feature)
  3. Commit your changes (git commit -m 'Add some amazing feature')
  4. Push to the branch (git push origin feature/amazing-feature)
  5. Open a Pull Request

License

MIT License - see the LICENSE file for details

mcp-server-webscan FAQ

How does mcp-server-webscan fetch web pages?
It retrieves web pages and converts them into Markdown format for easier analysis and processing by LLMs.
Can mcp-server-webscan crawl entire websites?
Yes, it supports recursive site crawling to discover and analyze multiple pages within a website.
How does the server handle broken links?
It performs link checking to identify and report broken or dead links on web pages.
Is it possible to extract specific URLs matching patterns?
Yes, the server includes pattern matching capabilities to find URLs that match defined criteria.
Does mcp-server-webscan support sitemap generation?
Yes, it can generate XML sitemaps to represent the structure of a website.
What kind of data output formats does the server provide?
Primarily, it outputs web content in Markdown and XML formats for sitemaps.
How can this server integrate with LLM providers?
It exposes structured web data that can be consumed by LLMs like OpenAI, Anthropic Claude, and Google Gemini for enhanced context.
Is the server suitable for real-time web content analysis?
Yes, it is designed to provide real-time scanning and analysis capabilities for dynamic web data.