web content extractor.com
web content extractor.com logo

Web Content Extractor

Extracts and processes web content using TypeScript, Cheerio, and Turndown for tasks like scraping, summarization, and d...

Created byApr 23, 2025

MCP Webscan Server

[![smithery badge](https://smithery.ai/badge/mcp-server-webscan)](https://smithery.ai/server/mcp-server-webscan)
A Model Context Protocol (MCP) server for web content scanning and analysis. This server provides tools for fetching, analyzing, and extracting information from web pages.
<a href="https://glama.ai/mcp/servers/u0tna3hemh"><img width="380" height="200" src="https://glama.ai/mcp/servers/u0tna3hemh/badge" alt="Webscan Server MCP server" /></a>

Features

  • **Page Fetching**: Convert web pages to Markdown for easy analysis
  • **Link Extraction**: Extract and analyze links from web pages
  • **Site Crawling**: Recursively crawl websites to discover content
  • **Link Checking**: Identify broken links on web pages
  • **Pattern Matching**: Find URLs matching specific patterns
  • **Sitemap Generation**: Generate XML sitemaps for websites

Installation

Installing via Smithery

To install Webscan for Claude Desktop automatically via [Smithery](https://smithery.ai/server/mcp-server-webscan):

Manual Installation

Usage

Starting the Server

The server runs on stdio transport, making it compatible with MCP clients like Claude Desktop.

Available Tools

  1. `fetch-page` - Fetches a web page and converts it to Markdown. - Parameters: - `url` (required): URL of the page to fetch. - `selector` (optional): CSS selector to target specific content.
  1. `extract-links` - Extracts all links from a web page with their text. - Parameters: - `url` (required): URL of the page to analyze. - `baseUrl` (optional): Base URL to filter links. - `limit` (optional, default: 100): Maximum number of links to return.
  1. `crawl-site` - Recursively crawls a website up to a specified depth. - Parameters: - `url` (required): Starting URL to crawl. - `maxDepth` (optional, default: 2): Maximum crawl depth (0-5).
  1. `check-links` - Checks for broken links on a page. - Parameters: - `url` (required): URL to check links for.
  1. `find-patterns` - Finds URLs matching a specific pattern. - Parameters: - `url` (required): URL to search in. - `pattern` (required): JavaScript-compatible regex pattern to match URLs against.
  1. `generate-site-map` - Generates a simple XML sitemap by crawling. - Parameters: - `url` (required): Root URL for sitemap crawl. - `maxDepth` (optional, default: 2): Maximum crawl depth for discovering URLs (0-5). - `limit` (optional, default: 1000): Maximum number of URLs to include in the sitemap.

Example Usage with Claude Desktop

  1. Configure the server in your Claude Desktop settings:
  1. Use the tools in your conversations:

Development

Prerequisites

  • Node.js >= 18
  • npm

Project Structure (Post-Refactor)

Building

Development Mode

Error Handling

The server implements comprehensive error handling:
  • Invalid parameters
  • Network errors
  • Content parsing errors
  • URL validation
All errors are properly formatted according to the MCP specification.

Contributing

  1. Fork the repository
  1. Create your feature branch (`git checkout -b feature/amazing-feature`)
  1. Commit your changes (`git commit -m 'Add some amazing feature'`)
  1. Push to the branch (`git push origin feature/amazing-feature`)
  1. Open a Pull Request

License

MIT License - see the LICENSE file for details