Files
2025-11-29 17:57:56 +08:00

1.9 KiB

allowed-tools, description, model
allowed-tools description model
mcp__mcp-server-firecrawl__firecrawl_scrape, Write, Bash Scrape websites using Firecrawl MCP and save content to research folders claude-sonnet-4-5-20250929

Scrape Site

This command scrapes websites using the Firecrawl MCP and intelligently saves the content to organized research folders within the desktop-commander documentation system.

$ARGUMENTS

Usage Examples:

  • /scrape-site https://docs.anthropic.com/claude/guide - Scrape and auto-organize in research folder
  • /scrape-site https://example.com/api "api-docs" - Scrape and save to specific subfolder
  • /scrape-site https://github.com/owner/repo/wiki "github-wiki" - Save with custom folder name

Instructions

  • Extract the URL from $ARGUMENTS (first argument is always the URL to scrape)
  • If second argument provided, use it as the subfolder name; otherwise auto-generate from URL domain
  • Use Firecrawl MCP to scrape the site with markdown format and main content extraction
  • Create organized folder structure in docs/research/[domain-or-subfolder]/
  • Generate descriptive filename based on URL path or page title
  • Save scraped content as markdown file with metadata header (URL, date, source)
  • Create or update an index file in the research folder listing all scraped content
  • Provide summary of scraped content and file location

Context

  • Research folder structure: docs/research/ (organized by domain/topic)
  • Existing research: !ls -la docs/context7-research/ docs/research/ 2>/dev/null | head -10
  • Firecrawl MCP status: Available for web scraping with markdown output
  • Current date: !date "+%Y-%m-%d"
  • Content organization: domain-based folders (anthropic, github, etc.) or custom subfolder names
  • File naming: descriptive names based on URL path, avoiding special characters
  • Metadata format: YAML frontmatter with url, scraped_date, domain, and title fields