12 boosters for "scraping" — open source, verified from GitHub, ready to install
An MCP server that integrates Bright Data's web scraping, browser automation, and data extraction tools into Claude, enabling AI agents to collect and process structured data from web sources while handling anti-bot detection.
"name": "@apitap/core", "version": "1.10.1", "description": "Intercept web API traffic during browsing. Generate portable skill files so AI agents can call APIs directly instead of scraping.",
ScrapeGraphAI Scrapegraph MCP enables Claude and other language models to perform intelligent, enterprise-grade web scraping directly within their conversations. Developers, data analysts, and researchers benefit from automated data extraction without building custom scraping infrastructure.
An MCP server that integrates Crawlbase API for web scraping capabilities into AI agents and code editors like Claude, Cursor, and Windsurf. Developers building AI applications needing reliable web data extraction benefit from this seamless integration.
"name": "anysite-skills", "description": "Professional skills for data extraction and analysis using anysite MCP server. Covers LinkedIn, Instagram, Twitter/X, Reddit, YouTube, Y Combinator, SEC, and web scraping for lead generation, competitive intelligence, market research, social media analytics,
Heuristic scoring (no AI key configured).
A system prompt for automating browser-based tasks using AI agents with Playwright and Gemini, enabling developers to build automated workflows for web interactions like job searches and email management across multiple AI platforms.
A skill that imports JIRA tickets via API or web scraping and converts them into local story files, enabling developers to quickly start implementation from their backlog. Useful for teams managing JIRA-based workflows who want to streamline ticket-to-code handoffs.
A knowledge curation agent that researches, validates, and determines optimal storage methods (URL reference, local excerpt, or embedding) for information sources using parallel web scraping. Ideal for AI engineers building knowledge-intensive agent systems who need intelligent source evaluation and integration.
A specialized agent that intelligently researches, validates, and categorizes knowledge sources for AI systems, determining optimal storage methods (URL references, local extracts, or embeddings) using parallel web scraping. Ideal for developers building knowledge-intensive agent systems who need automated source curation and validation.