64 boosters for "extract" — open source, verified from GitHub, ready to install
Extracts and summarizes YouTube video transcripts into structured notes with key learnings and actionable suggestions. Useful for researchers, students, and content consumers who want quick, clean summaries without manual transcription.
Arcanea InfoGenius Pro combines Google Search research with Gemini 3 Pro to generate visually rich infographics in 10 styles, helping developers and content creators quickly transform complex topics into engaging visual intelligence.
Create production-ready AI skills by extracting domain expertise and system ontologies, ensuring reliable performance in real-world applications. Ideal for developers building AI assistants that need deep contextual knowledge.
This MCP server enables extraction of text from documents, PDF manipulation, and OCR on images, making it useful for developers building document processing workflows in Claude Desktop, Claude Code, and Cursor.
A practical integration guide for enhancing OCR text extraction with visual and language LLM capabilities using local Ollama models in Caption Extractor. Developers working with document processing, image analysis, and text correction workflows benefit from this reusable agent framework.
Junipr MCP Server enables Claude to capture screenshots, extract PDFs, and retrieve metadata from web content through a standardized MCP interface. Developers and AI builders benefit from seamless integration of web-scraping capabilities into Claude-based applications.
Automates bulk annotation of CSV data by applying OpenAI prompts to each row and adding results as a new column. Ideal for developers who need to classify, extract, or summarize spreadsheet data at scale.
This MCP server automates cloud browser interactions to navigate websites, interact with page elements, and extract structured data—ideal for developers building web automation, data scraping, and testing workflows.
A web scraping MCP server that extracts and converts web pages into clean HTML, links, or Markdown format, supporting dynamic and complex pages. Useful for developers building AI applications that need to process web content programmatically.
This skill enables AI assistants to control a live Chrome browser session for web automation tasks like navigation, interaction, data extraction, and screenshots directly within Factory CLI workflows. Developers and AI agents benefit from live browser control for testing, scraping, and task automation without leaving the coding environment.
A simple MCP server that greets users by name and scrapes webpage content for quick reference. Useful for developers integrating basic greeting functionality and web content extraction into Claude-based applications.
A knowledge curation agent that researches, validates, and determines optimal storage methods (URL reference, local excerpt, or embedding) for information sources using parallel web scraping. Ideal for AI engineers building knowledge-intensive agent systems who need intelligent source evaluation and integration.
Extracts and summarizes YouTube video transcripts into structured notes with key learnings and actionable suggestions. Useful for researchers, students, and content consumers who want to quickly capture and organize video insights without manual note-taking.
Arjunkmrm Fetch is an MCP server that enables AI assistants to fetch and extract specific content from web pages using CSS selectors. It's valuable for developers and AI builders who need to integrate web scraping and data extraction capabilities into their Claude-powered applications.
Gradle Dependency Checker analyzes Gradle project dependencies, extracts version information, and visualizes dependency trees for Kotlin/kotlinx/skiko/androidx libraries. Developers working on Gradle-based projects benefit from quick dependency auditing and version management.
A specialized agent that intelligently researches, validates, and categorizes knowledge sources for AI systems, determining optimal storage methods (URL references, local extracts, or embeddings) using parallel web scraping. Ideal for developers building knowledge-intensive agent systems who need automated source curation and validation.