15 boosters for "llama" — open source, verified from GitHub, ready to install
A specialized AI agent that automatically detects, classifies, and fixes data anomalies in production pipelines using local SLMs and semantic clustering, with zero data loss guarantee. Data engineers and platform teams benefit most when dealing with broken pipelines that can't afford downtime.
An AI-powered VSCode extension that auto-generates standardized Git/SVN commit messages using multiple AI services (OpenAI, ChatGPT, Ollama, Gemini, etc.) with multi-language support and weekly report generation. Developers working with Git/SVN in VSCode benefit from faster, consistent commit workflows.
Heuristic scoring (no AI key configured).
VT.ai provides Copilot-specific coding instructions for a multimodal AI chat application, establishing standards for Python development including naming conventions, style guides, and testing practices. Developers building AI-powered features with language models will benefit from these standardized guidelines.
RagCode MCP is a semantic code navigation tool that integrates RAG-powered code search into Windsurf and other IDEs, enabling developers to intelligently query and understand multi-language codebases using local LLMs. It's ideal for developers working with Laravel, Go, Python, and PHP who need fast, context-aware code exploration without leaving their IDE.
Transforms rough prompts into production-ready LLM prompts using advanced techniques like Chain-of-Thought and RAG optimization. Ideal for developers, AI engineers, and teams building LLM applications who need to craft effective prompts across Claude, GPT, Llama, and other models.
Automates synchronization of custom Ollama models across distributed Vera AI clusters using YAML configuration and templating. Benefits DevOps engineers and teams managing multi-instance LLM deployments who need centralized model management.
Intelli-router automatically triages user messages by complexity and routes them to the most cost-effective AI model (local Ollama, Claude Sonnet, Codex, or Claude Opus). Developers building multi-model AI systems benefit from reduced costs and optimized latency.
Geist provides Cursor-integrated coding rules and best practices for Python, SQLAlchemy, logging, and server development using FastAPI and LLM models. It helps developers maintain consistent code standards and proper patterns when building AI-powered applications.
A JavaScript-based system for dynamically creating and modifying AI agents without recompilation, enabling local LLM developers to test and share agent configurations interactively.
An AI-powered code auditing tool that integrates with Claude Desktop via MCP protocol, using local Ollama models to analyze code for security vulnerabilities, performance issues, and quality concerns without sending code to external services. Developers and security teams benefit from automated, privacy-preserving code reviews with customizable analysis parameters.
A template and guidance document for customizing system prompts for hackathon presentation automation, particularly for integrating with demo tools and avatar systems. Useful for hackathon organizers and participants building AI-powered presentation systems.
A practical integration guide for enhancing OCR text extraction with visual and language LLM capabilities using local Ollama models in Caption Extractor. Developers working with document processing, image analysis, and text correction workflows benefit from this reusable agent framework.
A comprehensive system prompt and .cursorrules template for building a manufacturing defect detection MVP in Cursor AI, providing structured guidance on FastAPI backend, React frontend, and LLaVA vision model integration.
Kryptoskatt MCP Server enables AI assistants to query DeFi data from DefiLlama API through standard MCP protocol. It's useful for developers and analysts building AI applications that need real-time blockchain and DeFi information.