12 boosters for "ollama" — AI-graded, open source, ready to install
Local persistent memory store for LLM applications including claude desktop, github copilot, codex, antigravity, etc.
An AI-powered VSCode extension that auto-generates standardized Git/SVN commit messages using multiple AI services (OpenAI, ChatGPT, Ollama, Gemini, etc.) with multi-language support and weekly report generation. Developers working with Git/SVN in VSCode benefit from faster, consistent commit workflows.
Heuristic scoring (no AI key configured).
VT.ai provides Copilot-specific coding instructions for a multimodal AI chat application, establishing standards for Python development including naming conventions, style guides, and testing practices. Developers building AI-powered features with language models will benefit from these standardized guidelines.
RagCode MCP is a semantic code navigation tool that integrates RAG-powered code search into Windsurf and other IDEs, enabling developers to intelligently query and understand multi-language codebases using local LLMs. It's ideal for developers working with Laravel, Go, Python, and PHP who need fast, context-aware code exploration without leaving their IDE.
ADHARA is an LLM system prompt that enables early detection of learning friction by analyzing behavioral metrics like mouse hesitation, attention, and task performance against age-appropriate baselines. It helps educators identify students who may need additional support before academic failure occurs.
A production-grade agentic chatbot server built in Rust with multi-provider LLM support, tool calling, RAG, MCP integration, and advanced research capabilities
A Self-Modifying Multi-Agent Cognitition Architecture with Proactive Background Reflection (SMMAC-PBR)
Chat with AI via CLI
vala based gtk & terminal based desktop Ollama chat with tools - designed for local LLM work
AI-powered code auditing via MCP using local Ollama models for security, performance, and quality analysis
This document explains how to use the AI agent features in Caption Extractor, which enhance OCR processing with visual LLM models using local Ollama.