Showing projects tagged with "mcp-server".
High-performance local LLM server with GPU and NPU acceleration support, featuring multiple inference engines, OpenAI-compatible API, and cross-platform model deployment for AMD Ryzen AI processors.
Open-source MCP Host desktop application with universal LLM support for ChatGPT, Anthropic, Ollama, and OpenAI-compatible models, featuring OAPHub.ai cloud integration, cross-platform deployment via Tauri/Electron, 24+ language support, and OAuth-enabled MCP server integration.
Native macOS LLM server with MLX-optimized Apple Silicon inference, OpenAI/Ollama-compatible APIs, MCP server support for Cursor and Claude Desktop, remote provider connections, plugin ecosystem, menu bar chat, and Apple Foundation Models integration for macOS 26+.
Desktop app to manage MCP servers on Windows and macOS. Organize servers into Projects and Workspaces. Connects with Claude, Cline, Windsurf and Cursor. All data stored locally for privacy.