Back to Categories

MCP Server

Discover the best MCP Server MCP servers for AI agents. Browse tools, use cases, installation guides, and integration documentation for mcp server-focused Model Context Protocol implementations.

3 results found

mcp-server-with-spring-ai
mcp-server-with-spring-ai
mcp-server-with-spring-ai
mcp-server-with-spring-ai is a Spring Boot integrated MCP (Model Context Protocol) server example that showcases how to expose executable tools from an MCP server to clients (including LLMs) and how to wire a MCP client to consume those tools. The documentation explains MCP at a high level, outlines the three-layer MCP Java SDK architecture (Client/Server Layer, Session Layer, Transport Layer), and demonstrates two sample tools implemented in SellerAccountTools. This repo emphasizes how an MCP server can connect to external data sources (e.g., a PostgreSQL DB) and expose tools that an AI model can invoke to retrieve data, with the example illustrating tool invocation and automatic tool selection by prompts.
OpenClaw MCP Server
OpenClaw MCP Server
OpenClaw MCP Server
OpenClaw MCP Server is a secure Model Context Protocol (MCP) bridge that connects Claude.ai with a self-hosted OpenClaw assistant, enabling OAuth2 authentication and safe, controlled communication between the Claude AI ecosystem and your local or hosted OpenClaw deployment. This MCP server acts as an orchestration layer that exposes MCP tools to Claude.ai, manages authentication, and enforces security boundaries like CORS and transport options. It is designed to be deployed via Docker or run locally, with detailed installation, configuration, and security guidance provided in the documentation. By serving as a bridge, it enables Claude.ai to delegate tasks to your OpenClaw bot while preserving control over data flow and access controls, in line with MCP specifications and best security practices.
External MCP Server
External MCP Server
External MCP Server
Neurolink includes an External MCP Server capability, enabling seamless integration with external Model Context Protocol (MCP) servers. This feature loads and manages external MCP servers from a dedicated configuration file (.mcp-config.json), enables real JSON-RPC based communication, and supports end-to-end tool execution within the NeuroLink platform. It is designed for multi-provider AI workflows, allowing providers to delegate tool execution to external servers while preserving type safety, robust error handling, and deterministic behavior. The documentation highlights how to configure external MCP servers, register and discover tools, and perform end-to-end tool execution through the CLI, ensuring a production-ready MCP ecosystem.