This repository provides a Python-based Model Context Protocol (MCP) client using LangChain, enabling interaction with MCP servers through LangChain ReAct Agent and supporting LLMs from Anthropic, OpenAI, and Groq.
This project is a Model Context Protocol (MCP) client demonstrating the use of MCP server tools with a LangChain ReAct Agent. It uses convert_mcp_to_langchain_tools() from the langchain_mcp_tools library to initialize multiple MCP servers and convert their tools into a list of LangChain-compatible tools. It supports LLMs from Anthropic, OpenAI, and Groq.
Prerequisites: Python 3.11+, optional uv/uvx and npm/npx, and API keys from Anthropic, OpenAI, and/or Groq.
Setup: Install dependencies using make install, configure API keys in .env (copied from .env.template), and configure LLM and MCP server settings in llm_mcp_config.json5. The configuration file format is similar to Claude for Desktop, using mcp_servers instead of mcpServers, and supports JSON5 with environment variable substitution using ${...}.
Usage: Run the app with make start (verbose mode: make start-v, help: make start-h). Example queries, configurable in llm_mcp_config.json5, are available by pressing Enter at the prompt.
hideya/mcp-client-langchain-py
January 12, 2025
March 27, 2025
Python