Paper | Features | Installation | Usage | CLI | Development
A Model Context Protocol (MCP) based LLM deep evaluation framework.
This project provides a framework for evaluating Large Language Models using the Model Context Protocol. It enables automating end- to-end task generation and deep evaluation of LLM agents across diverse dimensions.
🎬 Watch Full Demo Video (with audio)
Click above to download and view the complete MCPEval demonstration with audio explanation
MCPEval system architecture showing the complete evaluation pipeline from task generation to analysis
MCPEval web interface providing intuitive access to all evaluation features
- v1.1.0 — Multi-turn simulation web UI, conversation replay viewer, model comparison dashboard with statistical testing, SQLite persistence & v1 REST API, SFRGateway proxy, CI pipeline, and comprehensive test suite
- Supporting GPT-5
- Using model-config for using any model to generate and evaluate
- A new revalidation cli is released for generating high-quality data
- 🚀 Automated End-to-End Evaluation — Single-command pipeline from task generation to analysis with parallel execution
- 🔧 MCP Protocol Integration — 15+ built-in MCP servers spanning enterprise, utility, and public API domains
- 📊 Comprehensive Analysis & Insights — Statistical model comparison with bootstrap confidence intervals and paired tests
- 💻 User-Friendly Web Interface — Conversation replay viewer, model comparison dashboard, and multi-turn simulation UI
- ⚡ Advanced CLI Commands — Generate, verify, evaluate, simulate, and judge with flexible model configuration
- 🗄️ SQLite Persistence & REST API — Durable storage for evaluation runs with a v1 leaderboard and runs API
- 🔬 Multi-Turn Simulation — LLM-as-user simulation with scenario generation, persona support, and 5-dimension LLM judging
- 🌐 SFRGateway Proxy — Self-contained LLM inference via the Salesforce Research gateway (no direct API keys needed)
- ✅ CI & Test Suite — GitHub Actions pipeline with unit and integration tests
If you find our system or paper useful, please cite
@misc{liu2025mcpevalautomaticmcpbaseddeep,
title={MCPEval: Automatic MCP-based Deep Evaluation for AI Agent Models},
author={Zhiwei Liu and Jielin Qiu and Shiyu Wang and Jianguo Zhang and Zuxin Liu and Roshan Ram and Haolin Chen and Weiran Yao and Huan Wang and Shelby Heinecke and Silvio Savarese and Caiming Xiong},
year={2025},
eprint={2507.12806},
archivePrefix={arXiv},
primaryClass={cs.AI},
url={https://arxiv.org/abs/2507.12806},
}
For complete setup including both CLI and Web UI:
# Clone the repository
git clone https://github.com/SalesforceAIResearch/MCPEval.git
cd MCPEval
# Run unified setup script (installs CLI, backend API, and frontend UI)
./setup.shThis will set up:
- ✅ Core CLI evaluation framework
- ✅ Flask REST API backend
- ✅ React web interface
- ✅ All dependencies using uv package manager
For command-line usage only:
# Make sure uv is installed
curl -LsSf https://astral.sh/uv/install.sh | sh
# Install the package
uv sync
uv sync --extra devcp .env.template .env
Edit the .env file to add your OpenAI API key:
OPENAI_API_KEY=YOUR_OPENAI_API_KEY_HERE
OR export the key in your terminal:
export OPENAI_API_KEY=YOUR_OPENAI_API_KEY_HERE
For self-contained LLM inference without managing API keys directly, use the bundled SFRGateway proxy:
cd sfrgateway
cp .env.template .env # edit .env with your X_API_KEY
PROXY_PORT=8008 uv run python server.pyThen point model configs at http://localhost:8008/v1 with "api_key": "dummy". See sfrgateway/README.md for details.
After running the setup script:
-
Start the backend API:
cd backend uv run app.pyBackend will run on
http://localhost:22358 -
Start the frontend (in a new terminal):
cd frontend npm startFrontend will run on
http://localhost:22359 -
Access the web application:
- Open
http://localhost:22359in your browser - Use the intuitive interface to generate tasks, run evaluations, and view results
- Conversation Replay — Browse and inspect multi-turn conversations turn by turn
- Model Comparison — Side-by-side model comparison with statistical significance testing
- Multi-Turn Simulation — Generate scenarios, run user simulations, and evaluate conversations from the UI
- Real-time progress tracking for all operations
- Open
Note: The frontend automatically proxies API requests to the backend server (port 22358). No additional configuration is needed.
For advanced users and automation:
We provide an example about a special calculator MCP application. We define an example special calculator MCP server and use OpenAI client to interact with the server.
Quick start:
# Basic example with local MCP server
uv run mcp_clients/example_openai_client/client.py --servers mcp_servers/special_calculator/server.py
# Multiple servers with environment variables (use ^ for env vars)
uv run mcp_clients/example_openai_client/client.py --servers @modelcontextprotocol/server-sequential-thinking mcp-server-nationalparks^NPS_API_KEY=your-api-key-here
# Combined example with arguments and environment variables
uv run mcp_clients/example_openai_client/client.py --servers @openbnb/mcp-server-airbnb:--ignore-robots-txt mcp-server-nationalparks^NPS_API_KEY=your-api-key-hereFor more details on the OpenAI client usage, see the OpenAI Client README.
MCPEval includes a diverse set of MCP servers spanning enterprise domains, public APIs, and computation utilities. Each server exposes tools that LLM agents are evaluated against.
These servers are fully deterministic with embedded data or pure computation — ideal for reproducible evaluation.
| Server | Tools | Domain | Description |
|---|---|---|---|
| hr_management | 10 | Enterprise | Departments, employees, leave requests, performance reviews, org chart. Embedded SQLite with 70+ rows. |
| ecommerce | 11 | Enterprise | Products, orders, customers, inventory, sales summaries. Embedded SQLite with 80+ rows. |
| datetime_tools | 7 | Utility | Timezone conversion, date difference, business days, holiday support (US/UK/DE/FR/JP). |
| unit_converter | 6 | Utility | Length, weight, temperature, volume, speed, data size conversion with strict enum schemas. |
| special_calculator | 4 | Demo | Basic arithmetic with special transformations (add+double, subtract+halve, etc.). |
| sqlite | 8 | Database | General-purpose SQLite operations — create tables, query, insert, with sample datasets. |
| filesystem | 14 | System | Local file operations (read, write, search, directory listing). npm: @modelcontextprotocol/server-filesystem |
| memory | 9 | Knowledge | Knowledge graph with entities, relations, and observations. npm: @modelcontextprotocol/server-memory |
| Server | Tools | Domain | Description |
|---|---|---|---|
| book | 8 | Library | Open Library search — books by title/ISBN, authors, advanced search. |
| youtube | 4 | Media | YouTube transcript extraction, search, and summarization. |
| healthcare | 5 | Medical | FDA drug lookup, PubMed search, clinical trials, ICD-10 codes. |
| sports | 4 | Sports | NBA, MLB, NFL teams, players, and game data via balldontlie.io. |
| Server | Tools | Domain | Credentials |
|---|---|---|---|
| travel_assistant | 6 | Travel | Flights, hotels, restaurants, local events. Requires SERPAPI_API_KEY, YELP_API_KEY. |
| airbnb | 2 | Travel | Airbnb listing search and details. npm: @openbnb/mcp-server-airbnb |
| yfinance | 10 | Finance | Stock prices, financials, options, analyst recommendations via Yahoo Finance. |
| national_park | 6 | Parks | U.S. National Parks info, alerts, campgrounds, events. Requires NPS_API_KEY (free). |
| crm_bench | 11 | CRM | Salesforce CRM operations (stub implementation for benchmarking). |
MCPEval supports multi-turn user simulation where a simulator LLM plays the user role and an agent LLM is tested:
# Generate scenarios from verified tasks
mcp-eval generate-scenarios \
--servers mcp_servers/hr_management/server.py \
--output scenarios.jsonl \
--num-scenarios 5
# Run multi-turn simulation
mcp-eval simulate \
--servers mcp_servers/hr_management/server.py \
--simulator-model-config simulator_model.json \
--agent-model-config agent_model.json \
--scenarios-file scenarios.jsonl \
--output multiturn_results.jsonl
# Evaluate conversations with LLM judge
mcp-eval evaluate-multiturn \
--input multiturn_results.jsonl \
--output multiturn_evaluation.jsonlThe judge evaluates on 5 dimensions: clarification handling, context maintenance, tool usage efficiency, goal achievement, and response quality.
# Complete development environment
./setup.sh
# Start backend API (Terminal 1)
cd backend && uv run app.py
# Start frontend UI (Terminal 2)
cd frontend && npm start
# Access at http://localhost:22359For each benchmark contribution, please follow the following steps:
- Create a new directory in the
benchmarks/your_benchmark_namefolder. - If you are developing a new MCP server, please create a new folder and add the server script in the
mcp_serversfolder. - If you are developing a new MCP client, please create a new folder and add the client script in the
mcp_clientsfolder. - Add your benchmark scripts to the
benchmarks/your_benchmark_namefolder.
For web interface contributions:
- Frontend components:
frontend/src/components/andfrontend/src/pages/ - Backend API endpoints:
backend/app.py
See our detailed Development Roadmap for the current progress and planned features across all components.
The MCPEval CLI provides a comprehensive toolkit for managing MCP servers and evaluating LLMs. For detailed documentation, parameter descriptions, and advanced usage examples, see the CLI README.
Auto Workflow (Recommended) - Complete evaluation pipeline in one command:
# Automatically generate tasks, verify, evaluate, and analyze results
mcp-eval auto \
--servers @openbnb/mcp-server-airbnb:--ignore-robots-txt \
--working-dir evaluation_results/airbnb_eval \
--task-model gpt-4o-2024-11-20 \
--eval-model-configs benchmarks/airbnb/eval_models/gpt-4o.json \
--num-tasks 50For more control over each step:
# 1. Generate tasks
mcp-eval generate-tasks \
--servers @openbnb/mcp-server-airbnb:--ignore-robots-txt \
--model-config benchmarks/airbnb/eval_models/gpt-4o.json \
--num-tasks 200 \
--output data/airbnb/evaluation_tasks.jsonl
# 2. Verify tasks work correctly
mcp-eval verify-tasks \
--servers @openbnb/mcp-server-airbnb:--ignore-robots-txt \
--tasks-file data/airbnb/evaluation_tasks.jsonl \
--output data/airbnb/evaluation_tasks_verified.jsonl
# 3. Revalidate task descriptions based on execution data (optional but recommended)
mcp-eval revalidate-tasks \
--verified-tasks-file data/airbnb/evaluation_tasks_verified.jsonl \
--model-config benchmarks/airbnb/eval_models/gpt-4o.json \
--output data/airbnb/evaluation_tasks_final.jsonl
# 4. Evaluate model performance
mcp-eval evaluate \
--servers @openbnb/mcp-server-airbnb:--ignore-robots-txt \
--model-config benchmarks/airbnb/eval_models/gpt-4o.json \
--tasks-file data/airbnb/evaluation_tasks_final.jsonl \
--output benchmarks/airbnb/results/gpt4o_evaluation.json \
--max-turns 30
# 5. Analyze results and generate reports
mcp-eval analyze \
--predictions benchmarks/airbnb/results/gpt4o_evaluation.json \
--ground-truth data/airbnb/evaluation_tasks_final.jsonl \
--generate-report
# 6. Optional: Run LLM judge evaluation
mcp-eval judge \
--input-file benchmarks/airbnb/results/gpt4o_evaluation.json \
--output-dir benchmarks/airbnb/results \
--model-config benchmarks/airbnb/eval_models/gpt-4o.json
# 7. Optional: Analyze LLM judgment results
mcp-eval judge-rubric \
--trajectory-file benchmarks/airbnb/results/gpt4o_evaluation_trajectory.json \
--completion-file benchmarks/airbnb/results/gpt4o_evaluation_completion.json \
--output-dir benchmarks/airbnb/reportNote: The revalidation step (step 3) analyzes the actual tool conversations from verified tasks and improves task descriptions to be more accurate and specific. This leads to higher-quality evaluation datasets and better task clarity for subsequent evaluations.
generate-tasks- Generate evaluation tasks for MCP serversverify-tasks- Verify tasks can be executed successfullyrevalidate-tasks- Improve task descriptions based on actual execution dataevaluate- Evaluate models using MCP servers and tasksanalyze- Analyze evaluation results and generate reportsjudge- Run LLM-based evaluation of execution trajectoriesjudge-rubric- Analyze LLM judgment resultsgenerate-scenarios- Generate multi-turn scenarios from tasks or serverssimulate- Run multi-turn user simulation conversationsevaluate-multiturn- Evaluate multi-turn conversations with LLM judgeconvert-data- Convert data to different formats (e.g., XLAM)auto- Complete automated evaluation workflow
Models are configured using JSON files. Examples:
{
"model": "gpt-4o-2024-11-20",
"temperature": 0.01,
"max_tokens": 16384
}For custom endpoints:
{
"model": "mistral-24b",
"api_key": "default",
"temperature": 0.01,
"max_tokens": 3000,
"base_url": "http://<IP_Address>:<port>/v1"
}# General help
mcp-eval --help
# Command-specific help
mcp-eval generate-tasks --help
mcp-eval evaluate --helpFor comprehensive documentation, examples, and advanced usage patterns, see the Complete CLI Documentation.
This project is licensed under the Apache 2.0 License. See the LICENSE file for details.
For any questions or feedback, please contact Zhiwei Liu at zhiweiliu@salesforce.com.

