Servers/MLflow Evaluate

MLflow Evaluate MCP Server

Grade: Best Budget Option

MLflow built-in LLM evaluation with custom metrics

GitHub24.8k starsUpdated todayApache-2.0
4.9ToolRoute
Value Score:7.4
Sample size:0 runs
Confidence:Accumulating
Last updated:No data yet

Score Breakdown

7.8
Output Quality
How good are the results?
8.0
Reliability
Does it work consistently?
8.2
Efficiency
How heavy is it to use?
9.2
Cost
Is it worth the price?
10.0
Trust
Is it safe to use?

All scores are out of 10. Based on accumulated telemetry.

About

The open source AI engineering platform for agents, LLMs, and ML models. MLflow enables teams of all sizes to debug, evaluate, monitor, and optimize production-quality AI applications while controlling costs and managing access to models and data.

Quick Install

See GitHub repo for install instructions

Fallback Intelligence

Fallback routing available via POST /api/route — the routing engine automatically selects the best alternative when this server is unavailable or underperforming.

Add Badge to Your README

[![ToolRoute Score](https://toolroute.io/api/badge/mlflow-eval)](https://toolroute.io/mcp-servers/mlflow-eval)
ToolRoute|4.9/10
Badge updates automatically as your score changes

Help improve this score

Used this MCP server? Report your execution outcome and earn routing credits that improve your future recommendations.

Report an outcome+3 to +10 routing credits
Compare two servers+8 to +25 routing credits
Submit a benchmark package+15 to +40 routing credits
POST /api/report { "skill_slug": "mlflow-eval", "outcome": "success" }
See full API docs →