MT-Bench
BenchmarkFreeMulti-turn conversation benchmark — 80 questions, 8 categories, GPT-4 as judge.
Capabilities10 decomposed
multi-turn conversation quality evaluation with gpt-4 judging
Medium confidenceMT-Bench evaluates LLM responses across 80 curated multi-turn questions using GPT-4 as an automated judge. The system submits model responses to GPT-4 with structured prompts that assess instruction following, reasoning coherence, and conversation consistency across turns. Responses are scored on a numeric scale, enabling quantitative comparison of model capabilities without human annotation overhead.
Uses GPT-4 as a scalable automated judge rather than crowdsourced human evaluation, enabling rapid iteration and reproducible scoring across 70+ models. The 80-question set is specifically designed for multi-turn reasoning (not single-turn), with questions spanning writing, roleplay, reasoning, math, coding, and knowledge domains.
Faster and cheaper than human evaluation (HELM, AlpacaEval use crowdsourcing) but more expensive than single-turn metrics; provides multi-turn context that single-turn benchmarks (MMLU, HellaSwag) cannot capture.
question-answer pair dataset curation and versioning
Medium confidenceMT-Bench maintains a curated set of 80 high-quality multi-turn questions across 8 semantic categories (writing, roleplay, extraction, reasoning, math, coding, knowledge, common-sense). Questions are stored as structured JSON with turn-by-turn prompts, enabling reproducible evaluation. The dataset is version-controlled in the FastChat repository, allowing tracking of changes and ensuring consistent benchmark definitions across research papers.
Explicitly structures questions as multi-turn conversations (not single-turn), with each question containing 2-3 sequential turns that build on prior context. Questions are manually curated by LMSYS researchers rather than automatically generated, ensuring semantic diversity and avoiding trivial or duplicate questions.
More rigorous than auto-generated benchmarks (HELM uses templates) but smaller in scale; provides explicit multi-turn structure that single-turn benchmarks (MMLU, ARC) cannot evaluate.
batch evaluation orchestration with distributed model inference
Medium confidenceMT-Bench integrates with FastChat's distributed serving infrastructure to evaluate multiple models in parallel. The evaluation pipeline submits each question to candidate models via the FastChat controller (which routes to model workers), collects responses, and batches them for GPT-4 judging. This architecture enables evaluating 70+ models without sequential bottlenecks, leveraging the controller-worker pattern for load distribution.
Leverages FastChat's controller-worker architecture (documented in DeepWiki) to distribute inference across multiple model workers, avoiding the need to implement custom parallelization. The evaluation pipeline is tightly integrated with FastChat's conversation templates and model adapters, ensuring consistent prompt formatting across models.
More efficient than sequential evaluation (HELM evaluates models one-at-a-time) but requires FastChat infrastructure; simpler than building custom distributed evaluation (e.g., Ray, Kubernetes) because it reuses existing controller-worker pattern.
leaderboard ranking and elo rating calculation
Medium confidenceMT-Bench scores feed into LMSYS's Elo rating system, which computes relative model strength based on pairwise comparison results. The Elo algorithm treats benchmark scores as implicit pairwise wins/losses, updating model ratings iteratively. Leaderboard rankings are published on lmarena.ai and updated weekly, providing a public-facing metric for model comparison that accounts for both absolute performance and relative positioning.
Applies Elo rating system (borrowed from chess) to LLM evaluation, converting absolute benchmark scores into relative rankings that account for the strength of competing models. This approach is more robust to benchmark saturation than absolute scores — as models improve, Elo ratings naturally spread to maintain discrimination.
More sophisticated than simple score ranking (HELM publishes raw scores) because it accounts for relative model strength; enables confidence intervals and trend analysis that raw scores cannot provide.
conversation template application for model-specific prompt formatting
Medium confidenceMT-Bench questions are formatted according to model-specific conversation templates (defined in FastChat's conversation.py) before submission to each model. Templates handle differences in prompt structure, special tokens, and role markers (e.g., Llama uses [INST], ChatGLM uses different role tags). This ensures that each model receives questions in its native format, preventing unfair evaluation due to prompt formatting mismatches.
Centralizes model-specific prompt formatting in FastChat's conversation template system (documented in DeepWiki), avoiding scattered prompt engineering across evaluation code. Templates are versioned and tested, ensuring consistency across benchmark runs. The system supports 40+ model families with a single template registry.
More maintainable than ad-hoc prompt engineering (HELM requires custom prompts per model) because templates are reused across FastChat's serving, training, and evaluation pipelines.
response collection and storage with turn-level granularity
Medium confidenceMT-Bench collects model responses at the turn level (not just final responses) and stores them in structured JSON format. Each turn's response is timestamped, includes metadata (model name, inference time, token count), and is linked to the corresponding question turn. This enables post-hoc analysis of how models handle multi-turn context and allows re-judging with different judges without re-running inference.
Stores responses at turn granularity rather than aggregating to final answer, enabling analysis of how models handle context accumulation. Metadata (inference time, token count) is captured alongside responses, supporting performance analysis beyond quality metrics.
More detailed than simple score storage (HELM stores only final scores) but requires more storage; enables re-judging and post-hoc analysis that single-run evaluation cannot support.
gpt-4 judge prompt engineering and consistency validation
Medium confidenceMT-Bench uses carefully engineered prompts to instruct GPT-4 to evaluate responses on dimensions like instruction following, reasoning, and coherence. The judge prompt includes examples of good/bad responses and explicit scoring rubrics to reduce variance. Consistency is validated by re-judging a subset of responses and computing inter-judge agreement (e.g., Spearman correlation between first and second judgments).
Validates judge consistency through re-judging and correlation analysis, rather than assuming GPT-4 is a perfect judge. The approach acknowledges that automated judging introduces variance and provides metrics to quantify it. Judge prompts are published alongside results, enabling reproducibility and external validation.
More rigorous than single-pass judging (most benchmarks don't validate judge consistency) but more expensive; provides transparency that proprietary judges (e.g., Claude-based evaluation) cannot offer.
correlation analysis between benchmark scores and human preferences
Medium confidenceMT-Bench scores are validated against human preferences collected via Chatbot Arena (side-by-side model battles). The system computes correlation metrics (Spearman, Kendall) between MT-Bench rankings and Chatbot Arena Elo ratings, validating that the automated benchmark aligns with human judgment. This validation is critical for establishing benchmark credibility and identifying cases where the benchmark may be misaligned with real-world preferences.
Uniquely validates MT-Bench against human preferences from Chatbot Arena (1.5M+ votes), providing empirical evidence that automated scores align with human judgment. This validation is published alongside benchmark results, establishing transparency about benchmark limitations.
More credible than benchmarks without human validation (MMLU, HumanEval lack large-scale human preference data) but requires access to human evaluation infrastructure that most teams don't have.
category-level performance breakdown and capability analysis
Medium confidenceMT-Bench questions are organized into 8 semantic categories (writing, roleplay, extraction, reasoning, math, coding, knowledge, common-sense), enabling per-category performance analysis. The evaluation pipeline computes separate scores for each category, revealing which models excel at specific capabilities and which have gaps. This breakdown is more informative than aggregate scores and helps identify model strengths/weaknesses.
Explicitly structures evaluation around semantic categories (writing, math, coding, etc.) rather than treating all questions equally. This enables capability-level analysis that aggregate scores cannot provide, supporting task-specific model selection.
More actionable than single-number benchmarks (MMLU provides only aggregate score) but less granular than domain-specific benchmarks (HumanEval for coding, MATH for mathematics).
benchmark reproducibility through fixed question sets and seed management
Medium confidenceMT-Bench ensures reproducibility by using a fixed, versioned set of 80 questions and managing random seeds for model inference (temperature, sampling parameters). The system records evaluation metadata (model version, inference parameters, GPT-4 model version, timestamp) enabling exact reproduction of results. Questions are publicly available, allowing external researchers to verify results or run independent evaluations.
Treats reproducibility as a first-class concern by versioning questions, recording all inference parameters, and publishing metadata alongside results. Questions are public, enabling external verification.
More reproducible than proprietary benchmarks (which don't publish questions); more rigorous than informal evaluation practices that don't track parameters.
Capabilities are decomposed by AI analysis. Each maps to specific user intents and improves with match feedback.
Related Artifactssharing capabilities
Artifacts that share capabilities with MT-Bench, ranked by overlap. Discovered automatically through the match graph.
LMSYS Chatbot Arena
Crowdsourced LLM evaluation — side-by-side blind voting, Elo ratings, most trusted LLM benchmark.
UltraChat 200K
200K high-quality multi-turn dialogues for instruction tuning.
WildChat
1M+ real user-AI conversations with demographic metadata.
xAI: Grok 4
Grok 4 is xAI's latest reasoning model with a 256k context window. It supports parallel tool calling, structured outputs, and both image and text inputs. Note that reasoning is not...
OpenAI: GPT-4 (older v0314)
GPT-4-0314 is the first version of GPT-4 released, with a context length of 8,192 tokens, and was supported until June 14. Training data: up to Sep 2021.
OpenAI: GPT-5.3 Chat
GPT-5.3 Chat is an update to ChatGPT's most-used model that makes everyday conversations smoother, more useful, and more directly helpful. It delivers more accurate answers with better contextualization and significantly...
Best For
- ✓LLM researchers benchmarking model families (Llama, Mistral, GPT variants)
- ✓Teams building Chatbot Arena or similar competitive evaluation platforms
- ✓Organizations selecting production LLMs based on multi-turn capability metrics
- ✓Researchers publishing LLM evaluation papers requiring standardized benchmarks
- ✓Teams building internal LLM leaderboards that need consistent question sets
- ✓Model developers analyzing performance breakdown by question category
- ✓Teams running Chatbot Arena with 70+ models requiring daily/weekly evaluations
- ✓Organizations with distributed GPU clusters wanting to parallelize benchmark runs
Known Limitations
- ⚠GPT-4 judge introduces cost (~$0.03-0.06 per evaluation) and dependency on OpenAI API availability
- ⚠Judge bias: GPT-4 may favor models with similar reasoning patterns to its own training
- ⚠No human validation layer — automated scoring can miss nuanced quality differences
- ⚠Fixed question set limits evaluation to 8 predefined categories; custom domains require new question curation
- ⚠Fixed 80-question set may not cover domain-specific tasks (medical, legal, scientific)
- ⚠English-only questions; multilingual evaluation requires separate benchmark
Requirements
Input / Output
UnfragileRank
UnfragileRank is computed from adoption signals, documentation quality, ecosystem connectivity, match graph feedback, and freshness. No artifact can pay for a higher rank.
About
Multi-turn conversation benchmark with 80 high-quality questions across 8 categories. Tests multi-turn reasoning, instruction following, and conversation coherence. Uses GPT-4 as judge. Part of the LMSYS evaluation suite.
Categories
Alternatives to MT-Bench
Are you the builder of MT-Bench?
Claim this artifact to get a verified badge, access match analytics, see which intents users search for, and manage your listing.
Get the weekly brief
New tools, rising stars, and what's actually worth your time. No spam.
Data Sources
Looking for something else?
Search →