lm-evaluation-harnessFramework43/100
via “metric computation with bootstrapped confidence intervals”
EleutherAI's evaluation framework — 200+ benchmarks, powers Open LLM Leaderboard.
Unique: Integrates bootstrapped confidence interval computation directly into the metrics pipeline, automatically resampling predictions to estimate metric variance. The system supports both built-in metrics (accuracy, F1, BLEU, ROUGE) and custom metric functions, with aggregation at task and suite levels. Bootstrapping is configurable (default 100k iterations) and cached to avoid recomputation.
vs others: Provides confidence intervals by default (not optional), which alternatives like simple accuracy reporting lack; bootstrapping approach is more robust than analytical CI formulas for non-normal distributions