ReasonBENCH is a benchmark suite and open-source library for controlled multi-run evaluation of LLM reasoning. It measures both the quality and stability of reasoning strategies by running repeated independent trials and reporting variance-aware metrics — including confidence intervals, run deviation, and global noise — rather than relying on single-run averages.
Preliminary work. Under review by the International Conference on Machine Learning (ICML).
LLM reasoning is typically evaluated using single runs, masking how much performance can vary across repeated executions. This practice obscures both reliability and cost, and can lead to misleading comparisons between methods and models. ReasonBENCH addresses this by repeating every model-strategy-task configuration with 10 independent trials and reporting distributional metrics alongside averages.
Key findings from our evaluation:
- Run-to-run variability is substantial — often large enough to change model/method rankings relative to single-run averages
- Quality and cost stability decouple — the most accurate strategy is not necessarily the most stable, and vice versa
- Model scaling improves both quality and stability — larger models within a family yield tighter distributions
- Prompt refinements improve quality but not stability — clarifying prompts and parsers boosts accuracy without reducing run-to-run variance
- Reasoning effort scales cost, not quality — increasing test-time reasoning effort primarily raises cost with limited and statistically insignificant quality gains
We implement 10 representative reasoning strategies using a standardized interface:
| Strategy | Type | Reference |
|---|---|---|
| IO | Direct | — |
| CoT | Direct | Wei et al., 2022 |
| CoT-SC | Direct | Wang et al., 2023 |
| ReAct | Adaptive | Yao et al., 2023b |
| Reflexion | Adaptive | Shinn et al., 2023 |
| ToT-BFS | Structured | Yao et al., 2023a |
| ToT-DFS | Structured | Yao et al., 2023a |
| GoT | Structured | Besta et al., 2024 |
| RAP | Planning | Hao et al., 2023 |
| FoA | Evolutionary | Klein et al., 2025 |
6 tasks spanning diverse reasoning domains:
| Task | Domain | Metric | Size |
|---|---|---|---|
| Game of 24 | Mathematical reasoning | Accuracy | 100 |
| SciBench | Scientific reasoning | Accuracy (exact match) | 109 |
| HumanEval | Code generation | pass@1 | 100 |
| HotPotQA | Multi-hop QA | Exact match | 100 |
| Sonnet Writing | Creative writing | Accuracy (rhyme + words) | 50 |
| HLE | General reasoning (Humanity's Last Exam) | Accuracy | 50 |
10 contemporary reasoning models from 6 providers:
| Model | Provider |
|---|---|
| GPT-4.1 Nano, GPT-4.1 Mini | OpenAI |
| GPT-5 Nano, GPT-5 Mini | OpenAI |
| GPT-OSS 120B | Together AI |
| DeepSeek R1 | Together AI |
| Llama 4 Maverick | Together AI |
| Qwen3-235B Thinking | Together AI |
| Claude Haiku 4.5 | Anthropic |
| Gemini 3 Flash |
pip install -r requirements.txtYou also need CacheSaver — a client-side inference optimization framework for efficient, affordable, and reproducible LLM inference:
pip install cachesaverSet your API keys as environment variables:
export OPENAI_API_KEY_CLAN="sk-..."
# and/or other provider keysThe simplest way to run an experiment is via the shell script:
bash scripts/simple/simple.shEdit the variables at the top of scripts/simple/simple.sh to change the benchmark, method, model, and split.
For direct invocation:
python scripts/simple/simple.py \
--benchmark game24 \
--method tot_bfs \
--split mini \
--provider openai \
--api_key OPENAI_API_KEY_CLAN \
--model gpt-4.1-nano \
--temperature 1.0 \
--max_completion_tokens 10000 \
--top_p 1.0 \
--batch_size 1 \
--timeout 2.0 \
--correctness 1 \
--allow_batch_overflow 1 \
--ns_ratio 0.0 \
--value_cache| Argument | Description |
|---|---|
--benchmark |
Task name: game24, humaneval, hotpotqa, scibench, hle, sonnetwriting |
--method |
Reasoning method: io, cot, cot_sc, foa, tot_bfs, tot_dfs, got, react, rap |
--split |
Dataset split: train, validation, test, mini |
--provider |
LLM provider: openai, gemini, anthropic, groq, together |
--model |
Model identifier (e.g., gpt-4.1-nano, claude-haiku-4-5) |
--ns_ratio |
Namespace ratio (0.0–1.0) for controlling parallel execution |
For each model-strategy-task configuration, we report metrics along two dimensions:
Quality:
- Average — stratified bootstrap mean over runs; benchmarks treated as strata
- Run Deviation — typical run-to-run deviation from the strategy mean per benchmark
- Noise (Global) — variance of z-scored outcomes across all benchmarks
- Noise (Run) — average within-benchmark z-score variance
Cost:
- Same four metrics computed over token usage and wall-clock time, expressed in USD
Method hyperparameters are defined per task in YAML files under scripts/configs/:
# scripts/configs/game24.yaml
tot_bfs:
num_selections: 3
num_steps: 4
num_evaluations: 3
got:
num_selections: 5
num_steps: 4
num_generate: 10
num_evaluations: 3
num_best: 2Decoding parameters (temperature, top_p, max tokens) are sourced from scripts/configs/<task>.env.
ReasonBENCH is organized around four core abstractions:
- Method — specifies the reasoning strategy independently of the model or task. Integrates agents, the environment, and the model, and exposes a standard
solve()interface. - Environment — formalizes task-specific dynamics: state transitions, action validation, terminal conditions, and evaluation.
- Agent — defines the interface between methods, models, and states. Agents construct prompts, issue queries, and parse responses into actions.
- Model — uniform interface for LLM providers, supporting async execution and integrated with CacheSaver for response caching and deduplication.
src/
├── models/ # LLM provider adapters (OpenAI, Anthropic, Groq, Together, Gemini)
├── methods/ # Reasoning strategy implementations
├── tasks/ # Task definitions (state, environment, agents, prompts)
│ ├── game24/
│ ├── humaneval/
│ ├── hotpotqa/
│ └── ...
├── __init__.py # Factory registrations
├── typedefs.py # Core ABCs and type definitions
└── utils.py # Logging and utility functions
scripts/
├── simple/ # Single-run experiment scripts
├── repeats/ # Batch/repeated experiment scripts
├── cached/ # Cached inference scripts
└── configs/ # YAML and .env configuration files
datasets/ # Gzip-compressed task datasets
tests/ # Pytest test suite
pytest # run all tests
pytest tests/got/test_game24.py # single file
pytest tests/got/test_game24.py -k "test_x" # single testTests use async fixtures and require valid API keys (Groq/OpenAI) for the mock LLM clients.
@inproceedings{reasonbench2025,
title={ReasonBENCH: Benchmarking the (In)Stability of LLM Reasoning},
author={Anonymous},
year={2025},
note={Under review at ICML}
}