Compare benchmark suites (mean + variance)
All computation runs locally in your browser

The Benchmark Builder is a lightweight tool for comparing multiple sets of timing measurements (called “suites”). You paste or type the numbers you measured (for example, execution time in milliseconds), and it instantly shows a ranked table with each suite’s mean and variance.
What problem does it solve?
It helps you answer questions like: “Which implementation is faster on average?” and “Which one is more stable?” without manually calculating statistics.
Developers
Compare two versions of a function, query, or API call.
Teams
Share a standardized summary in a PR discussion.
Learners
Build intuition for mean vs variance using real numbers.
If you often need to share structured results, you may also like our List Converter for quickly formatting copied data.
Worked example (two measurements)
Suppose Suite A has two runs: and .
The mean tells you the typical runtime. The variance tells you how much the runs fluctuate. If two suites have similar means, prefer the one with smaller variance (more stable performance).
Comparing two implementations
Background: You rewrote a function and want a quick sanity-check before merging.
Inputs: Suite “Old”: 8, 9, 10 ms; Suite “New”: 7, 7, 8 ms
Result: New has a lower mean (closer to ) and smaller variance.
How it helps: Use the exported Markdown table directly in the PR description.
Tracking a regression
Background: A change made the API slower and you want to quantify the impact.
Inputs: Suite “Before”: 120, 118, 125 ms; Suite “After”: 160, 155, 170 ms
Result: The mean increases (positive delta), and the ratio highlights magnitude.
How it helps: Paste the bullet list into a ticket so anyone can read it quickly.
Comparing multiple strategies
Background: You have 3–5 approaches (cache, batching, different queries).
Inputs: One suite per approach, 10–30 samples each
Result: The table ranks suites by mean (best first) while variance reveals consistency.
How it helps: Pick the best mean unless high variance would hurt user experience.
If you want to quickly normalize or reformat your raw measurements before pasting them here, try Text to ASCII Binary for quick transformations and copy-friendly output.
PR performance notes
Summarize benchmark runs in a consistent table.
Micro-optimizations
Check if a change actually moved the needle.
Choosing a default strategy
Compare multiple approaches and pick the best mean.
Stability checks
Spot noisy measurements via high variance.
Sharing results
Use Share + Favorites to keep common configs.
Quick exports
Copy Markdown for docs or bullet list for tickets.
When it may not be a good fit
If your benchmark methodology is inconsistent (different machines, warmup not accounted for, background tasks running), the numbers can be misleading. Use the tool to summarize data — but make sure your experiment design is sound.
Use enough samples: Two samples are fine for a quick check, but 10–30 runs usually give a more reliable mean.
Keep units consistent: If one suite is in ms and another in µs, your comparison will be meaningless.
Watch variance, not just mean: High variance often means noise, GC pauses, or unstable environment.
Export the exact numbers: Copy as Markdown for PRs; bullet list for issues and tickets.
For each suite, the tool computes the mean and the population variance. If you have measurements , then:
The results table ranks suites by mean (lowest mean first). For any suite that is not the best, the tool also shows a comparison against the best mean: the absolute delta and the multiplicative ratio.
Mean summarizes typical performance. Variance summarizes stability. In benchmarking, both matter — users feel slow spikes even if the average looks good.
If you want to share results, use the Share button to generate a link and Favorite to save a configuration. When you’re done, Reset clears everything.
For timing benchmarks, lower usually means faster. The tool sorts suites by mean ascending. If your metric is “higher is better”, interpret the ranking accordingly.
It uses population variance: . This matches many quick benchmark summaries where your measured runs are treated as the full dataset.
It compares each suite to the best mean. is the absolute difference, and is the ratio to the best.
Yes. Use “Copy as markdown table” for docs and PRs, or “Copy as bullet list” for tickets and chat.
No — it runs locally in your browser. Share links only include what you choose to embed in the URL.
This tool summarizes the numbers you enter; it cannot fix benchmarking methodology. For reliable comparisons, keep environment and test conditions consistent (hardware, load, warmups, and cache effects).
The results are informational and should not be treated as a guarantee of real-world performance.
For a deeper dive into descriptive statistics:
Use a simple chronometer (stopwatch) to track elapsed time down to milliseconds. Runs locally in your browser.
Normalize email addresses to a standard format for easier comparison. Useful for deduplication and data cleaning. Runs locally in your browser.
Estimate the time needed to consume a total amount at a constant rate, and get an expected end time. Runs locally in your browser.
Parse and decode your JSON Web Token (JWT) and display its content. All computation runs locally in your browser.
Know which file extensions are associated to a MIME type, and which MIME type is associated to a file extension. Includes a full MIME types table.
Generate random Lorem Ipsum placeholder text with customizable paragraphs, sentences, and word counts. Runs locally in your browser.