article thumbnail

Performance audit: Lego.com

Speed Curve

We track LEGO.com, along with a handful of other leading ecommerce sites, in our public-facing Retail Benchmarks dashboard , which I encourage you to check out. A couple of things worth noting: All of the sites in the leaderboard sites are pretty speedy, so this is NOT a name-and-shame exercise. Those are already big wins.

article thumbnail

How to Assess MySQL Performance

HammerDB

Instead, focus on understanding what the workloads exercise to help us determine how to best use them to aid our performance assessment. Benchmarking the target Two of the more popular database benchmarks for MySQL are HammerDB and sysbench. We will not concern ourselves with the raw throughput of workload. 4.22 %usr 38.40

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

An analysis of performance evolution of Linux’s core operations

The Morning Paper

A micro-benchmark suite, LEBench was then built around tee system calls responsible for most of the time spent in the kernel. On the exact same hardware, the benchmark suite is then used to test 36 Linux release versions from 3.0 It adds overhead to tests that exercise the kernel memory controller, even when cgroups aren’t being used.

article thumbnail

Evaluating the Evaluation: A Benchmarking Checklist

Brendan Gregg

These have inspired me to summarize another performance activity: evaluating benchmark accuracy. Accurate benchmarking rewards engineering investment that actually improves performance, but, unfortunately, inaccurate benchmarking is more common. If the benchmark reported 20k ops/sec, you should ask: why not 40k ops/sec?

article thumbnail

Evaluating the Evaluation: A Benchmarking Checklist

Brendan Gregg

These have inspired me to summarize another performance activity: evaluating benchmark accuracy. Accurate benchmarking rewards engineering investment that actually improves performance, but, unfortunately, inaccurate benchmarking is more common. If the benchmark reported 20k ops/sec, you should ask: why not 40k ops/sec?

article thumbnail

The top 5 reasons to run your own database benchmarks

HammerDB

Some opinions claim that “Benchmarks are meaningless”, “benchmarks are irrelevant” or “benchmarks are nothing like your real applications” However for others “Benchmarks matter,” as they “account for the processing architecture and speed, memory, storage subsystems and the database engine.”

article thumbnail

Why Browsers Get Built

Alex Russell

In some sense it's a confidence-management exercise. Distinguishing Traits # Shipped to many OSes Tiny platform teams (<20 people or <10% of the total team) Little benchmark interest or focus No platform feature leadership No standards footprint Platform feature availability lags the underlying engine (e.g.,