Remove Benchmarking Remove Exercise Remove Latency Remove Testing
article thumbnail

Real user monitoring vs. synthetic monitoring: Understanding best practices

Dynatrace

These development and testing practices ensure the performance of critical applications and resources to deliver loyalty-building user experiences. Because pre-production environments are used for testing before an application is released to end users, teams have no access to real-user data. What is synthetic monitoring?

article thumbnail

Evaluating the Evaluation: A Benchmarking Checklist

Brendan Gregg

These have inspired me to summarize another performance activity: evaluating benchmark accuracy. Accurate benchmarking rewards engineering investment that actually improves performance, but, unfortunately, inaccurate benchmarking is more common. If the benchmark reported 20k ops/sec, you should ask: why not 40k ops/sec?

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Evaluating the Evaluation: A Benchmarking Checklist

Brendan Gregg

These have inspired me to summarize another performance activity: evaluating benchmark accuracy. Accurate benchmarking rewards engineering investment that actually improves performance, but, unfortunately, inaccurate benchmarking is more common. If the benchmark reported 20k ops/sec, you should ask: why not 40k ops/sec?

article thumbnail

A peculiar throughput limitation on Intel’s Xeon Phi x200 (Knights Landing)

John McCalpin

There was no deep goal — just a desire to see the maximum GFLOPS in action. The exercise seemed simple enough — just fix one item in the Colfax code and we should be finished. Single-core testing with the same DGEMM routine showed maximum values of just under 72% of the nominal peak (about 90% of “adjusted peak”).

Latency 40