Interesting post on benchmarking
By joe
- 2 minutes read - 214 wordsHere. In it, the author makes a number of points. Some I take no issue with, or don’t have direct knowledge of. Others …
Erp … You only get the “faster” speeds with easily compressible data. You get the far slower speeds when the data isn’t so easy to compress. We know. We measured this, and observed it. If you write all zeros, just like in the days when compilers special cased particular codes (cough cough), its possible disks don’t even do the writes. Yet when they have random data, its kinda hard to fake a write. So, I’d argue that the “wrong” results weren’t wrong. Just likely writing some non-compressible bits along with compressible bits. His conclusion is:
Yeah. I agree with this. Real measurement is hard. If people aren’t reporting averages over several cases, and scaling of results, and … then there’s probably a high likelihood of “wrongness”. FWIW, I’ve seen bad IO (and more generally performance) measurement everywhere, from popular blogs, articles from people who should know better, through supercomputer centers and national labs. The deer in the headlights rapid blinking isn’t fun to either experience, or provide the headlights for. Its better to ask for an independent review of results, to make sure they make sense. Peer review. Novel concept.