
#webperf #sitespeed #perfmatters
some people come to speed monitoring with an expectation to see the same results across different testing services.
this is common in a scenario of comparing service x to goog (pagespeed insights, http://web.dev measure, etc., which are treated as the only authority).
this is common in a scenario of comparing service x to goog (pagespeed insights, http://web.dev measure, etc., which are treated as the only authority).
it’s not an unreasonable expectation. in an ideal world, the results would be identical or very similar everywhere you track. that way, there are no trust issues in terms of reliability and no confusion in terms of metrics. teams focus on tracking & improving. perfect!
this expectation might be personal or driven by stakeholders. owners might be using goog tools as reference, focusing solely on speed for seo (thus only what goog reports matters), not knowing about variability or a myriad other reasons.
so, what happens when people start comparing speed metrics and scores?
they see divergent results. it brings confusion to the team about which results are "real" and "trustworthy". is there a bug? what’s happening?





many other things can happen too, but in many cases, comparisons bring confusion, frustration and lack of trust not only in specific tooling but speed monitoring in general. it might become even harder to convince them about the importance of speed. it's a lose-lose.
so why are results between speed platforms so different, you might ask? there's a multitude of reasons


different tools / frameworks = different results.

different infra = different results.


simulation off or on = different results.

with so many factors at play, it's impossible to compare metrics, and even more so, performance score (that consists of many metrics being weighted).
this is not the answer people want to hear, but it is true.
this is not the answer people want to hear, but it is true.
in a way, this feels like an impossible problem to solve, because we can't change how web's networking works or make all infrastructure everywhere the same to deliver on the expectation of the same results.
here's what we CAN do:
educate on how speed measurements are collected + what affects them
maximise reducing variability in results
understand with WHY people compare: what is their goal? what do they need?



if you are tracking perf, you can also familiarize yourself with those limitations and act accordingly when planning your speed strategy. i wrote about those limitations more extensively here: https://calibreapp.com/blog/common-mistakes-in-tracking-speed
also, because the most often compared metric (by far) is the performance | pagespeed score, i'd like to make it clear that the scores have no bearing on search engine ranking. core web vitals do. read more here: https://calibreapp.com/blog/site-speed-search-ranking-complete-guide
all in all:
comparing will cause confusion
you will inevitably see divergent results
knowing the reasons for variability is critical
broadly educating on variability will help with perf buy-in
think about ux and selected metrics, not only the perf score





