Chuyển đến nội dung chính

Performance as a Ranking Factor: The State of the Web and Core Web Vitals [Part 3]

In part one of this series, we talked about how Google and the web in general were not really ready for the Page Experience Update — Google’s CrUX data covered too few websites, the vast majority of which were not hitting the required thresholds. That was why, I suggested, the update had been so delayed and watered down.

In part two, we talked about the metrics themselves — their flimsiness, their arbitrariness, their openness to manipulation. This, too, I suggested, might be holding Google back.

However, the proof is in the pudding. Are Core Web Vitals, taken individually or as a whole, correlated with rankings? If so, is that any more true than it was before the Page Experience Update? In this third and final post of this series, we’ll see what the data tells us about the relationship between Core Web Vitals metrics and organic ranking performance.

Viewer discretion advised

This is, at most, a correlation study. There are many mechanisms by which something can be correlated with rankings without having directly influenced rankings.

For example, perhaps websites that take SEO seriously rank well, and also tend to work on their loading performance. If so, loading performance and ranking would be correlated even without any direct causal link.

We’ll talk through potential implications as we go, but please, proceed with caution!

Performance of passing vs. failing URLs

To start with, I decided to look only at the URLs that had CrUX data in the first place. You may remember from part two that, at the time of the update rolling out in August this year, that was some 38.3% of URLs. This is taken from the top 20 results for 10,000 MozCast keywords, across mobile and desktop device types.

Note that these URLs are all taken from the top 20, so it’s interesting that the averages are both well above the rank of 10.5 we’d expect. This is likely because higher traffic URLs are disproportionately likely to rank well, and also disproportionately likely to have CrUX data.

We see a solid 0.39 ranking position lead here for the URLs that pass all three CWV thresholds, above those that fail at least one.

Does that mean this is a ranking factor?

On the face of it, the above data looks very promising for CWV as a ranking factor. However, it’s worth tempering our excitement a bit.

Let’s have a look at the same data but from May, before the Page Experience update rolled out:

We can notice a few things here:

  • The average rank of URLs with CrUX data was generally worse in August than in May. This is to be expected, as more URLs had CrUX data by August, so it had worked its way further down the rankings.

  • URLs which pass the CWV thresholds already had a ranking difference even before the update. This suggests that perhaps URLs which pass the test were already better in other ways that already counted towards rankings (for example, perhaps rankings were rewarding URLs with a good user experience).

  • The difference between URLs which passed the thresholds and those which did not has grown from 0.38 in May to 0.39 in August — although this is probably very easily within the margin of error.

It’s also interesting to contrast with a performance metric which was not part of the Page Experience update: Speed Index, as reported in Lighthouse lab results.

As “passing” the three thresholds for CWV represents being in the top 36.3% of URLs by that metric, we can compare what ranking difference is associated with being in the top 36.3% for Speed Index.

We can see in this chart that Speed Index, despite not being an explicit ranking factor, has a modest improvement in average rank associated with this percentile breakdown (0.17, vs. 0.39 for passing all three CWV thresholds). This doesn’t mean that Speed Index is a ranking factor, it just means that these things can be related in more complex ways.

(If you’re a mathematics nerd like me and you’ve just noticed the weighted average rank of the two groups is not the same, that’s because there are a tiny number of URLs for which I was able to obtain CrUX data, but not lab data, due to server errors, etc.)

So did anything happen?

Actually, yes. But it’s more subtle.

The real impact was felt for URLs that failed all three tests. Although these URLs often started out ranking best of all (probably because they disproportionately represent some important, household name brands) they’ve taken a hit with the update. These URLs have had a 1.15 position ranking drop, compared to around 0.2 for URLs with CrUX data taken as a whole.

This, as I mentioned in part one, is different to what Google set out to do. Back in the original FAQs for the update in 2020, Google said:

"If a page hits the recommended targets for all three metrics, it passes the web vitals assessment… The page experience ranking impact will be the same for all pages that are in the good range for all Core Web Vitals, irrespective of their individual Core Web Vitals scores."

(emphasis mine, original source here)

For all the data reasons I covered in part one, likely they weren’t able to do this, and had to improvise a bit, instead only applying the relative penalty (or absence of a boost) for URLs that failed all metrics, rather than for URLs that failed one or more.

Wait, so all I need to do is pass one metric?

Well, no, that’s not quite the attitude. There are still lots of other reasons to want to pass all three, and more importantly, to have a generally good page experience. Google is only going to be looking for more ways to augment and ramp up these factors over time.

Also, the rest of SEO still counts. Check out the rather more pronounced difference associated with Page Authority, for example:

Good luck out there :)


Ready to see if your site's pages are passing one CWV or none? Head over to Moz Pro and check out the Performance Metrics beta within our Site Crawl toolset.

Sign up for Moz Pro to access the beta!

Already a Moz Pro customer? Log in to access the beta!

Nhận xét

Bài đăng phổ biến từ blog này

Three Steps to a Better-Performing About Page

Posted by AnnSmarty Somehow, many businesses I’ve come across online have one glaring problem in common: a very weak and unconvincing About Us page. This doesn’t make any sense in my mind, as the About page is one of the most important brand assets, and unlike link building and social media marketing, it doesn’t require any ongoing effort or investment. An About page is often part of a buying journey. It can drive people to your site and help convince them to deal with you. And, in these uncertain times , you can use it to help build trust in you and your business. Creating a solid About page is a one-time task, but it will boost both brand loyalty and conversions for many months to come. Why is your About page so important? It is often an entry page Whether you’re a business owner or blogger, your About page tends to rank incredibly well for brand-driven search queries (those that contain your name or your brand name). If nothing else, it shows up in your sitelinks: Or your m

How Lead Generation Tactics Can Boost Your Link Building Results

Posted by AnnSmarty How effective is your link building campaign? I bet your answer is “I wish it could be better.” Talking to business owners and executives on a daily basis, I have yet to meet one who would be satisfied with their link building strategy. Everyone needs links, yet they are getting harder and harder to get. The solution? Change your link building mindset. How link building is similar to lead generation In any business marketing strategy, we’re really interested in one thing: sales. Yet, if we keep focusing on that end goal, we won’t achieve much. A customer may need up to eight touchpoints before they finally make a purchase. If you only focus on that final sale, you’re missing out on all those extra steps that drive your customer to buy. It may sound obvious (so I’ll stop here) but the point I’m trying to make is: Marketers cannot focus on the final sale. We need something in between — a secondary metric that will bridge the gap between “a stranger” and a “

How to Improve Organic Clickthrough for Your Content

Google search result pages are becoming more diverse and even interactive, which makes any clickthrough study out there much less reliable, because no two sets of search results are the same. But how much control do writers and content creators have over how their content is represented in search? As it turns out, they do have quite a few options when it comes to optimizing their search snippets! The anatomy of a standard search snippet The standard Google search snippet has changed over the years, but in essence all the key elements are still there: The clickable title or headline of the snippet (in blue) The description of that page (about two lines long — it was lengthened for no particular reason a few years ago, but now seems to be back to two lines) The URL path (used to be in green, now it is black) On a mobile device, there’s also a tiny logo next to the URL: Here’s how much control you have over these standard elements of your search snippet (in the order th