⏲️ 8 minutes

Relevance Scores: Breaking Down The Needs Met Rating

Search engines algorithmically weigh variables, such as the presence of words in the title, then produce a score based on those calculations.

Relevance scoring is a quantitative way to measure the relevance of a given piece of content. In the Google Rater Guidelines, one relevance score is referred to as the "Needs Met Score."

As a professional SEO, I rarely see non-technical information on relevance scoring. As a result, some outdated ideas on relevance scoring persist today.

For example, some SEOs overemphasize the importance of putting keywords in a URL to help a webpage rank better.

I have seen website administrators change all their URLs only to experience a significant drop in organic traffic and wonder why this well-known piece of advice did not help.

Google has stated that keywords in a URL have minimal impact on the Needs Met Score, contributing only about 0.001% to the overall score.

Ten years ago, however, words in the URL were a significant part of the relevance scoring system, but today they are a minor part, which explains why SEOs still tout the importance of keywords in the URLs.

What changed? URLs on mobile are truncated, and users rarely see them anymore. And keywords in a URL don't necessarily help sort content accurately.

Relevance Score Definition

According to Wikipedia, "relevance denotes how well a retrieved document or set of documents meets the information need of the user." It makes sense that putting words in a URL does not necessarily improve a user's needs-met rating.

Do Links Signify Relevance?

In the same light, it makes sense that links don't signify authority or relevance all the time; for example, when you're looking for expert advice, should the number of links matter?

Do you ask your Doctor how many links he has to his site before taking his opinion seriously? Probably no.

The idea of relevance is the whole premise of why search engines exist. Relevance scoring is what an SEO attempts to manipulate, regardless of the search system.

I put this post together to try and break down relevance scoring into simpler terms. Hopefully, by reading this article, you will come away with a deeper understanding of how search works.

What Is Relevance Scoring?

Information science, or informetrics, is the parent field of bibliometrics and webometrics. The idea of relevance applies to information science as a whole, and it helps users find library books or helps users find information on the web.

Relevance scoring has gone through many iterations. The earliest forms of relevance originated in the 17th century in attempts to sort scientific journals. According to Wikipedia, the study of relevance as a formal subject comes from the 20th century, under bibliometrics - the field used to sort library books.

Imagine the original informetric systems used in libraries, with library cards and old 1980's Tandy computers that pulled up books based on a numbering system, where they ranked the books alphabetically and by genre. These systems were some of the first search engines.

Old informetric system a phone book
Old Informetric System A Phone Book

In that same light. Recall the old phone books; they were a business directory dropped off at people's houses for free.

Businesses would pay a fee to be included in the "Yellow Pages." The phone book was one of the primary ways businesses got customers for decades. The Yellow Pages would rank businesses by business category and alphabetically.

As a result, companies started popping up with multiple "A's" in their business names, like "AAA Plumbing" or "AAA Hardware."

The use of repetition to manipulate a search system sounds oddly familiar.

Systems that rank results alphabetically may not serve the best search results for every query. The original web-based search engines had similar, easy-to-manipulate flaws as phone books.

Some of those web-based search systems are still in use today.

2022: My college's ancient web-based relevance scoring system

Government sites and colleges still use some of the old relevance scoring systems.

The search function on an old website might return the results alphabetically based on the book's title. Or, in the case of my college, alphabetically based on the title and used data from the entire course description.

I typed the search term "economics" in my college's course directory, and it returned courses that included the term "economics" anywhere in their descriptions, in alphabetical order.

Now for a college with 38 results, it is not a big deal if the course at the top of the search page is not an economics course. ANTH 301 matched the term "economics" because it contained the term in its course description.

If you searched for "books about SEO," you would find more than 103 million results on Google; the last thing you'd want is to return results alphabetically.

Let's see if we can come up with a way to score the content on my college's search directory to help improve the relevance of the search results.

Relevance scoring

One way to help my school's search engine return better results would be to count and weight additional variables that determine relevance. For example, you could look for the phrase in other places and count the number of times the term is repeated.

I counted the term "economics" 1 time on ANTH 301's webpage. I counted the term "economics" 10 times on ECON 200's webpage.

I found 11 repetitions of "economics" on ECON 201's webpage. And 13 repetitions of "Economics" on the ECON 202 webpage.

Based on the above assessment. My former college's search engine only looked for the inclusion of the keyword at least once; it did not count the repetitions or weigh the results based on other factors.

We could add variables such as repetition, meta description, meta title, and page title. Or any variable that we think is relevant.

Each variable could be assigned a number, which can be weighted if we choose. For example, a page with 8 repetitions could be represented by the number 8 or the number 4.12534536 on a logarithmic scale. The choice is ours. In the end, the sum of all the ranking factors must equal 100%.

Working with my former college's course index. A simple relevance scoring system could look something like this.

Search TermPage Location# of Repetitions Including FootersPage Title RepetitionsSummary Section RepetitionsOutcome Section RepetitionsAuthor ScoreSUM of Scores Excluding Footers
"Economics"ANTH 301101001
"Economics"ECON 20010242513
"Economics"ECON 2011110258
"Economics"ECON 2021310157

If it finds the search term in the title, it returns a "1"; if the term is in the title twice, it returns a "2." To keep it simple, we are counting repetitions and weighing results based on where those results are presented.

With R studio, for example, we could create something more complex. We could double the weight of the presence of the keyword in the title or stop counting repetitions after a certain number.

Weighing unnecessary repetitions

In the example content scoring system above, I have chosen not to count the repetition of the words in the footer section because they don't help us determine if the content is relevant.

Weighing expertise

It's hard to quantify an author's expertise and relevance to a given query without the necessary credentials. Why? Because without credentials, how do I know the content is expertly written? Should I take the author's word for it? Over a factor of near-infinite search terms, the problem is exacerbated.

Based on the authors' degrees below, who is objectively authoritative on economics? The Master of Economics or the Master of Anthropology?

An author with a degree in the field is a more qualified expert than someone without a degree.

Thus, we should weigh articles written by qualified experts higher than those written by unqualified experts. And that makes perfect sense; plenty of people give non-expert opinions; think of all the high-profile personalities. It doesn't matter how many articles a given personality has read. They are not a qualified expert.

Search TermPage LocationAuthor Score
"Economics"ANTH 3010
"Economics"ECON 2005
"Economics"ECON 2015
"Economics"ECON 2025

I added the field "author score"; this is either 0 or 5, based solely on the professor's education. If one or more professors teaching a class has an advanced degree in economics, the class receives a score of 5, while a class taught by a non-economics professor gets a score of 0.

The final content scores

So the total number of repetitions is higher than the content scores at the end.

I imagine every index resembles a giant excel file, with simple references to where a piece of data is located. For visual purposes, if you appended that data to your Index's imaginary giant excel file, it could look like this.

Term NumberTermCourse Name with Appended Data
1"Economics"ANTH 301[0,1,0,0], ECON200[2,4,2.5], ECON201[1,0,2,5], ECON202[1,0,1,5]
2"Math"2,34,91,118,546,239
3"Marketing"5,21,91,112,546,222,2137
3"English"1,19,77,92,116,239

As a result of the content scoring system, when users look for the term "Economics," they are shown results based on the content's relevancy score, which we calculate below.

  • ANTH 301 Content Score = [1+0+0] = 1
  • ECON 200 Content Score = [2+4+2] = 13
  • ECON 201 Content Score = [1+0+2] = 8
  • ECON 202 Content Score = [1+0+1] = 7

When a user is shown the results, the are now ordered by their content score: ECON 200, ECON 201, ECON 202, ANTH 301

ANTH 301 moved to the bottom of the search results
ANTH 301 moved to the bottom of the search results

We've improved my old college directory's search results. You're welcome.

But now it's time to move on to bigger things. The web is vast, and sorting four college courses might not cut it.

Also, with our current system, one could easily manipulate the search rankings by repeating the search term all over a page, which is precisely how the original search engines were manipulated.

Also, a system ranked alphabetically might be preferred in some contexts, like searching for classes. So maybe we did all this work and made the results worse?

Now that we've explored relevance scoring. We can try to combine this concept with the idea of backlinking.

How do backlinks work with relevance scores?

Modern informetric systems are much more complex than the ones used in school systems, if only for the size and complexity involved in contemporary content organization.

So you'd take the content rankings from above and overlay backlinks. You’d then have two scores, one for content within the index and one for the number of citations a piece of content receives; in theory, content with more citations is more authoritative. Content with a higher relevance score is more relevant.

Combined, these two scores would produce exciting results because you could technically have a higher-quality page, but another site could have a better link profile. They could both populate in the top spots. However, you'd also limit the number of results appearing for a given query to a reasonable amount.

If you're interested in learning about the latest backlink algorithm, I encourage you to read this blog.

Conclusion

Relevance scoring is an essential part of any search index. There are obvious limits to how it can be approached. To score content, the algorithm has to be able to read the content; how it weighs the content is entirely up to the engineers.

Machines are not yet smart enough to make subjective interpretations, so the machine has to be able to interpret the content easily, and it has to do so through text that exists in the same online matrix, in an easily digestible format. Meaning offsite and onsite factors have to be known - they are things on a webpage or in the code somewhere.

Things that we humans interpret as subjective, like expertise or quality or trust, have to be measured objectively; otherwise, the algorithm cannot understand them.

About the author

October 23, 2022
Cedric Chambers is a humble polymath. He studied painting as an undergraduate and business analytics in his master's program. Throughout the past decade, Cedric has worked in a number of industries.
December 23, 2023
The Importance of Quality Score for CPC

The results are surprising, Quality Score and Cost Per Click are more clearly correlated when you look at Exact Match keywords.

Read More
January 29, 2023
How To Style The WordPress Code Block

I love WordPress for its customizations. Styling code snippets enhances user perceptions. Copy and paste the code below to style your WordPress code blocks.

Read More
January 29, 2023
How to Add A Title Field To User Profiles In WordPress

The code snippet plugin changed my life. I no longer edit the raw PHP files to add simple functions like titles to users' profiles.

Read More
December 24, 2022
Adwords: Reducing CPA by 59%

By restructuring your account and using hyper-targeted ad text, CPA can be reduced.

Read More
December 16, 2022
The Meaning + Examples Of YMYL Pages

YMYL stands for "Your Money or Your Life." Google uses YMYL to help understand if a web page's main content could impact people's safety, health, or financial well-being.

Read More
December 15, 2022
A/B Testing With Optinmonster: Increasing Revenue By 64.89%

With slight modifications to Optinmonster's native A/B testing capabilities, it becomes an excellent CRO testing tool.

Read More
  • Lubble Favicon
  • OUR BLOG
    December 23, 2023
    The Importance of Quality Score for CPC

    The results are surprising, Quality Score and Cost Per Click are more clearly correlated when you look at Exact Match keywords.

    Read More
    January 29, 2023
    How To Style The WordPress Code Block

    I love WordPress for its customizations. Styling code snippets enhances user perceptions. Copy and paste the code below to style your WordPress code blocks.

    Read More
    January 29, 2023
    How to Add A Title Field To User Profiles In WordPress

    The code snippet plugin changed my life. I no longer edit the raw PHP files to add simple functions like titles to users' profiles.

    Read More
    December 24, 2022
    Adwords: Reducing CPA by 59%

    By restructuring your account and using hyper-targeted ad text, CPA can be reduced.

    Read More
    December 16, 2022
    The Meaning + Examples Of YMYL Pages

    YMYL stands for "Your Money or Your Life." Google uses YMYL to help understand if a web page's main content could impact people's safety, health, or financial well-being.

    Read More
    December 15, 2022
    A/B Testing With Optinmonster: Increasing Revenue By 64.89%

    With slight modifications to Optinmonster's native A/B testing capabilities, it becomes an excellent CRO testing tool.

    Read More
    Lubble's mission is to be the go-to resource for digital marketers and local business owners.
    Insights