Don't miss the game changer LRT Smart | Contact Us+43 720 116440(866) 347-3660+44 800 011 9736(877) 736-7787
Home » Case studies » Discovering ways to improve your Google rankings

Discovering ways to improve your Google rankings

So how does Google decide who is going to be on the first page? It is a difficult job to try and work out what exactly Google is looking for, so the aim is try and look for some of the major signals that Google is likely to use to provide us with their Search Engine Results Pages (SERPs).

How to Find and Simulate Links in Bulk, with FIVE secret methods.
Understand how to simulate links and their risk before acquiring them.

How to Reverse Engineer the Top 10 Google Results

Understanding what Google wants

The secret of SEO today is to understand what exactly Google wants. You may hear some SEOs saying that you need social signals, press releases, editorial links etc. etc. To guess like this is inaccurate and dangerous. The only surefire way is to analyze your competition and see what they are doing and see why they are successful.
Google uses many signals to calculate search engine positions. This information is, of course, top secret, but by looking at the sites that already rank we can look at:

  • The type of backlinks that Google expects
  • How smaller more focused sites can compete with huge authority sites
  • How the geographical location of your web host server can affect rankings
  • The average backlink ratios (e.g. Start Page vs. Deep Link) in your niche
  • The acceptable Domain-Wide Link Detox Risk (Domain DTOXRISK™) for your niche
  • How Google determines the relevancy of your site

Table of contents

The Plan

Every day I read about people who think that Google has treated them unfairly. More often than not people complain about the sites that rank on the first page saying that they don't deserve to be there. Google isn't perfect, but it is getting better. It is not often that we see a really spammy site ranking well for a popular keyword like we did in 2012.

So how does Google decide who is going to be on the first page? Google is algorithmically driven and looks at many factors to determine ranking positions. Even in a hyper competitive niche like pharmacy, weight loss or payday loans Google must show results on the first page that keep users happy and make Google their search engine of choice.

I am going to take a keyword example , examine the pages found and see what we can learn about the results that Google shows.

Google's Algorithms

It states here :

... Google’s algorithms rely on more than 200 unique signals or “clues” that make it possible to guess what you might really be looking for. These signals include things like the terms on websites, the freshness of content, your region and PageRank.

It is a difficult job to try and work out what exactly Google is looking for, so the aim is try and look for some of the major signals that Google is likely to use to provide us with their Search Engine Results Pages (SERPs).

Villas for sale in Mallorca

For this example I decided to pick a search keyword close to my heart, 'villas for sale in mallorca', I'd love to live there one day and  it is a keyword that I am likely to Google. For those of you unfamiliar with Europe, Mallorca is a small island which is part of the Balearic Islands. The Balearic Islands are an archipelago of Spain in the western Mediterranean Sea. Mallorca is a favorite destination for the British and the majority of foreign property buyers in Spain are British. A search for this keyword on in English for all countries gives this result for the Top 10:


Keywords in the URL

Looking at the results we can see that most URLs do seem to contain related keywords, such as 'property', 'spain', 'balearic','realty' and the alternative spelling of Mallorca, Majorca. It is interesting to note that the word 'villa' is not in any of these results. Clearly Google does not require all the keywords to be contained in the URL. However it is clear that the structure of makes it very easy to understand what the page is about. The top result is a page called Mallorca.html in a directory called /overseas-property-for-sale/.

A JUICE Analysis at Page Level

To examine the 10 URLs I found, I pasted them into the JUICE Tool and got this result:


Here I was analyzing the actual pages, rather than the domains, that were found in Google. The interesting point here is the BL column: the top 2 results (both from and the result at #6 have no backlinks whatsoever. It seems the days of frantically building deep links are well and truly over. Links to the whole domains, seen in the BLdom column are a different matter . Our winner in 1st and 2nd place is that has more backlinks than any of the others. So now we should do some further analysis at domain level.

The Quick Domain Compare (QDC)

In a moment we are going to use the Competitive Landscape Analyzer  (CLA). The CLA can have an input of 10 competitor domains which can be compared against your own domain.  As the top 2 places are occupied by the same domain we now have a spare place, so I'll introduce #11 from the Google results.  At #11 is:

and for the sake of our experiment we can say that our site is the one that ranks at #12: is actually a huge site with phenomenal traffic, but Google only chooses to rank this at #12. If we have a look with the Quick Domain Compare (QDC) we can see some more details, showing how compares with our top 5 domains:

QDC results

Even though ranks for an incredible 622,512 keywords, there are two things that make it stand out.

  1. has a Power*Trust=25, whereas our winner has a very strong Power*Trust = 35. Power*Trust is certainly a very good metric to keep an eye on.
  2. is hosted in the US whereas the others are all hosted in Europe. at #1 is hosted in the UK and at #2 is hosted in Spain.

The Competitive Landscape Analyzer (CLA)

To make further analysis I set up the CLA like this:

CLA setup

Classifying the Keywords

After running the CLA there was a lot of keyword classifying to do, which was not quite as obvious as it first seemed. I had decided to use as if it was my own domain. The term ‘Realtor’ is not one that I am very familiar with. Realtor is a trademarked brand name that is short for The National Association of Realtors. While the term real estate agent describes anyone who is licensed by their state to work in the real estate industry, Realtors are members of a national association. After some research I realised that many people may search for ‘realtor’ rather than ‘real estate agent’ not realising that it is a brand keyword rather than a money keyword. This is excellent branding work from I classified 'realtor' as a brand keyword.

CLA Results

The Domain Statistics table shows what you would expect, Google favors the domains with the higher Power*Trust, but with a few exceptions:

CLA Dom Stats

The result that did surprise me here was the 9th result which has a low Power*Trust and only 12 Referring Root Domains. Looking at the site it does seem to have a very clean looking, responsive theme, thus providing a good user experience. By using this Google query:

site: Villas for sale in Mallorca

we can see if our keyphrase exists on the site. It does not, but a semantic version of it is evident in the Google results. Google helps us out a little here by making the keywords from your search bold as well as the related keywords:

Mallorca property house

It seems that ‘Properties for Sale in Mallorca’ is an acceptable related keyword to ‘Villas for sale in Mallorca’. The Google search has found a related keyword on the homepage of the site and all is well. If we look at the homepage of the site we can see the text, which is actually an H3 heading, 'All properties for Sale in Mallorca'. Below the heading are links to seven inner pages, one of which is, 'Villas / Houses in Mallorca'. This shows that Google has found some very relevant content which it ranks in the Top 10 with hardly any backlinks and no keyword stuffing.

Mallorca property house

Brand, Compound, Money and Other Anchor Text

The CLA shows the ratios of anchor text links with an easy to read histogram. The very dark green bar represents the average of the Top 3 sites, the dark green one is the average of the Top 5 sites and the light green one is the average of all the domains. The orange bar represents 'you' which in our case is

CLA brand

These ratios are different in every niche, in every country, for every keyword and for every language. In our example, I would be aiming to have anchor text with 79-83% brand keywords, 3-5% compound keywords, 8-12% money keywords and 4-6% other keywords.

Follow vs. No Follow Links

Sites that have all DoFollow links look highly unnatural and can get devalued by Google. Most normal sites have a mixture of DoFollow and No Follow links.

CLA follow

To rank in this niche I would be aiming for 78-82% Follow Links and 17-21% No Follow links.

Text vs. Image Links

A normal site will have a mixture of text vs. image links. It surprised me that in our example there are quite a lot of image links:

CLA image text

To rank a site here I would be aiming for 55-66% text links and 33-44% image links. The average is being pushed high by who have a whopping 52% of their links as images. Many of these do look a bit dubious such as this DoFollow one in the footer of a Scottish Estate Agent:

Cairn estate agent

I can't see any reason why Cairn Estate Agency would have such a link on their site, unless they were being remunerated somehow by Later you will see that I ran a Competive Link Detox report which showed this link to trigger the SUSP11 rule: Sitewide Footer Link detected, a typical way that paid links or 'friend' links are artificially placed.

Link Location, Deeplink Ratio and more

So far we have understood that the sites which Google ranks in the Top 10 have similar ratios to each other. Our example of (the orange one) has been standing out and not blending in with the rest. Even though it is a huge authority website, for our phrase it ranks the worst. There are lots of factors that Google take into account when calculating SERP positions, but certainly the ones we have looked at so far show us what Google expects. This pattern continues when we look at the Link Location:

CLA link location

...and the Deeplinking Ratios:

CLA deep linking

With the CLA we can optionally choose to include extra metrics like social votes from Twitter, Facebook and Google+, Dmoz and Wiki Links and much more. In each case you can take a look at your link profile and see how you can improve it to blend in with your competition.

Common Backlinks

It is highly likely that results on the Google SERP will share some common backlinks. The theory here is that domains with more common backlinks will be rewarded for their relevancy.

The goal of Google is to give a good user experience by providing relevant results.  For example this page you are reading now will probably not rank well for the phrase 'Villas for sale in Mallorca' even though I have mentioned the phrase several times. It is unlikely that will share any common links with our examples. For somebody wanting to buy a villa in Mallorca, would be a bad experience, for somebody learning about SEO it will hopefully be the opposite.

This is one area where Private Blog Networks (PBNs) completely fail to work as it is highly unlikely for competitors to have common links from a PBN. We can discover more about common backlinks using the Common Backlinks Tool (CBLT).

CBLT Commonality

Certainly the Google #1, has the most common backlinks, 391 with coming second with 380. Between these in the Google SERP is which has just 20 common backlinks. If Google is taking the common backlinks into account then they are almost certainly taking into account the topic and content of the site too. and both focus on UK properties with an international section, whereas only have properties on the islands of Mallorca and Ibiza. It would not make sense for to have as many common backlinks as

You can filter the CBLT Results Table to find the best links with the most commonality. Whilst doing this I did spot a Hungarian Link Directory, that was linking to 7 of our 10 domains. This looks very spammy, so now is a good time to take a look at the Link Detox Risk (DTOXRISK™) of these domains.

Competitive Link Detox (CDTOX)

The CDTOX compares your domain with 10 competitors and determines the risk associated with their backlinks. In this example our domain is which is compared to the 10 domains that we used above. As these sites are all ranking well for a popular keyword, we are not expecting any high Domain-wide Detox Risk scores (Domain DTOXRISK™) here. Generally a Domain DTOXRISK™ of around 700 is OK. A Domain DTOXRISK™ > 1000 is where we start to see problems, but this will always depend on the niche and some other factors. The total average is 784 which is about what I would expect:

CDTOX tachometer

The winner, has a below average Domain DTOXRISK™ of 284. do have some questionable links, but they seemed to be outweighed by many good ones. CDTOX shows:

  • 1052 Links with Low Link Audit Priority
  • 5869 Links with a Medium Link Audit Priority
  • 422 Links with a High Link Audit Priority do not fair so well, with a high Domain DTOXRISK™ of 1385. CDTOX shows:

  • 49 Links with Low Link Audit Priority
  • 280 Links with a Medium Link Audit Priority
  • 280 Links with a High Link Audit Priority

In these examples we do not know if any of these sites are using a disavow file. do not need to worry about this at the moment, but do. Without a disavow file are heading for Google Penguin hell. also have a high Domain DTOXRISK™ of 1402, but I would hope that a company of that stature would protect themselves with a good disavow file.


Link Detox Screener

Here are the main takeaways:

  • It is no longer necessary to have your exact keyword in your URL, but some related keywords seem to help
  • It is not necessary to have backlinks to an inner page for it to rank
  • Power*Trust is always going to be a major factor in ranking, but other factors do matter as well
  • Hosting a site in or near the country where your target audience is seems to be a good idea.
  • A smaller less powerful site can rank alongside a huge site if its content is more focused to the search query
  • Keyword stuffing is no longer necessary or advised. Using a few related keywords in your content is a much better idea
  • User experience should be one of your prime considerations when creating a website
  • You should always monitor your site's backlinks and try to blend in with your competition, paying attention to the average link ratios for your niche. Pay particular attention to brand vs. money anchors, No Follow vs. DoFollow, text vs. image links, startpage vs. deeplinks
  • Acquiring common backlinks seems to be a very good idea
  • There are quite wide variations in the Domain DTOXRISK™ for the domains that rank in the Top 10. Although there is no direct correlation with ranking, you should try and have nothing higher than an average Domain DTOXRISK™
  • Having some bad, high risk links is OK as long as you have plenty of good ones too

It is important to realise that your backlinks will change over time and Google will adjust their algorithms every day. For this reason you should monitor your link profile on a regular basis. I would recommend that you use, at least, Link Detox on a monthly basis.

superhero smallWith the Superhero Plans you can perform link audits, link cleanup, link disavow boost, competitive research, professional SEO and backlink analysis for your own or your competitor's sites.

You can fix or avoid a Google Penalty! Learn more about how you can Recover and Protect with LRT.




Rick built his first website in 1997 when the main search engine was Altavista and the word Google didn’t exist. Since then he has learnt SEO from the ground up and has lived through every update Google has thrown at us. Rick now has a great deal of knowledge about healthy link maintenance and link risk assessment. In these days of manual and algorithmic Google Penguin penalties it is more important than ever to monitor your backlink profile very carefully, this is where his expertise is ... 


  1. @lnkresearchtool on February 23, 2015 at 11:34

    Discovering ways to improve your Google rankings

  2. @Bojan_Miljevic on February 23, 2015 at 12:18

    How does Google decide who is on the first page? via @LnkResearchTool

  3. @serpolytics on February 23, 2015 at 13:10

    How does Google decide who is on the first page?

  4. @Irfan84_Yaqoob on February 23, 2015 at 14:43

    How does Google decide who is on the first page? via @LnkResearchTool

  5. @GamblingAffPro on February 24, 2015 at 02:33

    An insight into how Google may decide which sites appear on the first page #seo #google

  6. @mmhemani on February 24, 2015 at 04:33

    @ricklomas this is mind blowing: i am your newest and probably the biggest fan 🙂

    • Rick on February 24, 2015 at 09:45

      Wow! Thank You!

  7. @linkmyseoworld on February 24, 2015 at 06:49

    RT @adeelzahor: How does Google decide who is on the first page? via @LnkResearchTool

  8. @EJDijks on February 24, 2015 at 09:59

    How does Google decide who is on the first page? #SEO

  9. @WebsiteValueOT on February 24, 2015 at 10:00

    RT @EJDijks: How does Google decide who is on the first page? #SEO

  10. @JavieRuizGago on February 24, 2015 at 11:12

    How to Reverse Engineer the Top 10 Google Results –

  11. @hellosocialyou on February 24, 2015 at 12:55

    “User experience should be a prime consideration when creating a website.” SEO considerations via @lnkresearchtool

  12. @marccloutier89 on February 24, 2015 at 15:48

    How does #Google decide who is on the first page? via @LnkResearchTool

  13. Tommy on February 25, 2015 at 09:27

    We know backlink is important but do we really need to disavow every low quality links? Can’t Google determine which link is bad and good?

    • Rick Lomas on February 25, 2015 at 11:21

      Tommy, you don’t have to disavow every bad link, but you should aim to have a similar Domain-wide DETOXRISK as the sites on the first page of Google. If this means disavowing a lot of useless and potentially harmful links, then do it. If you have been doing 2011 style link building then this of utmost importance. Just blend in and then try and do just a tiny bit better 🙂

Leave a Comment

Clients speak for LinkResearchTools

William Sears - Growth & SEO at LinkedIn

William Sears

"A huge part of SEO success is your backlink portfolio.

 Not only do you need to understand where your links are coming from, you need to be able to take action to manage those links. LinkResearchTools is an indispensable collection of powerful tools to do just that. I’m a long-time customer of LinkResearchTools and highly recommend it to anyone who is serious about crushing it with links."

Kenneth Chan, Founder and CEO - Tobi


"Your tools are the best in the industry.

 The service is great.

Christoph, your passion is contagious."

Larry Markovitz -  Sr. Director of Organic Search at GroupM 


"Continuous improvement of our client's results

LinkResearchTools and specifically Link Detox gives us actionable insights that continuously improve our clients results in the search engines."


Read what other happy users of LRT say and see all the companies that licensed LRT.

Recover - Protect - Learn - Grow your SEO - Learn how LRT can help you.

Have you seen enough to make a decision?

Want to test with real data? Discounted trial for small websites. Just the export doesn't work.

LinkResearchTools is SEO Software for Link Analysis and Link Building

Trusted by the world’s most respected brands.

victoria' secret