Giter VIP home page Giter VIP logo

failed_backlink_analysis's Introduction

How Confirmation Bias Ruined My Backlink Data Analysis

In this page, I will show how confirmation bias has ruined my data analysis project, for which I spent countless hours. Had I done a very simple check in the beginning of the analysis, I probably would have been able to avoid this mishap. I am writing this so others won't repeat the mistake I made, hopefully.

Basic Terms

Definition of backlinks from Wikipedia:
Backlinks... are incoming links to a website or web page. Backlinks were originally important (prior to the emergence of search engines) as a primary means of web navigation; today, their significance lies in search engine optimization (SEO). The number of backlinks is one indication of the popularity or importance of that website or page (for example, this is used by Google to determine the PageRank of a webpage).

Definition of link building from Wikipedia:

link building describes actions aimed at increasing the number and quality of inbound links to a page.

History of Backlinks

Before Google introduced its Penguin algorithm in April 2012, many SEO specialists participated in spammy link building, in which they attempted--and often succeeded in--manupulating Google's search engine algorithm to boost their websites into users' top search results. They did so by placing high volumns of low-quality incoming links from other websites to their client's website without examining topical relevancy or trustworthiness of those linking sites. In other words, they placed a high priority only in the number of these backlinks, not the quality. This was a common practice in the SEO industry, including the company for which I am employeed now.

However, in the ever-changing world of SEO, Google announced its revamped search ranking algorithm in April 24, 2012, known as the Penguin update. Google Penguin caused a major historical shift in the world of SEO as it penalized websites benefitted from artificial spammy link building activities. As a result, many websites lifted in this fashion dropped from users' search results like dead flies. According to one poll, as much as 63% of website owners reported that their sites were negatively affected, with many never recovering from the hit.

What I Tried to Examine

When I first joined my current SEO company in February 2013, I was told that a former employee (to whom I will refer as "John") was an active link builder whose techniques became outdated, and in fact, harmful for SEO after the Penguin update. John left the company right before Google announced its Penguin algorithm, and when the company learned about it, it decided to deploy a new link building strategy for its SEO clients. Just curious, I wanted to examine how the company's new link building strategy improved over John's old link building methods.

My Misled Process

I started by downloading spreadsheet data for the 11 SEO clients' websites that existed either before or during John's employment. I didn't download data for the new SEO clients that came after Google's Penguin update because I wouldn't have any backlink data dating back to April 2012, which was fine. I also did not bother downloading data for non-SEO clients' websites since I figured I would have no use for them (first mistake). After all, I was trying to analyze the link building activities for our SEO clients, not for non-SEO clients. Just like that, I failed to establish a baseline for my findings.

After downloading and cleaning the data, I plotted the number of new sites linking to our clients' websites by time, which rendered this:
Alt text

I was interested to see that there was a noticeable peak near the end of 2011 that soared through the chart, followed by two additional waves of fresh linking websites, which I took as an evidence of link building activities (second mistake).

I placed color-coded timeframes onto the graph to provide a better visualization.
Alt text
Alt text

It should be noted that, while I was certain that John left his job sometime in mid-April, I did not have his definitive employment start date; I only knew he started working for the company sometime in the latter half of 2011. So instead of matching his exact employment start date, I looked at the above plot and assumed that John's employment began sometime near mid-September, 2011 (third mistake).

Looking at the graph, I noticed that there continued to have been additional websites that were linking to our SEO clients' websites after John parted the company. That should have raised a red flag. And it did, when I first started this analysis 6 months back. However, when I resumed the analysis couple weeks ago, I justified the finding by noting that John's used to hire a third party agecy, which I presumed to have continued to participate in link building activities even after John's departure (fourth mistake).

Lastly, I had no hard evidence to attribute the second wave of additional linking websites (marked by purple) to our company's new link building strategy. I assumed that it was the work of ours simply because there was a noticeable peak and I assumed that small business websites don't normally accrue backlinks on their own (fifth mistake).

Below is an excerpt from my initial report write-up:

We can clearly see that there had been continued link building activities, even after John left the company. That's because John had utilized a third party agency to do his link buildings. So even after he left, the third party agency continued to create backlinks because the orders had already been submitted by John. Hence, the green marks the (exntended) effective timeframe for John's link building work. After a hiatus, my company hired another SEO consultant under whom it resumed its link building activities with a new strategy and that's the purple mark on the graph.

Plotting the number of additional sites linking versus time for each of our SEO client's site only further confirmed my previous (misled) notion.
Alt text
Alt text
Alt text
Alt text
Alt text
Alt text
Alt text
Alt text

After nitpicking on the looks of each graph, I decided to analyze the differences of the linking sites created under "John's link building program" (call it Old Link-Building group) against those created under the company's "new strategy" (New Link-Building group). There were three things I noticed.

  1. Interestingly enough, there were less backlinks per site in the New Link-Building group (average of 12.23 backlinks per site) than in the Old Link-Building group (average of 14.91 backlinks per site). That made sense, provided that our company focused more on the quality of backlinks rather than the sheer quantity. Under the new strategy, we in fact focused on creating fewer but better quality backlinks (as opposed to more but lower quality backlinks). It sure felt like I was off to a good start.

  2. I also compared the average trustworthiness (Trust Flow metric) of linking websites in the Old Link-Building group to that in the New Link-Building group. Because I presumed John participated in excessive spammy link building during his employment, I expected to see lower trustworthiness score (a metric provided in Majestic SEO's data) in the Old Link-Building group than in the New Link-Building gruop. In this area, however, I observed no significant discrepency. Alt text

  3. Lastly, I compared the average site authority score (Citation Flow metric) of linking websites in the Old Link-Building group to that in the New Link-Building group. The average site authority score in the Old Link-Building group: 19.83. The average site authority score in the New Link-Building group: 23.58. This finding partially reaffirmed my belief that the new link building strategy was creating incoming links from websites with higher site authorithy (i.e. better quality, more legitimate websites), although I hadn't yet determined its statistical significance.
    Alt text

How I Found My Mistake

Toward the end of my analysis, for the sake of completeness, I decided to do a last-minute plotting of the new linking sites versus time for non-SEO sites. I anticipated a flat graph with no high spikes. However, this is what I got:
![Alt text](./output/linking_sites_vs_time_non_seo.png)

The graph looked almost identical to plot I drew for SEO-only clients. I couldn't believe it. How could the graph have signs of link building activities when our company (neither under John or under the new link building strategy) created backlinks for them?
For a while, I had to re-check that I wasn't redrawing the first graph and that I was using the right data.

Conclusion

So what happened? I cannot say until I conduct a further analysis. But it appears that there were spammy web directories that troll around and add spammy links to other websites found on the Internet for both SEO and non-SEO websites. The high peaks in the graph were results of neither John's link building work or the new strategy.

The lesson from this failed data analysis project is that one must ALWAYS, ALWAYS, ALWAYS have a baseline to compare and contrast his findings. In other words, there always needs to be two groups: control and variable.

I also learned that I should never, ever, ever, ever, ever assume anything. Confirmation bias is a trickery of mind. One should avoid falling victim to the trick of his own mind by placing heavy caution against confirmation bias (or any kind of bias for that matter).

Note

1. It is unclear how often Majestic SEO crawls sites to accurately determine "first link date" and "last link date". However, it is highly doubtful that their web crawler indexes every website in the world every day. (With such high data volumn, it would be an impossible feat.) Therefore, the first and last link date does not necessarily reflect the actual dates but rough estimates thereof. This is also mentioned in the Majestic SEO's website: "Due to limitations of storage, the first crawl date degrades in accuracy after 150 days." The analysis accounted for this
  1. For the sake of our SEO clients' privacy, their business names and web URLs have been masked. Also, "John" is a pseudonym for the discussed former employee.

Folder Structure

data: This folder contains the raw spreadsheet data downloaded from Majestic SEO.
output: This folder contains the visual graphs produced from analyzing the dataset.
script: This folder contains the R script that was used to analyzed the spreadsheet data.

Consideration for Further Analysis

Going forward, I would like to know how some of these small business websites accrused such high volumns of unnatural backlinks from topically unrelated websites with spammy contents. That is, what triggered some sites to be the target of backlinks while others have not?

After answering this question, it would be extremely interesting to see if a machine learning or cluster algorithm can be applied to the raw data to distinguish spammy backlinks from legitimate incoming links. If we can successfully pick out all the bad, spammy backlinks automatically, then my SEO company can submit a request to Google through the Disavow Tool, removing all spammy backlinks from Google's consideration when it calculates search rankings for our clients' websites, hence avoiding adverse search rank penalty from Google.

failed_backlink_analysis's People

Watchers

James Cloos avatar Howard avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.