matt-cutts-scraper-sites
SEO 101

SEO 101: Are You Being Outranked by Your Own Content?

Unique, and original, content drives the Internet (so we’re constantly told), and we SEOs spend much of of our time producing and honing that content in the hopes of making the all mighty Google happy as a clam.

Unfortunately, there has always been an ongoing problem with spammy websites lifting existing content verbatim, and passing it off as their own creation. These scraper sites not only devalue the original content they have stolen, but they can also impact the page rankings of the original page.

Many of us have been affected by scraper sites over the years, and there is nothing quite as nerve wracking as having your hard work stolen, and then watching it earn a better piece of SERP real estate. So what is Google doing about scraper sites, and what can we do to protect ourselves and our websites?

What Are Scraper Sites?

If you’ve been lucky enough to avoid scraper sites thus far, a short primer will put you in the picture. A scraper site is a website that lifts original content from another site and passes it off as it’s own. We’re not talking about paraphrasing or quoting here. Scrapers steal whole blocks of original content, copying and pasting it onto their own websites. They do this to generate revenue, to divert traffic from the original site, and to manipulate page rankings. Up until recently, there was little a webmaster could do about scraper sites that had stolen their content. That may be starting to change. Though I stress the phrase ‘may be’.

Google, Matt Cutts, and Scraper Sites

Over the last few years, Google has attempted to address the scraping problem through their various search algorithms. Unfortunately, that has had little effect on either scraper sites or their ability to gain ground in a given SERP. Just a few months back in February, Google’s Matt Cutts sent out a tweet announcing a new feature that allows webmasters to report scraper sites (shown above).

It’s not too much to say that Cutts’ tweet was met with consternation, some derision, and a fair bit of ill humor. It even prompted a response tweet that has since gone viral, but the less said about that the better. Whether you are a fan of Cutts and Google or not, it is clear that they have scraper sites in their cross-hairs and are working on a plan to address the situation.

Reporting a Scraper Site

Reporting a scraper site is easy, and only takes a few moments. Simply go to Google’s Scraper Report page, enter the URL of original site and the URL of the scraper site. The report page also asks that you enter the search results URL that demonstrates the problem. Finally, before you can submit your report, you must confirm that the original website is following Google’s webmaster guidelines, and that it has not received any recent manual penalties.

Google has yet to say what they are going to do with this information, and submitting a report is no guarantee that your site will improve in the page rankings or that the scraper site will be immediately penalized. But clearly, Google is gathering this information to help them form a suitable attack strategy to combat scraper sites.

Protecting Your Website’s Rankings

If your website has fallen pray to a scraper, there is not much you can do at the moment about the thievery itself. If your site is still outranking the scraper site, then you are in good shape as far as the SERPs go. However, if your site is being outranked by a scraper site, there are a few things you can do to help your site’s performance.

  • Check your website for technical errors that may be preventing the search engine bots from accurately crawling and evaluating your content. This includes checking your XML sitemap and robots.txt files and any other crawl errors using GWT or similar.
  • Optimize your title, description, and content so that search engines can better establish its topical relevance.
  • Build better back links and deep links to indicate the popularity and relevance of your content. Yes, you absolutely must make outreach a large part of you SEO efforts to garner high value links.
  • Remove any bad links that may be effecting your page ranking. Well, remove the ones you can, document the rest for possible disavow, and focus more effort on outreach over removal.

Despite the critical backlash against Matt Cutts’ February tweet, it is clear that Google is taking a fresher look at scraper sites. While reporting a scraper site may not result in any immediate action, the fact that Google is gathering this information is at least a step in the right direction. If you are being outranked by your own content, file a report and then take to your own website and look for ways to improve your optimization strategies. At some point, our content efforts and following all the rules has to get rewarded, right?

 

Featured image screenshot of tweet published Feb 2014

 SEO 101: Are You Being Outranked by Your Own Content?

Dario Zadro

Web Strategist at Zadro Web
Dario Zadro, is an SEO enthusiast, entrepreneur, and passionate web developer. To reach him, feel free to DM @dariozadro or visit Chicago SEO - Zadro Web for more insights and great tips.
 SEO 101: Are You Being Outranked by Your Own Content?

You Might Also Like

Leave a Reply

Your email address will not be published. Required fields are marked *

You may use these HTML tags and attributes: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <cite> <code> <del datetime=""> <em> <i> <q cite=""> <strike> <strong>

9 thoughts on “SEO 101: Are You Being Outranked by Your Own Content?

  1. Finding out such content scrapers of your site is really a tough job. But thanks for your nice post and tips I will act upon these tips mentioned in your post to trace content scrapers of my site.

    1. One easy way is to copy-paste your headline in Google and see results. If this won’t work, try by copying long sentences of your blog post instead and you should get some results.

      In my experience, when found some website scraping my contents, I usually do a whois research and directly contact the domain register and submit a “Duplicated content without any permission” issue.
      100% of the times I’ve been able to win against content scrapers :)

      1. Great advice! I’ve done this myself and it’s amazing how accurate google can be about finding plagiarism. I’ve seen cases where companies say things like “we write our own content to give you the best results” in their sell copy or a blog post but just by searching a sentence or two, a lot of their content was clearly lifted from other websites.

        Also, it’s a last resort if the webmaster at the offending site isn’t cooperating, but most hosts take DCMA complaints pretty seriously.

    2. Thanks for your comment and nice words. Looks like Matteo has a great response to help in finding the scraper sites by copying large blocks of text directly into Google. Another option is to use Copyscape and parse your entire site in one batch.

  2. I have used Copy Scape a few times. Have to admit I hate when this happens. I have written articles in the past and they have been copied and published within hours. The big push is for quality fresh content to help with ranking and search profile but its a pain when somebody else helps themselves to your work… Any reader feedback on this ‘service’ yet??

    Cheers

  3. You can also find scraper sites sometimes by looking through links in your Google Webmaster Tools. Some scraper sites scrape everything (including your internal links), so as a result, they link to you. Just look at sites that have lots of links pointing to your site as a filter to see which are probable ones.

    1. That’s a great suggestion Koby. I hadn’t thought to look specifically for scraper content in my GWT. I’ll have to try that. We use CopyScape here at SEJ (and I use it for my own client work as well) and I find it to be pretty reliable and affordable. But I can’t help being paranoid that they (or me!) am missing something! :)