Duplicate content has always been a hot topic among webmasters; mostly because no one really knows what it is and the rumors persist.
And Google doesn’t help much either. Sometimes, I think of it as a hyperactive 3 year-old, who is incredibly sharp in some areas, but not so much in others.
So the best way to go is to keep it simple, stay under the radar, and shoot for the middle of the road.
With that said, let’s figure out what duplicate content is, what it isn’t, and what you should do to stay on top of it.
What is duplicate content?
“Duplicate content generally refers to substantive blocks of content within or across domains that either completely match other content or are appreciably similar. Mostly, this is not deceptive in origin. Examples of non-malicious duplicate content could include:
- Discussion forums that can generate both regular and stripped-down pages targeted at mobile devices
- Store items shown or linked via multiple distinct URLs
- Printer-only versions of web pages”
Identical or substantially similar content.
Within your own domain or across others.
Most of it is normal and acceptable.
So far, so good.
Why Is Duplicate Content a Problem?
How would you like to search for the best pecan pie recipe just to find that every single result on the first page turned out the exact same recipe?
Users don’t like the same result and Google doesn’t like crawling the same results.
For a search engine, it’s also a processing consideration. If there is substantial duplication, the crawl/indexation rates might be dampened. In short, the site can lose some ‘trust’.
Two Types of Duplicate Content
We all have our own ideas of duplicate content and most of the times they boil down to “Don’t republish the same article to multiple directories. Instead, spend countless hours spinning that same article to the point where it doesn’t make sense any longer and THEN publish it to a zillion and one directories. That will surely trick all the PhDs working for Google into ranking my site pretty highly.”
Now in the spirit of “being informed”, let’s take a look at the 2 types of duplicate content you see around, shall we?
- Cross-domain type: this one is the most commonly thought of and includes the same content, which (often unintentionally) appears on several external sites.
- Within-your-domain type: the one that Google is actually mostly concerned about, i.e. that appears (often unintentionally) in several different places within your site.
Let’s now do a little more exploring into each type and see what Google really thinks about it.
Off-Site Content Syndication
There is absolutely nothing wrong with syndicating your content to different sites per se.
NOTHING WRONG WITH IT!
Here’s what happens when your content gets syndicated: Google will simply go through all the available versions and show the one that they find the most appropriate for a specific search.
Mind you the most appropriate version might not be the one you’d prefer to have ranked. That’s why it’s very important that each piece of syndicated content includes a link back to your original post – I assume it would be on your site. That way Google will trace the original version and will most likely (but not always) display it in its search results.
Per Matt Cutts:
I would be mindful that taking all your articles and submitting them for syndication all over the place can make it more difficult to determine how much the site wrote its own content vs. just used syndicated content. My advice would be 1) to avoid over-syndicating the articles that you write, and 2) if you do syndicate content, make sure that you include a link to the original content. That will help ensure that the original content has more PageRank, which will aid in picking the best documents in our index. (Source)
Black Hat Syndication
However, here’s the other side of content syndication coin: the content is deliberately duplicated across the web in an attempt to manipulate search engine rankings or to generate more traffic.
This results in repeated content showing up in SERPs, upsets the searchers, and forces Google to clean out the house.
“In the rare cases in which Google perceives that duplicate content may be shown with intent to manipulate our rankings and deceive our users, we’ll also make appropriate adjustments in the indexing and ranking of the sites involved. As a result, the ranking of the site may suffer, or the site might be removed entirely from the Google index, in which case it will no longer appear in search results.” (Resource – Google Webmaster Tools Help)
On-Site Content Syndication
On-site duplicate content problems are much more common and guess what: they are entirely UNDER YOUR CONTROL, which makes it very easy to fix them.
The first step to identifying the potential weak spots on your blog is learning more about your content management system.
For example, a blog post can show up on the home page of your blog, as well as category page, tag page, archives, etc. – THAT’S the true definition of duplicate content.
We, the users, have the common sense to understand that it’s still the same post; we just get to it via different URLs. However, search engines as unique pages with exactly same content = duplicate content.
How to Take Matters into Your Own Hands
Here are some practical “non-techie” steps you can take to minimize the presence to dupe content on your site:
- Take care of your canonicalization issues. In other words, www.trafficgenerationcafe.com, trafficgenerationcafe.com, trafficgenerationcafe.com/index.html are one and the same site as far as we are concerned, but 3 different sites as far as search engines are concerned. You need to pick your fave and stick with it. If you don’t know how, here’s are the instructions: WWW vs non-WWW: Why You Should Put All Your Links in One Basket
- Be consistent in your internal link building: don’t link to /page/ and /page and /page/index.htm – if links to your pages are split among the various versions, it can cause lower per-page PageRank.
- Include your preferred URLs in your sitemap
- Use 301 redirects: If you have restructured your site (for instance, changed your permalink structure to a more SEO-friendly one), use 301 redirects (“RedirectPermanent”) in your .htaccess file or, even simpler, use one of the many Redirection plugins available in your WordPress plugin directory.
- Use rel=”canonical”
- Use Google Webmaster Parameter Handling Tool
- Minimize repetition: i.e. don’t post your affiliate disclaimer on every single page; rather create a separate page for it and linked to it wheb needed.
- Managing your archive pages: Avoid duplicate content issues by displaying excerpts on your archive pages instead of full posts. You really want to give your readers just a hint of the content and direct them back to the original posts. To accomplish that, open your archive.php of your theme and replace the_content with the_excerpt. Hint: make sure your category and tag pages also display excerpts only.
- Country-specific content: Google is more likely to know that .de indicates Germany-focused content, for instance, than /de or de.example.com.
This is a good time to remind myself that I am not writing a novel…
Oh, wait a minute: one more important issue: robot.txt file.
According to TopRankBlog.com:
Google doesn’t recommend blocking duplicate URLs with robots.txt, because if they can’t crawl a URL they have to assume it’s unique. It’s better to let everything get crawled and to clearly indicate which URLs are duplicates…. Robots.txt controls crawling, not indexing. Google may index something (because of a link to it from an external site) but not crawl it. That can create a duplicate content issue.
Let’s move on, shall we?
Is There a Duplicate Content Penalty?
I’ll have Google answer this daunting question.
Here’s a quote by Susan Moskwa, Webmaster Trends Analyst from Google:
A lot of people think that if they have duplicate content that they’ll be penalized. In most cases, Google does not penalize sites for accidental duplication. Many, many, many sites have duplicate content.
Google may penalize sites for deliberate or manipulative duplication. For example: auto generated content, link networks or similar tactics designed to be manipulative.”
Susan further explained when webmasters should not worry about duplicate content:
- Common, minimal duplication.
- When you think the benefit outweighs potential ranking concerns. Consider your cost of fixing the duplicate content situation vs. the benefit you would receive.
- Remember: duplication is common and search engines can handle it.
How exactly does Google handle it?
While pulling up the search results, Google will basically collapse the duplicates leaving only the most relevant, in their opinion of course, page in the SERPs for that specific query. As I explained before, the way Google determines the most relevant result is based upon a myriad of factors and the only thing you can do for your part is to always link back to your original post.
Scraping Be Gone!
A word on the recent Google algorithm change:
My post mentioned that “we’re evaluating multiple changes that should help drive spam levels even lower, including one change that primarily affects sites that copy others’ content and sites with low levels of original content.” That change was approved at our weekly quality launch meeting last Thursday and launched earlier this week. (Matt Cutts – source)
What does it mean to the average webmaster?
We can all do a little chicken dance, since the probability of scraped (stolen, in other words) content ranking above the original articles that we put blood, sweat, and tears into, is minimal.
Google is rightfully going to war against all the autoblogs that don’t have what it takes to produce content of their own and all they do is republished other people’s work in hopes to rank highly in search engines, bring traffic to their crappy websites and make some money off AdSense, paid advertisement, and such.
If you find that another site is duplicating your content by scraping (misappropriating and republishing) it, it’s unlikely that this will negatively impact your site’s ranking in Google search results pages. If you do spot a case that’s particularly frustrating, you are welcome to file a DMCA request to claim ownership of the content and request removal of the other site from Google’s index.
Duplicate Content on Your Blog
Originally this duplicate content post was written by me as a guest post for Search Engine Journal.
This topic is so important though that I decided to bring it to Traffic Generation Cafe and make sure that you have your questions on the subject answered once and for all… well, until Google changes something again, that is.
Of course, I have some special additions to the post just for my readers – you, that is.
Here are some examples of duplicate content on your blog that you might not be aware of:
- Comment Pagination: believe it or not, when you break down your blog comments into pages, you create duplicate content.
- Tag/Category pages: the issue here is the same as with archives – if you display full posts on those pages, you create duplicate content. So make sure you switch to showing excerpts only and that will solve the issue.
- Author pages: when you are a blog of one or many authors, each author has an archive page, which collects all the posts he/she has written. You got it: make sure to display excerpts.
If I think about it, I am sure, I’ll find many other instances of dupe content on your site.
However, remember what I said above: most of these instances are considered natural and Google (most likely) will deal with them without a glitch.
Does it mean you should forget about duplicate content and viva la vida loca (live crazy life, in other words)?
In short, Google will give your site only that much “crawling” attention and you should spend those precious seconds wisely – not making Google figure out what content it actually should pay attention to.
If you want a longer version, here you go: Why you should still care about duplicate content from SearchNewsCentral.com. You really should read it.
Duplicate Content Marketing Takeaway
1. Dupe content doesn’t cause your site to be penalized.
2. Google is very good at picking the best version of your content to be displayed in SERPs and ignoring the rest.
3. Almost all dupe content issues are easy to fix and should be fixed.
4. Don’t worry, be happy – don’t be afraid, be informed.
Image credit: http://www.laughparty.com/
So, what think you? Comment to show me that you’re alive!