Duplicate content is an SEO issue many SEOs or content marketers probably have experienced once a time in their daily routine. Content marketers who spend the time to create qualitative content strategies do not want to get penalized for duplication or near duplicates. In 2013, Matt Cutts stated that 25% of the web was duplicate content, you can find this number in the video at the end of the article.
Duplication refers to blocks of content that appear more than one time inside or outside a website or which are pretty similar. Duplicate content or near duplicates can lead to SEO penalties.
In fact, crawlers have trouble indexing the right content between different versions. Bots are then obliged to pick the content likely to be the best one, but it can lead to a loss of relevancy as it is not always simple to choose the right version. Moreover, bots will face difficulties to deliver the link metrics to the right page or share it between the different versions. Then, duplicate content also leads to the inability to rank the right version for a given query. At the end, you will face a drop in traffic.
The thing is, Google gives a lot of credits to user experience and focuses on delivering the best content possible to its audience. This is also why duplicate content is penalized.
While not all types of duplicate content can hurt your SEO, some of them need to be on the look out to avoid SEO penalties.
This article will cover:
There are different types of duplicate content you should avoid.
Duplicate product forms
E-commerce websites often use manufacturer’s item descriptions to describe the products they sell. The problem is that those products are often sold to different e-commerce websites. Then, the same description appears on different websites and creates duplicate content.
Syndicated or copied content
Many bloggers use content, quotes or comments from other websites to illustrate their articles. There is nothing wrong with that if you link back to the original one. However, Google can still consider this as a duplication and will poorly value those pieces of content.
Sorting and multi-pages lists
Large e-commerce websites have filter and category options that generate unique URLs. Product pages can appear in different categories and be ordered differently depending on how the list is sorted. For instance, if you range 45 products by price or by alphabetical order, you will end up with two pages containing the same content, but with different URLs.
Google considers URLs in www, http, https, .com and .com/index.html as different ones even if they point to identical pages and will evaluate them as duplicate content.
Session IDs issues refer to different session IDs stored in the same URL that are assigned to a visitor when he or she comes to the website.
Printable versions of content can lead to duplicate content issues when different versions of a page are indexed.
Take a fresh look at what Relevance can do to elevate your brand.Learn More
There are different best practices to avoid duplicate content issues. The main solution with content located in different URLs is to canonicalize the original one. You can use a 301 redirect, a rel=canonical or parameters handling tool from Google Webmaster Central.
The 301 redirect is great for URL issues leading to duplications. It informs search engines which version of a page is the original and it links the duplications to that original one. Plus, when different well-ranked pages are linked to a single one, they are not in competition anymore and they create an overall stronger and more popular signal.
It works quite the same as 301 redirects, but it is easier to set up. That tag is located in the HTML head section of your web page and looks like this:
<link href=”http://www.mywebsite.com/canonical-version-of-page/” rel=”canonical” />
Then search engines know the above URL is a copy of the original one.
You can use it for content you integrated from other websites. It will inform search engines that you know the content is not from you and that the link metrics of that content should pass to the original one.
Use the noindex,nofollow meta tag to tell search engines not to index the content. Bots will be able to crawl the page, but won’t index it. Thus, you won’t be penalized for duplicate content.
A quite simple operation is to set a preferred domain for search engines. It will inform whether a site must be displayed under ‘www’ or not in the SERPs.
Unique product description
As we said, product information on e-commerce websites can lead to duplicate content issues. Take the time to write unique ones or enrich your descriptions as it will help you rank above sites whose descriptions are duplicated.
In order to save time, you can use different qualitative tools to help you eradicate duplicate content. Here are three different ones, with some being totally free.
This tool detects any duplicate content on your website. You just need to add your website’s URL and it will draw a full report with your content performances.
This onsite SEO semantic crawler also offers a duplicate and near duplicate detection feature. It shows you clusters of duplicates and near duplicates, types of duplication and clearly indicates which URLs are concerned.
The tool offers a 30 day-trial. If you want to enjoy all the functionalities, you just need to pick a plan and cancel your subscription before the end of the trial and you won’t pay anything.
Copyscape is a great partner because it detects duplicate content outside of your blog. You can thus easily know if someone has duplicated your content without your permission or without giving you credits.
Our publication contributors combine decades of experience with unique insights into the content promotion and distribution industry.
Welcome to a community dedicated to solving the challenge of getting online content seen and consumed by the right audience on the best channels.
Get the latest content promotion news and insights everyday.
Thank you for subscribing please check your email for verification.