Owning a Tourism Business is awesome

7 Jun 2007

The Most Common Reason for Dropped Rankings: Duplication

Repeatedly I find myself explaining that using duplicate content can and will negatively affect search engine rankings and it is heart breaking to see clients having to rebuild rankings due to such a simple mistake. As a result, I felt it was time to spread the word and this post to hopefully dispel many misled website owners.

Why write an entire post on something as simple as duplicate content? Well probably because it is not as simple as it sounds and many resort and tourism website owners find themselves in the gray area of duplication; where they don't know for sure whether they are risking rankings or not.

The following is a breakdown of the most common duplicate content issues we see defined from the standpoint of a question – hopefully making this article a little easier to read. After all, I have no illusions that reading up on duplicate content rules are exciting.

Duplicate Websites

Definition: a duplicate website is a website that has many if not all of the same pages as another live website. Note: the following questions are based on a resort who owns two websites that are duplicates.

Q: "Why is a duplicate website such a bad idea?"
A: The major SE's are constantly trying to improve the quality of their search engine results in an effort to provide the best quality content for users. When search engine spiders index duplicate content, valuable time and processing power is wasted. As a result, search engines have blocked sites that used duplicate content from their database, ultimately favoring the site that either had the content first, or the one site that has the greater online history. In addition, the major search engines have a bad taste after dealing with so many duplicates content created by spammers over the past several years. As a result, posting a duplicate website is an offense that can quite literally blacklist a domain; there are few things the search engine properties dislike more than being gamed by spammers.

Q: "What should I do with my duplicate website then? Just delete it?"
A:
Deleting the site is the only option unless you want to create an entirely new website with unique content and a unique purpose. That said, by deleting the website you can still ensure the effort you put into promoting the old site does not go to waste by pointing the domain to your new website's domain using a 301 redirect. A 301 is a term used to describe a server protocol which Google and other search engines will 'see' when they visit the old site. The protocol essentially says that your content from the old site can be found on the new site and that this is a permanent forwarding of all traffic. 301 redirects are by far the best way to minimize your losses from shutting down a website that just might have traffic or inbound links.

Q: "Which website should I shut down? Is there anything I should consider first?"
A:
Yes, it is very important that you choose the website that has the most back links and has been online the longest. The reason I say this is that Google tends to favor entrenched websites; they have been around a while, are well back linked and overall appear to have a positive history.

Whatever your decision is, it is vital you understand switching a website to a new domain is a dangerous step. This is because of Google's famed 'sandbox'. The 'sandbox' is really only an overused turn of phrase that represents a portion of the Google algorithm, which considers the age of the domain as a signifier of trust. Generally, new websites will require 6 months to a year before substantial rankings are evident; this is kind of a right of passage that Google appears to be enforcing on the average website. Sites that are obviously popular and quickly gain a load of legitimate link popularity will easily avoid the sandbox (because Google can not afford to miss a 'great' website) but this is not the common scenario.

Q: "Will using a 301 redirect pass on the benefit of the deleted site's link popularity?"
A:
Link popularity is passed onto the other website when a 301 is used but how much this pass-over will benefit the website seems to fluctuate on a case-by-case basis. Usually the fluctuation is only present when popularity from one domain is passed to another with differing content/topic. In this case, since the link popularity is being redirected to an identical website I expect the benefit to be virtually lossless.

Duplicate Content

Definition: content appearing within a website that is duplicated elsewhere on the same website or elsewhere on the Internet.

Q: "I need content for my website; can I just copy content from industry journals and benefit from that quality content?"
A:
No, aside from the copyright concerns of using content that is not yours, your rankings (if they exist) would suffer because it is highly likely the major search engines would detect the duplicate content. As a result, the page that you create may get flagged as duplicated and it would be ignored at the very least. The page could even devalue your site's overall credibility. Credibility is a critical component of Google's algorithm so sites with less credibility tend to have a harder time staying ('sticking' if you will) in a particular ranking.

Q: "I use a content management system to manage my site and it uses a particular set of templates. These templates have some duplicate content within them and they are spread all throughout my website. Should I be worried?"
A:
No, in most cases the amount of duplicate content used within a template in a content management system (CMS) is negligible. If, however, you have a large number of pages created using a page where 90% of the text is duplicated and only 10% is unique you do have a reason to make some changes. In my opinion it is crucial that every page within a website be composed mostly of unique content with the exception of catalogues and shopping carts where text simply has to be reused over and over.

Whatever your situation make certain that your site contains a large number of pages composed of unique content that has been well optimized by yourself or your search engine optimizer (SEO).

Q: "How much of my page should be unique? Is there a standard ratio or percentage you can share?"
A:
There is no industry standard formula but, if I had to state a percentage, I would say a minimum of 70% of the page should be completely unique to thwart any concerns of duplication. You may be able to get away with less than 70% unique content, but I would suggest this is playing with fire. Either way, this statistic is moot since every page you create needs to be created with the intention to provide a powerful resource; after all search engines are only a small part of the plan - you do need visitors to like what they see and büy your product or service!

Q: "My blog currently has many different ways to find content and depending on the route a visitor may find the page is actually shown on a different URL (i.e. archives, search by label, etc.). In this scenario am I not in danger of a duplicate content penalty?"
A:
Yes and no. Yes that this is duplicated content but no you are not likely to be penalized by this simply because a majority of blogs offer these additional methods of finding content so it would be detrimental if search engines penalized this application right now. That said, search engines do have to have some way to handle this duplicate content. I expect when Google (picking the most advanced search engine) finds duplicate blog postings on a website its algorithm chooses the most popular posting as the primary page to provide in its ranking results. In other words, the posting URL that has the most number of inbound links or was spidered first will be the page that attains rankings.

Conclusion

I am sure I didn’t cover every question regarding duplicate content but I am fairly certain I touched on the most common questions we see in basic SEO work.

No comments: