{"id":957,"date":"2010-08-02T10:18:07","date_gmt":"2010-08-02T07:18:07","guid":{"rendered":"http:\/\/www.webseoanalytics.com\/blog\/?p=957"},"modified":"2010-08-02T10:18:07","modified_gmt":"2010-08-02T07:18:07","slug":"duplicate-content-the-effects-on-search-engine-rankings","status":"publish","type":"post","link":"http:\/\/www.webseoanalytics.com\/blog\/duplicate-content-the-effects-on-search-engine-rankings\/","title":{"rendered":"Duplicate Content: the effects on Search Engine Rankings"},"content":{"rendered":"<p><img decoding=\"async\" style=\"margin:0 0 10px 40px; display:block;\" title=\"duplicatecontent\" src=\"http:\/\/www.webseoanalytics.com\/blog\/wp-content\/uploads\/2010\/07\/duplicatecontent.gif\" alt=\"duplicatecontent\" width=\"570\" \/>The duplicate content is one of the <strong>major problems<\/strong> that most of the websites face. Search engines are also affected since they waste valuable resources in order to <strong>crawl<\/strong>, <strong>analyze<\/strong> and <strong>detect<\/strong> those pages. Finally since the duplicate content <strong>reduces<\/strong> the <strong>quality<\/strong> of the search results, this problem also affects the search engine users. So how can we solve this problem once and for all?<\/p>\n<p>In this article we will focus on the root of the problem, we\u2019ll see why it can be a major issue for the search engines and for the websites and we will explain how duplicate content affects SEO. In the <a href=\"http:\/\/www.webseoanalytics.com\/blog\/duplicate-content-how-to-solve-the-problem\/\">next article<\/a> we\u2019ll examine in detail the most <strong>common<\/strong> <strong>web development<\/strong> and <strong>SEO mistakes<\/strong> that lead to duplicate content problems and we\u2019ll suggest ways to <strong>solve the issue<\/strong>.<\/p>\n<h2>What is duplicate content?<\/h2>\n<p>The term Duplicate Content is used to describe the situation where multiple URLs have the <strong>same<\/strong> or <strong>almost the same<\/strong> content. Note that those pages can be part of the same or of different websites.<\/p>\n<h2>How duplicate content is created?<\/h2>\n<p>One of the common ways to get this result is by copy-pasting the same text in different pages or by submitting the same content\/article\/review in multiple sites. Additionally duplicate content can be the result of using <strong>poor Web development techniques<\/strong> or of <strong>developing a bad link structure<\/strong>.<\/p>\n<h2>Why duplicate content is a problem?<\/h2>\n<p>In order to understand why the Duplicate Content is a problem you need to see it from the search engine\u2019s point of view. Search engines need to <strong>crawl<\/strong>, <strong>analyze<\/strong> &amp; <strong>index<\/strong>, find the <strong>reputation<\/strong> of each page and be able to <strong>search<\/strong> <strong>fast<\/strong> through their index in order to return the results to the users. Having lots of duplicate content in a website is bad for search engines since they waste their resources on pages that do not usually have a significant value for the users.<\/p>\n<p>Matt Cutts, a well known Google employee, has mentioned in one Google-Webmaster-Help video that in order to crawl a large part of the web you need a relatively small number of machines (more than 25 less than 1000). This means that <strong>crawling<\/strong> a website requires a relatively <strong>small amount of resources<\/strong>. Nevertheless the <strong>analysis<\/strong> of the page, the evaluation of the links and the indexation is a much more <strong>time consuming process<\/strong>. Those of you, who have coded web-spiders in the past, know that the analysis requires lots of CPU and memory comparing to the web requests. This is due to the complexity of the algorithms that are used in the text analysis.<\/p>\n<div style=\"text-align:center;margin:10px 0 10px 0;\"><object classid=\"clsid:d27cdb6e-ae6d-11cf-96b8-444553540000\" width=\"640\" height=\"385\" codebase=\"http:\/\/download.macromedia.com\/pub\/shockwave\/cabs\/flash\/swflash.cab#version=6,0,40,0\"><param name=\"allowFullScreen\" value=\"true\" \/><param name=\"allowscriptaccess\" value=\"always\" \/><param name=\"src\" value=\"http:\/\/www.youtube.com\/v\/usyYXNNBRjc&amp;hl=el_GR&amp;fs=1\" \/><param name=\"allowfullscreen\" value=\"true\" \/><embed type=\"application\/x-shockwave-flash\" width=\"640\" height=\"385\" src=\"http:\/\/www.youtube.com\/v\/usyYXNNBRjc&amp;hl=el_GR&amp;fs=1\" allowscriptaccess=\"always\" allowfullscreen=\"true\"><\/embed><\/object><\/div>\n<p>Clearly the duplicate content is a problem for the search engine users because it affects the <strong>quality of the search results<\/strong>. But why this is a problem for the webmasters? Well, since this problem requires additional resources that <strong>cost money to the search engine companies<\/strong>, they try to force the webmasters and the SEOs to help them solve the issue. And the cheapest way to solve it is by motivating the webmasters to eliminate their duplicate pages.<\/p>\n<h2>Does duplicate content issue affects the rankings?<\/h2>\n<p>Even if duplicate content <strong>will not directly lead to bans<\/strong> from search engines (read the article \u201c<a href=\"..\/common-on-page-optimization-mistakes\/\">Why my SEO campaign failed? Part 1: Common On-page Optimization mistakes<\/a>\u201d),<strong> it does affect the SEO status <\/strong>of a website. When search engines identify cases of duplicate content they try to find out which is the best version of the page that should appear in the search results. Normally this choice is made based on the <strong>age<\/strong> of the page, the <strong>authority<\/strong> of the domain, the <strong>number of incoming links<\/strong>, the <strong>PageRank<\/strong> etc. So if few pages of your site contain lots of copy-pasted text from another pages or websites there is a good chance that they will not appear in the search results.<\/p>\n<div style=\"margin:10px 0 10px 0; text-align:center;\"><img decoding=\"async\" loading=\"lazy\" class=\"alignnone size-full wp-image-971\" title=\"duplicate-content-key\" src=\"http:\/\/www.webseoanalytics.com\/blog\/wp-content\/uploads\/2010\/07\/duplicate-content-key.jpg\" alt=\"duplicate-content-key\" width=\"500\" height=\"310\" srcset=\"http:\/\/www.webseoanalytics.com\/blog\/wp-content\/uploads\/2010\/07\/duplicate-content-key.jpg 500w, http:\/\/www.webseoanalytics.com\/blog\/wp-content\/uploads\/2010\/07\/duplicate-content-key-300x186.jpg 300w, http:\/\/www.webseoanalytics.com\/blog\/wp-content\/uploads\/2010\/07\/duplicate-content-key-483x300.jpg 483w\" sizes=\"(max-width: 500px) 100vw, 500px\" \/><\/div>\n<p>Additionally as we said before, duplicate pages can be a result of <a href=\"..\/common-web-development-mistakes\/\">poor programming<\/a> or link structure development. Usually the dynamic websites tend to pass variables in each dynamic URL in order to get a particular record from the Database:<\/p>\n<div style=\"margin:10px 0 10px 0; color:#67b7da;\">http:\/\/www.example.com\/product.php?id=3012&amp;color=red<\/div>\n<p>In the above example the product with id 3012 could have lots of different pages (one for every color). If the content of every page does not change significantly then this could lead to duplicate content problems.<\/p>\n<p>The <strong>PageRank distribution<\/strong> is negatively affected by the presence of duplicate pages. Since PageRank flows through the links, a lot of important link juice is directed to duplicate pages or gets evaporated. As a result the Rankings of the website are affected.<\/p>\n<p>Finally another reason why duplicate content can negatively affect your rankings is that search engines find and index a <strong>particular number of pages<\/strong> from every website. The number of pages depends on the <strong>domain authority<\/strong>. If your site contains lots of duplicate pages, then the re-crawl period will increase and the new important pages that you add in your website will be indexed much slower.<\/p>\n<h2>Should duplicate content be a problem for all websites?<\/h2>\n<p>Certainly having the same content in many different pages of the website is not particularly useful for the users, but this does not mean that it\u2019s a critical mistake. As we mentioned above it becomes a <strong>major problem<\/strong> for webmasters because it affects their Search Engine rankings.<\/p>\n<p>Generally speaking, <strong>in most of the cases<\/strong>, unique quality content is very important for the users. But the question is, should duplicate content affect <strong>all<\/strong> webmasters or there are cases where they <strong>should not affect<\/strong> their SEO status? Many users have asked Matt Cutts in the past whether it is a problem for online stores to have the same generic product descriptions that are used also by other websites. Matt Cutts replied that this is a problem for their SEO campaigns and that <em>if they want to attract more users they should do it by providing unique quality content and by differentiating from other e-commerce sites<\/em>. I am sorry Matt but I have to <strong>strongly disagree<\/strong> with you on that.<\/p>\n<div style=\"text-align:center;margin:10px 0 10px 0;\"><object classid=\"clsid:d27cdb6e-ae6d-11cf-96b8-444553540000\" width=\"640\" height=\"385\" codebase=\"http:\/\/download.macromedia.com\/pub\/shockwave\/cabs\/flash\/swflash.cab#version=6,0,40,0\"><param name=\"allowFullScreen\" value=\"true\" \/><param name=\"allowscriptaccess\" value=\"always\" \/><param name=\"src\" value=\"http:\/\/www.youtube.com\/v\/z07IfCtYbLw&amp;hl=el_GR&amp;fs=1\" \/><param name=\"allowfullscreen\" value=\"true\" \/><embed type=\"application\/x-shockwave-flash\" width=\"640\" height=\"385\" src=\"http:\/\/www.youtube.com\/v\/z07IfCtYbLw&amp;hl=el_GR&amp;fs=1\" allowscriptaccess=\"always\" allowfullscreen=\"true\"><\/embed><\/object><\/div>\n<p>My personal opinion is that when a user searches for a particular branded product, for example the new Xbox 360 console, he\/she does not really care about the well-written description, the number of incoming links of the domain, the PageRank of the page, or the authority etc. He\/she does care about the <strong>price<\/strong>, the <strong>product<\/strong> and the <strong>services<\/strong> that come with it. Using the same algorithms or principles to evaluate blogs and ecommerce websites is not the best practice. Even if I do understand that there are lots of <strong>technical difficulties<\/strong> in evaluating correctly the products, still you can\u2019t force or suggest to the e-tailers that they have to <strong>spend lots of time<\/strong> and effort in rewriting their product descriptions in order to avoid a duplicate content disaster. <strong>E-tailers don\u2019t differentiate from the competition by providing unique descriptions but by providing unique quality products and services.<\/strong><\/p>\n<p>That was the first part of the article. The next part will focus on more <strong>technical subjects<\/strong> and on <a href=\"http:\/\/www.webseoanalytics.com\/blog\/duplicate-content-how-to-solve-the-problem\/\">how to solve the duplicate content problem<\/a>.<\/p>\n<p>Images by <a rel=\"nofollow\" href=\"http:\/\/www.searchenginejournal.com\/duplicate-content-infographics\/10292\/\" target=\"_blank\">searchenginejournal<\/a>, <a rel=\"nofollow\" href=\"http:\/\/www.seodenver.com\/duplicate-content-over-multiple-domains-seo-issues\/\" target=\"_blank\">seodenver<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>The duplicate content is one of the major problems that most of the websites face. Search engines are also affected since they waste valuable resources in order to crawl, analyze and detect those pages. Finally since the duplicate content reduces &hellip; <a href=\"http:\/\/www.webseoanalytics.com\/blog\/duplicate-content-the-effects-on-search-engine-rankings\/\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":2,"featured_media":962,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[10],"tags":[],"aioseo_notices":[],"_links":{"self":[{"href":"http:\/\/www.webseoanalytics.com\/blog\/wp-json\/wp\/v2\/posts\/957"}],"collection":[{"href":"http:\/\/www.webseoanalytics.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/www.webseoanalytics.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/www.webseoanalytics.com\/blog\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"http:\/\/www.webseoanalytics.com\/blog\/wp-json\/wp\/v2\/comments?post=957"}],"version-history":[{"count":0,"href":"http:\/\/www.webseoanalytics.com\/blog\/wp-json\/wp\/v2\/posts\/957\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"http:\/\/www.webseoanalytics.com\/blog\/wp-json\/wp\/v2\/media\/962"}],"wp:attachment":[{"href":"http:\/\/www.webseoanalytics.com\/blog\/wp-json\/wp\/v2\/media?parent=957"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/www.webseoanalytics.com\/blog\/wp-json\/wp\/v2\/categories?post=957"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/www.webseoanalytics.com\/blog\/wp-json\/wp\/v2\/tags?post=957"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}