7 deadly mistakes to avoid in Google Webmaster Tools Console

google_webmastertoolsThe Google Webmaster Tools is a free service that allows webmasters to submit their websites & sitemaps to Google, affect their indexing process, configure how the search engine bots crawl their sites, get a list of all the errors that were detected on their pages and get valuable information, reports and statistics about their websites. Certainly it is a very powerful console that can help Webmasters not only optimize their websites but also control how Google accesses their sites.

Unfortunately though, even if using the basic functionalities and changing the default configuration is relatively straight forward, one should be very careful while using the advanced features. This is because some of the options can heavily affect the SEO campaign of the website and change the way that Google evaluates it. In this article we discuss the 7 most important mistakes that one can make while using the advanced features of Google Webmaster Tools and we explain how to avoid a disaster by configuring everything properly.

1. Geographic Targeting feature

The Geographic Targeting is a great feature that allows you to link your website to a particular location/country. It should be used only when a website targets to users from a particular country and when it does not really make sense to attract visitors from other countries.

geotargeting

This feature is enabled if your domain has a generic (neutral) top-level-domain such as .com, .net, .org etc while it is preconfigured if it has a country specific TLD. By pointing your site to a particular country you can achieve a significant boost on your rankings on the equivalent Country version of Google. So for example by setting the Geographic Targeting to France you can positively affect your rankings on Google.fr. Unfortunately this will also lower your rankings on the other Google Country versions. So this means that you might be able to boost your rankings on Google.fr but you can actually hurt your results on Google.com and Google.co.uk.

So don’t use this feature unless your website targets only on people from a specific country.

2. Sitemap Submission

Google allows you to submit via the Console a list of sitemaps and this will speed up drastically the indexing of your website. Even though this is a great feature for most of the webmasters, in few cases it can cause two major problems.

The first problem is that sometimes the XML sitemaps appear on Search Results, so if you don’t want your competitors to be able to see them make sure you gzip them and submit the .gz version on Google. Also obviously you shouldn’t name the files “sitemap.xml” or add their URLs on your robots.txt file.

sitemaps

The second problem appears when you index a large number of pages (several hundred thousands) really really fast. Even though Sitemaps will help you index those pages (given that you have enough PageRank), you are also very likely to reach a limit and raise a spam flag for your website. Don’t forget that the Panda Update targets particularly on websites with low quality content that try to index thousands of pages very fast.

So if you want to avoid exposing your site architecture be extra careful with sitemap files and if you don’t want to reach any limitation use a reasonable number of sitemaps.

3. Setting the Crawl Rate

Google gives you the ability to setup how fast or slow you want Googlebot to crawl your website. This will not only affect the crawl rate but also the number of pages that get indexed every day. If you decide that you want to change the default setting and set a custom crawl rate remember that a very high crawl rate can consume all of the bandwidth of your server and create a significant load on it while a very low crawl rate will reduce the freshness and the number of crawled pages.

crawl-rate

For most webmasters it is recommended to let Google determine their crawl rate.

4. Crawler Access

On the Crawler Access Tab, Google gives you the opportunity to see which parts of your website are allowed to be crawled by Google. Also it gives you the opportunity to Generate a robots.txt file and request a URL Removal. These features are powerful tools in the hands of experienced users, nevertheless a misconfiguration can lead to catastrophic results affecting heavily the SEO campaign of the website. Thus it is highly recommended not to block any parts of your website unless you have a very good reason for doing it and you know exactly what you are doing.

url-removal

Our suggestion is that most webmasters should not block Google from any directory and let it crawl all of their pages.

5. Parameter handling

After crawling both your Website and the external links that point to your pages, Google presents to you the most popular URL parameters that are passed to your pages. In the Parameter Handling console Google allows you to select whether a particular parameter should be “Ignored”, “Not Ignored” or “Let Google decide”.

param-handling

These GET parameters can be from product Ids to Session Ids and from Google Analytics campaign parameters to page navigation variables. These parameters in many cases can lead to duplicate content issues (for example Session IDs or Google Analytics Campaign parameters), while in other cases they are essential parts of your website (product or category IDs). A misconfiguration on this console could lead to disastrous results on your SEO campaign because you can either make Google index several duplicate pages or make it drop from the index useful pages of your website.

This is why it is highly recommended to let Google decide how to handle each parameter, unless you are absolutely sure what a particular param is all about and whether it should be removed or maintained. Last but not least, I remind you that this console is not the right way to deal with duplicate content. It is highly recommended to work with your internal link structure, use 301s and Canonical URLs before trying to ignore or block particular parameters.

6. Sitelinks

As we explained several times in the past, the sitelinks are links that appear under some results in SERPs in order to help users navigate the sites easier. The sitelinks are known to attract user attention and they are algorithmically calculated based on the link structure of the website and the anchor texts that are used. Through the Google Webmaster Tools console the webmaster is able to block the Sitelinks that are not relevant, targeted or useful to the users. Failing to remove the irrelevant sitelinks or removing the wrong sitelinks can lead to low CTR on Search Engine Results and thus affect the SEO campaign.

VodeblogWebmasterToolsSitelinks

Revisiting regularly the console and providing the appropriate feedback about each unwanted sitelink can help you get better and more targeted sitelinks on the future.

7. Change of Address

The Change of Address option is available only for domains and not for subdomains or sub directories. From this console you are able to transfer your website from a particular domain to a new one. Most of the users will never need to use this feature nevertheless if you find yourself in this situation you need to proceed with caution. Using 301 redirections and mapping the old pages to the new ones are extremely important to make the transition successful. A misconfiguration can cause a lot of problems nevertheless it is highly unlikely that someone will accidentally set up this option.

change-address

Avoid using this feature when your old domain has been banned by Google and have always in mind that changing your domain will have a huge impact on your SEO campaign.

The Google Webmaster Tools is a great service because it provides all the necessary resources to the webmasters in order to optimize, monitor and control their websites. Nevertheless caution is required while changing the standard configuration because mistakes can become very costly.

TRY THE WSA TOOLBOX
  • Pingback: 7 deadly mistakes to avoid in Google Webmaster … – Web SEO Analytics :PC & Internet Lifestyle

  • Pingback: 3 errores comunes cometidos en Google Webmasters

  • Logo Design

    That’s very useful informative that I was looking for a long time. You shared it very briefly. Now webmasters can do their best.. Good luck.

  • Pingback: 7 deadly mistakes to avoid in Google Webmaster Tools Console | Web … | Webmaster

  • Pingback: Google Plus, SEO, Online Business, Social Web – Speedlink 26

  • Pingback: Google Plus, SEO, Online Business, Social Web – Speedlink 26 – seo

  • LA seo

    Very informative tips for webmasters. I found that part with submitting sitemap is the great tool to index your pages. I didn’t realize that before. Thanks!

  • SEO Company

    Certainly it is a very powerful console that can help Webmasters not only optimize their websites but also control how Google accesses their sites.
    _______________
    James

  • Pingback: 7 deadly mistakes to avoid in Google Webmaster Tools Console | Web SEO Analytics | SEO Strategies & Tactics | Scoop.it

  • Pingback: 44 Google Webmaster Tools Resources | SEOptimise

  • Seo Agency

    That’s very useful informative that I was looking for a long time. You shared it very briefly. Now webmasters can do their best.. Good luck.
    _______
    James

  • Aleisteir Roberts

    great posts, im just recently found your blog thru Twitter and I am hoocked, it was like a bag full of treasure for a newbie SEO like me.

  • Wakefield locksmiths

    Thanks for the useful guide, although I already knew quite a few of the points I was still able to learn something from it

  • becas guarderia

    Lately google give more importance to freshness so if you update the content, as the post said better you have their crawls running faster for your site..

  • Suplementos Antioxidantes

    Hi Guys
    Great tips & we follow your recommendations frequently. However the lastest Panda actualizations are doing some weird things around the Web & with SERPs. Think we need a bit more than these 7 tips

  • Pingback: Google Webmaster Tools: Gaining An Edge in Managing a Successful Website | Wil Gradwell

  • Najam Sethi

    I am having a huge cpu usage most of the time on my server and when I see the analytics in real-time sometimes it only shows 3 4 visitors per minutes but still a high cpu usage, I suspect this is due to Google Crawler, I have set the crawler to recommended settings but I wanna know how slow is the slowest point, that is 5 seconds it says, but what does 5 seconds mean? does it mean 5 secs for the whole website?

    Love to see any responses to my query,

    Thanks
    Najam Sethi

  • ego-t

    i have several hundred blocked pages on webmasters control panel, is there any tool where i can see which pages are blocked?

  • Doreen

    Hey,

    I recently tampered with the geographic target thing and lost my rankings in google.com. I changed it back but I’m still nowhere in sight. What can I do?

    Please help!

Leave a Reply