Merchant Stories

Page Indexing by Google: 3 Ways Google Indexes Your Store

Apr 8, 2014 1 min read 567 views
Listen audio
Page Indexing by Google: 3 Ways Google Indexes Your Store

Fast and effective site indexing is the main part of any internet projects success. Let us review how Google indexes pages, what a ‘bot’ is and what their different types are.

General Page Indexing information

A bot is part of server software, the main purpose of which is checking your website for updates, investigation of all site pages, adding them to the database and spreading the information about the site among the search results in a specific search engine.

Google search engine has several known bots: Adsensebot, Freshbot, and DeepCrawl.

Adsensebot is mainly intended for webmasters who use Adsense context advertisement on their sites. At the moment when the site is updated (a new page, product, text added) JavaScript from Adsense code sends the message to Adsensebot, it visits the page within 10 – 15 minutes and indexes its content. This action is necessary in order to have the ability to place useful search advertisements on the site pages.

Freshbot is a bot of popularity. It visits the most popular and well-visited pages of a certain site. The number of its visits may vary depending on the frequency of the site updates and customers’ visits. For instance, such gigantic platforms as eBay or Amazon are visited by Freshbot every 10 minutes or even more often. Average sites with the number of users up to 500 per 24 hours are visited once in 4 – 10 days. This bot checks all website links and pages and places them in the search engine database, and after that, they are checked by DeepCrawl.

DeepCrawl is the most complicated bot, it checks all data that are placed into the database by Freshbot, even its name shows how deep that check is. This is necessary in order to place as many sites as possible into the full Google index. It visits a site once in a month, sometimes twice, if it is a large web store or a news portal.

Sometimes, checking the site by bots is complicated by so-called ‘trash’ in the results of their work. For example, a Supplemental filter may ban a category page, search and filtering results. This negatively affects the site rating and often prevents bots from checking the site. In order to avoid that you need to add several code lines into robots.txt file. They are as follows:

Disallow: */tag/*

Disallow: */author/*

Disallow: */page/*

 

You may also use GoMage Advanced Navigation extension that allows you to hide product filtering and search results from indexing, which significantly reduces the number of pages banned by a Supplemental filter.

And have a look at the other Extensions for Magento ® which will help you to increase sales on your Magento ® store.

That's where you contact us!

    By submitting this form you agree to GoMage's Terms of Use and Privacy Policy
    woo-hoo! Now its time to keep checking your inbox, as we will be getting in touch soon. Promise :)
    oops! Thanks. But it seems like some kind of technical issues stop you from meeting GOMAGE. Could you try again?