Apr 202014

A multi-language site offers several versions of the same content, one version for each of the languages in the site. To make search engines aware of this structure, each page much include additional information that allows the search engine to identify the alternate versions of the page. Otherwise, those pages could be treated as duplicates, or not be included in the index.

Google gives some guidelines on how to include these metadata. Those guidelines are briefly summarized in this post.

Continue reading »

 Posted by at 3:06 pm
Mar 172013

An XML sitemap is a document that lists the URLs in a web site, with information about their relative importance and how often they are modified. This information is taken into account by search engines, such as Google and Yahoo, to adjust their crawling rate to index the site.

This post explains how to automate the generation of the sitemap and its submission to Google and Bing

Continue reading »

 Posted by at 10:20 am
Mar 102013

Nowadays, many websites implement dynamic pages: When the page is loaded, or when the user interacts with the page, some of the page elements are updated with new content retrieved from the server by means of ajax javascript calls. This technique avoids having to send all the content that does not change (html code of the header and sidebar menu, css, js, images,…) with every request.

The problem with dynamic pages is that Googlebot (and crawlers of other search engines) does not know about the existence of those dynamic pages, and therefore they do not get crawled and indexed and do not appear in the search results pages of the search engine. To solve this issue, Google has established a way to setup the website to allow crawling of dynamic content. The details of this setup are explained in this post. Continue reading »

 Posted by at 3:55 pm