Wednesday, 12 October 2011

Indexation of a site

   Before the site will appear in search results, it should be indexed a search engine. Indexation means that the search robot has visited your site, has analysed it and has brought the information in a search engine database.

 
   If some page is brought in an  
index of the searcher 
it can be shown in search results. If the page in an index is absent, the search engine knows nothing about it, and, hence, cannot use the information from this page in any way.

   The majority of sites of the average size (that is tens containing some or hundreds pages) usually have no problems with correct indexation by search engines. However, there is a number of the moments which should be considered at work on a site. 
   The search engine can learn about again created site two ways:

   - Manual addition of the address of a site through the corresponding form of a search engine. In this case you inform a search engine on a new site and its address gets to turn on indexation. It is necessary to add only main page of a site, the others will be found by the search robot under links;

   - To give to the search robot independently to find your site. If on your new resource there is at least one deep link from other resources, already indexed a search engine the search robot in short terms itself will visit and проиндексирует your site. In most cases it is recommended to use this variant, that is to receive some deep links to a site and simply to wait robot arrival. Manual addition of a site can even extend a robot waiting time.

   Time necessary for site indexation makes, as a rule, from 2-3 days till 2 weeks, depending on a search engine. Sites the Google search engine fastest indexes.

   Try to make a site friendly for search robots. For this purpose consider following factors:

   - Try, that any pages of your site were accessible under links from the main page no more than for 3 transitions. If the structure of a site of it does not suppose, make a so-called sitemap which will allow to carry out the specified rule;

   - Do not repeat widespread errors. Identifiers of sessions complicate indexation. If you use navigation through scripts necessarily duplicate links in the usual image – search engines are not able to read scripts (more in detail about these and other errors it is told in chapter 2.3);

    - Remember that search engines index no more than 100-200 кб the text on page. For pages more volume the page beginning (the first 100-200 кб will be indexed only.). The rule From this follows – do not use page in the size more than 100 kb if want that they have been indexed completely.

   To operate behaviour of search robots it is possible by means of a file robots.txt, in it it is possible to resolve or forbid obviously for those indexations or other pages. Exists also special тег «NOINDEX», allowing to close for indexation separate parts of page, however this тег is supported only by the Russian search engines.

   Databases of search engines are constantly updated, records in base can be exposed to changes, disappear and appear again, therefore the number of indexed pages of your site can periodically change.

   One of the most frequent reasons of disappearance of page of an index is inaccessibility of the server, that is the search robot at attempt of indexation of a site could not get access to it. After restoration of working capacity of the server the site should appear in an index again after a while.

   It is necessary to notice also that the more deep links has your site, the there is its reindexation faster.

   To trace process of indexation of a site it is possible by means of the analysis of broad gulls-files of the server in which all visits of search robots register. In corresponding section we will in detail tell about programs which allow to make it.

No comments:

Post a Comment