Page 1 of 1

Indexing identical documents in different encodings

Posted: Tue Jan 28, 2025 5:24 am
by subornaakter20
If web servers issue identical documents in different encodings, then the resource consumption during their indexing is very high. In the databases of domestic search engines, documents are stored in one encoding. It is recommended to use only one encoding for indexing.

Download a useful document:

Checklist: How to Achieve Your Goals in Negotiations with Clients
If encodings are issued via server ports, then mail marketing to doctors different robots.txt must be issued on different servers/ports. That is, in each server/port, you must additionally register:

user-agent: *


If encodings are issued by directories, then you will need a single robots.txt file, in which you should write:



Indexing one site on different servers
The problem may arise if the server has mirrors or if encodings are obtained by changing the prefix in the host name. For example, encodings for the may be as follows: win.chto-to.ru, koi-The robot cannot determine which of these addresses is the main one. It only sees that there are two completely matching documents and indexes one of them.

That is, it is possible that your content will be indexed on different hosts. To index a resource at one address, you need to specify the corresponding robots.txt file. After a certain time, the indexed documents will refer to the desired address.

Read also!

"Page dropped from index: obvious and hidden reasons"
Read more
It is not recommended to open your resource content to search engine robots under different names or ports. You will not be able to improve indexing. The server will experience the load, as will the network. This can lead to negative consequences.

For example, search engines do not always understand which of your site's "mirrors" is the main one. They can choose any. Some search engines even have settings that allow deleting sites whose content is identical.

You can avoid this problem by disallowing indexing of any addresses that are not primary when writing your site's robots.txt file.

To identify existing errors, we recommend conducting a site audit. We will conduct a full analysis of internal and external optimization errors for you, and also provide 47+ recommendations for increasing traffic and sales from your site.