[Search engine optimization] how Baidu search engines to identify content replication

Original content for inclusion and ranking of sites is critical, so a lot of copy website content will affect the site keyword ranking. For example, you originally wrote the article, should have been ranked very good, but others copied or reproduced in your article, but unfortunately, the search engine determines thing I read was copied or reproduced on the page is the original source, you should just rankings It will be taken away that page.

We illustrate Baidu search engine discovery process to copy the contents of the Web. There are three assumptions:

First, the text of each page are assumed to copy the contents of the page contains.

Second, copy the contents of each page on the assumption that in different domains.

Third, the following steps have been simplified to make the process simple and clear as possible. Almost certainly with Baidu methods used by search engines is not exactly the same.

There are a few facts about the copied content worth mentioning, or to copy the contents of the issue relatively unknown owners may make mistakes.

Copy the contents of the location

If you are on my website, copy the contents of it? Yes, in fact copy the contents may occur within a site, it can also appear on different sites.

Copy the contents of proportion

What percentage of copied content will be filtered out? Unfortunately, the search engine will never disclose this information because it would hurt their ability to prevent this problem on a page.

And almost certainly, the percentage of use of search engines often change more than once simply copy the contents of a direct comparison is detected. Page does not have to be exactly the same as before replication.

The proportion of the code text

If there is a lot of code on the page but not many unique HTML elements of how to do? Google will not think these pages are copied contents? Not. Search engines do not care about your code, they are interested in the content of the page. Only in very extreme cases the amount of code became a problem.

The ratio of unique content and navigation elements

My website each page has a lot of navigation, a lot of header and footer elements, but only a little content, Google will not put these pages as a copy? Not. Google before evaluating whether to copy a page common page elements such as navigation will be excluded. They are very familiar with the site layout, know all the pages on a site have the same result as very normal. They are more concerned about the "unique" part of each page, often ignore the rest.

Reproduced with permission from the content

If I were black heart authorization to upload content from the Internet to other places, how to avoid copying the contents of the problem? Use this line of code in the page header, the search engine will know the content is not prepared for them. Usually this is the best approach, because users can also access the page, links the past, the links on the page can also pass the weight.

This article from: http://www.qzgi.cn/archives/1127

Guess you like

Origin blog.csdn.net/weixin_44905281/article/details/91039109