Jump to content

Cloaking: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
m Reverted edits by Dhanashree2019 (talk) to last version by Kuru
→‎See also: Remove duplicate link
 
(26 intermediate revisions by 21 users not shown)
Line 2: Line 2:
{{For|cloaking as used in extended DOS device drivers|Helix Cloaking}}
{{For|cloaking as used in extended DOS device drivers|Helix Cloaking}}


{{More citations needed|date=October 2021}}
'''Cloaking''' is a [[search engine optimization]] (SEO) technique in which the content presented to the [[search engine spider]] is different from that presented to the user's [[Web browser|browser]]. This is done by delivering content based on the [[IP address]]es or the [[User-Agent]] [[HTTP]] header of the user requesting the page. When a user is identified as a search engine spider, a server-side [[Scripting language|script]] delivers a different version of the [[web page]], one that contains content not present on the visible page, or that is present but not searchable. The purpose of cloaking is sometimes to deceive [[search engine]]s so they display the page when it would not otherwise be displayed ([[black hat hacking|black hat]] SEO). However, it can also be a functional (though antiquated) technique for informing search engines of content they would not otherwise be able to locate because it is embedded in non-textual containers such as video or certain [[Adobe Flash]] components. Since 2006, better methods of accessibility, including [[progressive enhancement]], have been available, so cloaking is no longer necessary for regular SEO.{{fact|date=November 2018}}
'''Cloaking''' is a [[search engine optimization]] (SEO) technique in which the content presented to the [[search engine spider]] is different from that presented to the user's [[Web browser|browser]]. This is done by delivering content based on the [[IP address]]es or the [[User-Agent]] [[HTTP]] header of the user requesting the page. When a user is identified as a search engine spider, a server-side [[Scripting language|script]] delivers a different version of the [[web page]], one that contains content not present on the visible page, or that is present but not searchable. The purpose of cloaking is sometimes to deceive [[search engine]]s so they display the page when it would not otherwise be displayed ([[black hat hacking|black hat]] SEO). However, it can also be a functional (though antiquated) technique for informing search engines of content they would not otherwise be able to locate because it is embedded in non-textual containers, such as video or certain [[Adobe Flash]] components. Since 2006, better methods of accessibility, including [[progressive enhancement]], have been available, so cloaking is no longer necessary for regular SEO.<ref>{{cite web |title=Cloaking {{!}} Google Search Central |url=https://developers.google.com/search/docs/advanced/guidelines/cloaking |website=Google Developers |language=en}}</ref>


Cloaking is often used as a [[spamdexing]] technique to attempt to sway search engines into giving the site a higher ranking. By the same method, it can also be used to trick search engine users into visiting a site that is substantially different from the search engine description, including delivering [[pornography|pornographic]] content cloaked within non-pornographic search results.
Cloaking is often used as a [[spamdexing]] technique to attempt to sway search engines into giving the site a higher ranking. By the same method, it can also be used to trick search engine users into visiting a site that is substantially different from the search engine description, including delivering [[pornography|pornographic]] content cloaked within non-pornographic search results.
Line 10: Line 11:
A similar technique is used on [[DMOZ]] [[web directory]], but it differs in several ways from search engine cloaking:
A similar technique is used on [[DMOZ]] [[web directory]], but it differs in several ways from search engine cloaking:
* It is intended to fool human editors, rather than computer search engine spiders.
* It is intended to fool human editors, rather than computer search engine spiders.
* The decision to cloak or not is often based upon the HTTP [[referrer]], the user agent or the visitor's IP; but more advanced techniques can be also based upon the client's behaviour analysis after a few page requests: the raw quantity, the sorting of, and latency between subsequent HTTP requests sent to a website's pages, plus the presence of a check for [[robots.txt]] file, are some of the parameters in which search engines spiders differ heavily from a natural user behaviour. The referrer tells the [[Uniform Resource Locator|URL]] of the page on which a user clicked a link to get to the page. Some cloakers will give the fake page to anyone who comes from a web directory website, since directory editors will usually examine sites by clicking on links that appear on a directory web page. Other cloakers give the fake page to everyone ''except'' those coming from a major search engine; this makes it harder to detect cloaking, while not costing them many visitors, since most people find websites by using a search engine.
* The decision to cloak or not is often based upon the HTTP [[referrer]], the user agent or the visitor's IP; but more advanced techniques can be also based upon the client's behaviour analysis after a few page requests: the raw quantity, the sorting of, and latency between subsequent HTTP requests sent to a website's pages, plus the presence of a check for [[robots.txt]] file, are some of the parameters in which search engines' spiders differ heavily from a natural user behaviour. The referrer tells the [[Uniform Resource Locator|URL]] of the page on which a user clicked a link to get to the page. Some cloakers will give the fake page to anyone who comes from a web directory website, since directory editors will usually examine sites by clicking on links that appear on a directory web page. Other cloakers give the fake page to everyone ''except'' those coming from a major search engine; this makes it harder to detect cloaking, while not costing them many visitors, since most people find websites by using a search engine.

==Black hat perspective==
{{unreferenced section|date=September 2011}}
Increasingly, for a page without natural popularity due to compelling or rewarding content to rank well in the search engines, webmasters design pages solely for the search engines. This results in pages with too many keywords and other factors that might be search engine "friendly", but make the pages difficult for actual visitors to consume. As such, black hat SEO practitioners consider cloaking to be an important technique to allow webmasters to split their efforts and separately target the search engine spiders and human visitors. Cloaking allows user experience to be high while satisfying the necessary minimum keyword concentration to rank in a search engine.

In September 2007, [[Ralph Tegtmeier]] and Ed Purkiss coined the term "mosaic cloaking" whereby dynamic pages are constructed as tiles of content and only portions of the pages, [[JavaScript]] and [[Cascading Style Sheets|CSS]] are changed, simultaneously decreasing the contrast between the cloaked page and the "friendly" page while increasing the capability for targeted delivery of content to various spiders and human visitors.


==Cloaking versus IP delivery==
==Cloaking versus IP delivery==
[[Geo targeting|IP delivery]] can be considered a more benign variation of cloaking, where different content is served based upon the requester's [[IP address]]. With cloaking, search engines and people never see the other's pages, whereas, with other uses of IP delivery, both search engines and people can see the same pages. This technique is sometimes used by graphics-heavy sites that have little textual content for spiders to analyze.
[[Geo targeting|IP delivery]] can be considered a more benign variation of cloaking, where different content is served based upon the requester's [[IP address]]. With cloaking, search engines and people never see the other's pages, whereas, with other uses of IP delivery, both search engines and people can see the same pages. This technique is sometimes used by graphics-heavy sites that have little textual content for spiders to analyze.<ref>{{Cite book |last=Eberwein |first=Helgo |url=https://www.worldcat.org/oclc/885168276 |title=Wettbewerbsrechtliche Aspekte von Domains und Suchmaschinen die Rechtslage in Deutschland und Österreich |date=2012 |isbn=978-3-8329-7890-7 |edition=1. Aufl |location=Baden-Baden |oclc=885168276}}</ref>


One use of IP delivery is to determine the requester's location, and deliver content specifically written for that country. This isn't necessarily cloaking. For instance, Google uses IP delivery for [[AdWords]] and [[AdSense]] [[advertising]] programs to target users in different geographic locations.
One use of IP delivery is to determine the requester's location, and deliver content specifically written for that country. This isn't necessarily cloaking. For instance, Google uses IP delivery for [[AdWords]] and [[AdSense]] [[advertising]] programs to target users in different geographic locations.
Line 30: Line 25:
** [[Keyword stuffing]]
** [[Keyword stuffing]]
** [[Link farm]]s
** [[Link farm]]s
** [[Spamdexing]]
** [[URL redirection]]
** [[URL redirection]]
** [[Page hijacking]]
* '''Technology:'''
* '''Technology:'''
** [[Content negotiation]]
** [[Content negotiation]]
Line 39: Line 32:
==References==
==References==
{{reflist}}
{{reflist}}

==Further reading==
* Baoning Wu and Brian D. Davison: "[http://airweb.cse.lehigh.edu/2005/wu.pdf Cloaking and Redirection: A Preliminary Study]". Workshop on Adversarial Information Retrieval on the Web, Chiba, Japan, 2005.
* Baoning Wu and Brian D. Davison: "[http://airweb.cse.lehigh.edu/2005/wu.pdf Cloaking and Redirection: A Preliminary Study]". Workshop on Adversarial Information Retrieval on the Web, Chiba, Japan, 2005.


{{Spamming}}
[[Category:Black hat search engine optimization]]
[[Category:Black hat search engine optimization]]

Latest revision as of 18:55, 8 June 2024

Cloaking is a search engine optimization (SEO) technique in which the content presented to the search engine spider is different from that presented to the user's browser. This is done by delivering content based on the IP addresses or the User-Agent HTTP header of the user requesting the page. When a user is identified as a search engine spider, a server-side script delivers a different version of the web page, one that contains content not present on the visible page, or that is present but not searchable. The purpose of cloaking is sometimes to deceive search engines so they display the page when it would not otherwise be displayed (black hat SEO). However, it can also be a functional (though antiquated) technique for informing search engines of content they would not otherwise be able to locate because it is embedded in non-textual containers, such as video or certain Adobe Flash components. Since 2006, better methods of accessibility, including progressive enhancement, have been available, so cloaking is no longer necessary for regular SEO.[1]

Cloaking is often used as a spamdexing technique to attempt to sway search engines into giving the site a higher ranking. By the same method, it can also be used to trick search engine users into visiting a site that is substantially different from the search engine description, including delivering pornographic content cloaked within non-pornographic search results.

Cloaking is a form of the doorway page technique.

A similar technique is used on DMOZ web directory, but it differs in several ways from search engine cloaking:

  • It is intended to fool human editors, rather than computer search engine spiders.
  • The decision to cloak or not is often based upon the HTTP referrer, the user agent or the visitor's IP; but more advanced techniques can be also based upon the client's behaviour analysis after a few page requests: the raw quantity, the sorting of, and latency between subsequent HTTP requests sent to a website's pages, plus the presence of a check for robots.txt file, are some of the parameters in which search engines' spiders differ heavily from a natural user behaviour. The referrer tells the URL of the page on which a user clicked a link to get to the page. Some cloakers will give the fake page to anyone who comes from a web directory website, since directory editors will usually examine sites by clicking on links that appear on a directory web page. Other cloakers give the fake page to everyone except those coming from a major search engine; this makes it harder to detect cloaking, while not costing them many visitors, since most people find websites by using a search engine.

Cloaking versus IP delivery

[edit]

IP delivery can be considered a more benign variation of cloaking, where different content is served based upon the requester's IP address. With cloaking, search engines and people never see the other's pages, whereas, with other uses of IP delivery, both search engines and people can see the same pages. This technique is sometimes used by graphics-heavy sites that have little textual content for spiders to analyze.[2]

One use of IP delivery is to determine the requester's location, and deliver content specifically written for that country. This isn't necessarily cloaking. For instance, Google uses IP delivery for AdWords and AdSense advertising programs to target users in different geographic locations.

IP delivery is a crude and unreliable method of determining the language in which to provide content. Many countries and regions are multilingual, or the requestor may be a foreign national. A better method of content negotiation is to examine the client's Accept-Language HTTP header.

See also

[edit]

References

[edit]
  1. ^ "Cloaking | Google Search Central". Google Developers.
  2. ^ Eberwein, Helgo (2012). Wettbewerbsrechtliche Aspekte von Domains und Suchmaschinen die Rechtslage in Deutschland und Österreich (1. Aufl ed.). Baden-Baden. ISBN 978-3-8329-7890-7. OCLC 885168276.{{cite book}}: CS1 maint: location missing publisher (link)

Further reading

[edit]