The Beginners Guide To SEO
The Beginners Guide To SEO
The Beginners Guide To SEO
Imagine the World Wide Web as a network of stops in a big city subway
system.
Each stop is a unique document (usually a web page, but sometimes a PDF, JPG, or other
file). The search engines need a way to crawl the entire city and find all the stops along the
way, so they use the best path availablelinks.
The link structure of the web serves to bind all of the pages together.
Links allow the search engines' automated robots, called "crawlers" or "spiders," to reach the
many billions of interconnected documents on the web.
Once the engines find these pages, they decipher the code from them and store selected
pieces in massive databases, to be recalled later when needed for a search query. To
accomplish the monumental task of holding billions of pages that can be accessed in a
fraction of a second, the search engine companies have constructed datacenters all over the
world.
These monstrous storage facilities hold thousands of machines processing large quantities of
information very quickly. When a person performs a search at any of the major engines, they
demand results instantaneously; even a one- or two-second delay can cause dissatisfaction,
so the engines work hard to provide answers as fast as possible.
1.
2.
Providing Answers
Search engines are answer machines. When a person performs an online search, the search
engine scours its corpus of billions of documents and does two things: first, it returns only
those results that are relevant or useful to the searcher's query; second, it ranks those results
according to the popularity of the websites serving the information. It is both relevance and
popularity that the process of SEO is meant to influence.
There is perhaps no greater tool available to webmasters researching the activities of the engines than the freedom to use the search engines
themselves to perform experiments, test hypotheses, and form opinions. It is through this iterativesometimes painstakingprocess that a
considerable amount of knowledge about the functions of the engines has been gleaned. Some of the experiments weve tried go something
like this:
1.
2.
5.
ishkabibbell.com).
6.
3.
4.
7.
on other domains.
1.
2.
We like to say, "Build for users, not for search engines." There are three types of search
queries people generally make:
3.
4.
5.
Click on a result.
6.
"Do" Transactional Queries: I want to do something, such as buy a plane ticket or listen to a
song.
"Know" Informational Queries: I need information, such as the name of a band or the best
restaurant in New York City.
"Go" Navigation Queries: I want to go to a particular place on the Intrernet, such as Facebook
solution.
7.
users. So ask yourself what your target customers are looking for and make sure your site
delivers it to them.
It all starts with words typed into a small box.
8.
view
view
goals.
when the majority of the people who would visit your website
are from Japan.
Mixed contextual signals: For example, the title of your blog
post is "Mexico's Best Coee" but the post itself is about a
vacation resort in Canada which happens to serve great coee.
These mixed messages send confusing signals to search
engines.
Take a look at any search results page and you'll find the answer to why search marketing
has a long, healthy life ahead.
There are, on average, ten positions on the search results page. The pages that fill those positions are
ordered by rank. The higher your page is on the search results page, the better your click-through
rate and ability to attract searchers. Results in positions 1, 2, and 3 receive much more traffic than
results down the page, and considerably more than results on deeper pages. The fact that so much
attention goes to so few listings means that there will always be a financial incentive for search
engine rankings. No matter how search may change in the future, websites and businesses will
compete with one another for this attention, and for the user traffic and brand visibility it provides.
Search engines are limited in how they crawl the web and
interpret content. A webpage doesn't always look the same to
you and me as it looks to a search engine. In this section, we'll
focus on specific technical aspects of building (or modifying)
web pages so they are structured for both search engines and
human visitors alike. Share this part of the guide with your
programmers, information architects, and designers, so that all
parties involved in a site's construction are on the same page.
Indexable Content
To perform better in search engine listings, your most important content should be in HTML
text format. Images, Flash files, Java applets, and other non-text content are often ignored or
devalued by search engine crawlers, despite advances in crawling technology. The easiest
way to ensure that the words and phrases you display to your visitors are visible to search
engines is to place them in the HTML text on the page. However, more advanced methods are
available for those who demand greater formatting or visual display styles:
1.
3.
4.
content.
2.
Link tags can contain images, text, or other objects, all of which provide a clickable area on the page that users can engage to move to
another page. These links are the original navigational elements of the Internet known as hyperlinks. In the above illustration, the "<a"
tag indicates the start of a link. The link referral location tells the browser (and the search engines) where the link points. In this example,
the URL http://www.jonwye.com is referenced. Next, the visible portion of the link for visitors, called anchor text in the SEO world,
describes the page the link points to. The linked-to page is about custom belts made by Jon Wye, thus the anchor text "Jon Wye's Custom
Designed Belts." The "</a>" tag closes the link to constrain the linked text between the tags and prevent the link from encompassing other
elements on the page.
This is the most basic format of a link, and it is eminently understandable to the search engines. The crawlers know that they should add
this link to the engines' link graph of the web, use it to calculate query-independent variables (like Google's PageRank), and follow it to
index the contents of the referenced page.
Submission-required forms
either do not crawl or give very little weight to the links embedded
within. Standard HTML links should replace JavaScript (or
accompany it) on any page you'd like crawlers to crawl.
Frames or iframes
The links embedded inside the Juggling Panda site (from our
above example) are perfect illustrations of this phenomenon.
Although dozens of pandas are listed and linked to on the page, no
crawler can reach them through the site's link structure, rendering
them invisible to the engines and hidden from users' search
queries.
Keyword Domination
Keywords dominate how we communicate our search intent and
interact with the engines. When we enter words to search for, the
engine matches pages to retrieve based on the words we entered.
The order of the words ("pandas juggling" vs. "juggling pandas"),
spelling, punctuation, and capitalization provide additional
information that the engines use to help retrieve the right pages
and rank them.
Search engines measure how keywords are used on pages to help
determine the relevance of a particular document to a query. One
of the best ways to optimize a page's rankings is to ensure that the
keywords you want to rank for are prominently used in titles, text,
and metadata.
Generally speaking, as you make your keywords more specific,
you narrow the competition for search results, and improve your
changes of achieving a higher ranking. The map graphic to the left
compares the relevance of the broad term "books" to the specific
title Tale of Two Cities. Notice that while there are a lot of results for
the broad term, there are considerably fewer results (and thus, less
competition) for the specific result.
Keyword Abuse
Since the dawn of online search, folks have abused keywords in a
misguided eort to manipulate the engines. This involves "stung"
keywords into text, URLs, meta tags, and links. Unfortunately, this
tactic almost always does more harm than good for your site.
In the early days, search engines relied on keyword usage as a
prime relevancy signal, regardless of how the keywords were
actually used. Today, although search engines still can't read and
comprehend text as well as a human, the use of machine learning
has allowed them to get closer to this ideal.
The best practice is to use your keywords naturally and
strategically (more on this below). If your page targets the keyword
phrase "Eiel Tower" then you might naturally include content
about the Eiel Tower itself, the history of the tower, or even
recommended Paris hotels. On the other hand, if you simply
sprinkle the words "Eiel Tower" onto a page with irrelevant
content, such as a page about dog breeding, then your eorts to
rank for "Eiel Tower" will be a long, uphill battle. The point of
using keywords is not to rank highly for all keywords, but to
rank highly for the keywords that people are searching for
when they want what your site provides.
On-Page Optimization
Keyword usage and targeting are still a part of the search engines'
1.
2.
3.
4.
The Conclusion:
What should optimal page density look like then? An optimal page
for the phrase running shoes would look something like:
In the title tag at least once. Try to keep the keyword phrase as
close to the beginning of the title tag as possible. More detail on
title tags follows later in this section.
Once prominently near the top of the page.
At least two or three times, including variations, in the body
copy on the page. Perhaps a few more times if there's a lot of
text content. You may find additional value in using the keyword
or variations more than this, but in our experience adding more
instances of a term or phrase tends to have little or no impact
on rankings.
At least once in the alt attribute of an image on the page. This
not only helps with web search, but also image search, which
can occasionally bring valuable trac.
Once in the URL. Additional rules for URLs and keywords are
discussed later on in this section.
At least once in the meta description tag. Note that the meta
description tag does not get used by the engines for rankings,
but rather helps to attract clicks by searchers reading the results
page, as the meta description becomes the snippet of text used
by the search engines.
And you should generally not use keywords in link anchor text
pointing to other pages on your site; this is known as Keyword
Cannibalization.
Title Tags
The title element of a page is meant to be an accurate, concise
description of a page's content. It is critical to both user
experience and search engine optimization.
As title tags are such an important part of search engine
optimization, the following best practices for title tag creation
makes for terrific low-hanging SEO fruit. The recommendations
below cover the critical steps to optimize title tags for search
engines and for usability.
Be mindful of length
The title tag of any page appears at the top of Internet browsing
software, and is often used as the title when your content is shared
through social media or republished.
Search engines display only the first 65-75 characters of a title tag
in the search results (after that, the engines show an ellipsis "..."
to indicate when a title tag has been cut o). This is also the
general limit allowed by most social media sites, so sticking to this
limit is generally wise. However, if you're targeting multiple
keywords (or an especially long keyword phrase), and having them
in the title tag is essential to ranking, it may be advisable to go
longer.
Include branding
Using keywords in the title tag means that search engines will bold
those terms in the search results when a user has performed a
At Moz, we love to end every title tag with a brand name mention,
as these help to increase brand awareness, and create a higher
click-through rate for people who like and are familiar with a brand.
Sometimes it makes sense to place your brand at the beginning of
query with those terms. This helps garner a greater visibility and
a higher click-through rate.
the title tag, such as your homepage. Since words at the beginning
of the title tag carry more weight, be mindful of what you are trying
to rank for.
Meta Tags
Meta tags were originally intended as a proxy for information about a website's content.
Several of the basic meta tags are listed below, along with a description of their use.
Meta Robots
The Meta Robots tag can be used to control search engine crawler activity (for all of the major
engines) on a per-page level. There are several ways to use Meta Robots to control how
search engines treat a page:
index/noindex tells the engines whether the page should be crawled and kept in the engines'
index for retrieval. If you opt to use "noindex," the page will be excluded from the index. By
default, search engines assume they can index all pages, so using the "index" value is
generally unnecessary.
follow/nofollow tells the engines whether links on the page should be crawled. If you elect to
employ "nofollow," the engines will disregard the links on the page for discovery, ranking
purposes, or both. By default, all pages are assumed to have the "follow" attribute.
Example: <META NAME="ROBOTS" CONTENT="NOINDEX, NOFOLLOW">
noarchive is used to restrict search engines from saving a cached copy of the page. By
default, the engines will maintain visible copies of all pages they have indexed, accessible to
searchers through the cached link in the search results.
nosnippet informs the engines that they should refrain from displaying a descriptive block of
text next to the page's title and URL in the search results.
noodp/noydir are specialized tags telling the engines not to grab a descriptive snippet about
a page from the Open Directory Project (DMOZ) or the Yahoo! Directory for display in the
search results.
The X-Robots-Tag HTTP header directive also accomplishes these same objectives. This
technique works especially well for content within non-HTML files, like images.
Meta Description
The meta description tag exists as a short description of a page's content. Search engines do
not use the keywords or phrases in this tag for rankings, but meta descriptions are the primary
source for the snippet of text displayed beneath a listing in the results.
The meta description tag serves the function of advertising copy, drawing readers to your site
from the results. It is an extremely important part of search marketing. Crafting a readable,
compelling description using important keywords (notice how Google bolds the searched
keywords in the description) can draw a much higher click-through rate of searchers to your
page.
Meta descriptions can be any length, but search engines generally will cut snippets longer
than 160 characters, so it's generally wise to stay within in these limits.
In the absence of meta descriptions, search engines will create the search snippet from other
elements of the page. For pages that target multiple keywords and topics, this is a perfectly
valid tactic.
URL Structures
URLsthe addresses for documents on the webare of great value from a search
perspective. They appear in multiple important locations.
Shorter is better
While a descriptive URL is important, minimizing length and trailing
slashes will make your URLs easier to copy and paste (into emails,
blog posts, text messages, etc.) and will be fully visible in the
search results.
Go static
The best URLs are human-readable and without lots of
parameters, numbers, and symbols. Using technologies like
mod_rewrite for Apache and ISAPI_rewrite for Microsoft, you can
easily transform dynamic URLs like this http://moz.com/blog?
Instead, if the site owner took those three pages and 301redirected them, the search engines would have only one
strong page to show in the listings from that site.
From an SEO perspective, the Canonical URL tag attribute is similar to a 301 redirect. In
essence, you're telling the engines that multiple pages should be considered as one (which a
301 does), but without actually redirecting visitors to the new URL. This has the added bonus
of saving your development sta considerable heartache.
For more about dierent types of duplicate content, this post by Dr. Pete deserves special
mention.
Rich Snippets
Ever see a 5-star rating in a search result? Chances are, the search
engine received that information from rich snippets embedded on
the webpage. Rich snippets are a type of structured data that
allow webmasters to mark up content in ways that provide
information to the search engines.
While the use of rich snippets and structured data is not a required
element of search engine-friendly design, its growing adoption
means that webmasters who employ it may enjoy an advantage in
some circumstances.
Structured data means adding markup to your content so that
search engines can easily identify what type of content it is.
Schema.org provides some examples of data that can benefit from
structured markup, including people, products, reviews,
businesses, recipes, and events.
Often the search engines include structured data in search results,
such as in the case of user reviews (stars) and author profiles
(pictures). There are several good resources for learning more
Ask yourself...
Is the keyword relevant to your website's content? Will
searchers find what they are looking for on your site
when they search using these keywords? Will they be
happy with what they find? Will this traffic result in
financial rewards or other organizational goals? If the
answer to all of these questions is a clear "Yes!" then
proceed ...
Keyword Research
Resources
Where do we get all of this knowledge about keyword demand and
keyword referrals? From research sources like these:
Google AdWords Keyword Planner Tool
Google Trends
Microsoft Bing Ads Intelligence
Wordtrackers Free Basic Keyword Demand
Keyword Difficulty
What are my chances of success?
In order to know which keywords to target, it's essential to not only
understand the demand for a given term or phrase, but also the
work required to achieve high rankings. If big brands take the top
10 results and you're just starting out on the web, the uphill battle
for rankings can take years of eort. This is why it's essential to
understand keyword diculty.
2. Machine Learning
In 2011 Google introduced the Panda update to its ranking algorithm, significantly changing
the way it judged websites for quality. Google started by using human evaluators to manually
rate thousands of sites, searching for low quality content. Google then incorporated machine
learning to mimic the human evaluators. Once its computers could accurately predict what the
humans would judge a low quality site, the algorithm was introduced across millions of sites
spanning the Internet. The end result was a seismic shift that rearranged over 20% of all of
Google's search results. For more on the Panda update, some good resources can be found
here and here.
3. Linking Patterns
The engines discovered early on that the link structure of the web could serve as a proxy for
votes and popularity; higher quality sites and information earned more links than their less
useful, lower quality peers. Today, link analysis algorithms have advanced considerably, but
these principles hold true.
Crafting Content
For search engine success
"Develop great content" may be the most oft-repeated suggestion in the SEO world. Despite
its clichd status, though, this is sound advice. Appealing, useful content is crucial to search
engine optimization. Every search performed at the engines comes with an intentto find,
learn, solve, buy, fix, treat, or understand. Search engines place web pages in their results in
order to satisfy that intent in the best possible way. Crafting fulfilling, thorough content that
addresses searchers' needs improved your chance to earn top rankings.
Transactional Searches
Navigational Searches
Informational Searches
Fulfilling these intents is up to you. Creativity, high-quality writing, use of examples, and inclusion
of images and multimedia can all help in crafting content that perfectly matches a searcher's goals.
Your reward is satisfied searchers who demonstrate their positive experience through engagement
with your site or with links to it.
For search engines that crawl the vast metropolis of the web,
links are the streets between pages. Using sophisticated link
analysis, the engines can discover how pages are related to each
other and in what ways.
Since the late 1990s search engines have treated links as votes for popularity and importance
in the ongoing democratic opinion poll of the web. The engines themselves have refined the
use of link data to a fine art, and use complex algorithms to perform nuanced evaluations of
sites and pages based on this information.
Links aren't everything in SEO, but search professionals attribute a large portion of the
engines' algorithms to link-related factors (see Search Engine Ranking Factors). Through links,
engines can not only analyze the popularity websites and pages based on the number and
popularity of pages linking to them, but also metrics like trust, spam, and authority.
Trustworthy sites tend to link to other trusted sites, while spammy sites receive very few links
from trusted sources (see MozTrust). Authority models, like those postulated in the Hilltop
Algorithm, suggest that links are a very good way of identifying expert documents on a given
subject.
Link Signals
Used by search engines
How do search engines assign value to links? To answer this, we
need to explore the individual elements of a link, and look at how
the search engines assess these elements. We don't fully
understand the proprietary metrics that search engines use, but
through analysis of patent applications, years of experience, and
hands-on testing, we can draw some intelligent assumptions that
hold up in the real world. Below is a list of notable factors worthy
of consideration. These signals, and many more, are considered by
professional SEOs when measuring link value and a site's link
profile. You may also enjoy some further on the Moz Blog reading
about search engine valuation of links.
Global Popularity
The more popular and important a site is, the more links from that
site matter. A site like Wikipedia has thousands of diverse sites
linking to it, which means it's probably a popular and important
site. To earn trust and authority with the engines, you'll need the
help of other link partners. The more popular, the better.
Local/Topic-Specific Popularity
The concept of "local" popularity, first pioneered by the Teoma
Anchor Text
One of the strongest signals the engines use in rankings is anchor
text. If dozens of links point to a page with the right keywords, that
page has a very good probability of ranking well for the targeted
phrase in that anchor text. You can see examples of this in action
with searches like "click here," where many results rank solely due
to the anchor text of inbound links.
TrustRank
It's no surprise that the Internet contains massive amounts of
spam. Some estimate as much as 60% of the web's pages are
spam. In order to weed out this irrelevant content, search engines
use systems for measuring trust, many of which are based on the
link graph. Earning links from highly-trusted domains can result in
a significant boost to this scoring metric. Universities, government
websites and non-profit organizations represent examples of hightrust domains.
Link Neighborhood
Spam links often go both ways. A website that links to spam is
likely spam itself, and in turn often has many spam sites linking
back to it. By looking at these links in the aggregate, search
engines can understand the "link neighborhood" in which your
website exists. Thus, it's wise to choose those sites you link to
carefully and be equally selective with the sites you attempt to earn
links from.
Freshness
Link signals tend to decay over time. Sites that were once popular
often go stale, and eventually fail to earn new links. Thus, it's
important to continue earning additional links over time. Commonly
referred to as "FreshRank," search engines use the freshness
signals of links to judge current popularity and relevance.
Social Sharing
The last few years have seen an explosion in the amount of
content shared through social services such as Facebook, Twitter,
and Google+. Although search engines treat socially shared links
dierently than other types of links, they notice them nonetheless.
There is much debate among search professionals as to how
exactly search engines factor social link signals into their
algorithms, but there is no denying the rising importance of social
channels.
In a word: no. Although there is evidence that social shares such as Tweets, Likes, and Plusses affect rankings, at this time links are
considered a far superior and more lasting way to promote the popularity of your content than any other method.
As with any marketing activity, the first step in any link building campaign is the creation of
goals and strategies. Unfortunately, link building is one of the most dicult activities to
measure. Although the engines internally weigh each link with precise, mathematical metrics,
it's impossible for those on the outside to access this information.
SEOs rely on a number of signals to help build a rating scale of link value. Along with the data
from the link signals mentioned above, these metrics include the following:
Competitor's Backlinks
MozRank
Domain Authority
Moz Domain Authority (or DA) is a query-independent measure of
how likely a domain is to rank for any given query. DA is calculated
by analyzing the Internet's domain graph and comparing a given
domain to tens of thousands of queries in Google.
It takes time, practice, and experience to build comfort with these variables as they relate to search engine traffic. However, using your
websites analytics, you should be able to determine whether your campaign is successful.
Success comes when you see increases in search traffic, higher rankings, more frequent search engine crawling and increases in
referring link traffic. If these metrics do not rise after a successful link building campaign, it's possible you either need to seek better
quality link targets, or improve your on-page optimization.
Be newsworthy
Earning the attention of the press, bloggers and news media is an eective, time-honored way
to earn links. Sometimes this is as simple as giving away something for free, releasing a great
new product, or stating something controversial.
SEOs tend to use a lot of tools. Some of the most useful are provided by the
search engines themselves. Search engines want webmasters to create sites
and content in accessible ways, so they provide a variety of tools, analytics
and guidance. These free resources provide data points and unique
opportunities for exchanging information with the engines.
Below we explain the common elements that each of the major search engines support and
identify why they are useful.
XML
Extensible Markup Language (recommended format)
This is the most widely accepted format for sitemaps. It is
extremely easy for search engines to parse and can be
produced by a plethora of sitemap generators. Additionally, it
allows for the most granular control of page parameters.
Relatively large file sizes. Since XML requires an open tag and
a close tag around each element, file sizes can get very large.
RSS
Txt
Text File
Extremely easy. The text sitemap format is one URL per line
up to 50,000 lines.
2. Robots.txt
Disallow
Prevents compliant robots from accessing specific pages or
folders.
Sitemap
Indicates the location of a websites sitemap or sitemaps.
Crawl Delay
Indicates the speed (in milliseconds) at which a robot can crawl a
server.
An Example of Robots.txt
#Robots.txt www.example.com/robots.txt
User-agent: *
Disallow:
# Dont allow spambot to crawl any pages
User-agent: spambot
disallow: /
sitemap:www.example.com/sitemap.xml
3. Meta Robots
The meta robots tag creates page-level instructions for search
engine bots.
The meta robots tag should be included in the head section of the
HTML document.
<body>
<h1>Hello World</h1>
</body>
</html>
4. Rel="Nofollow"
Remember how links act as votes? The rel=nofollow attribute
allows you to link to a resource, while removing your "vote" for
search engine purposes. Literally, "nofollow" tells search engines
not to follow the link, although some engines still follow them to
discover new pages. These links certainly pass less value (and in
most cases no juice) than their followed counterparts, but are
useful in various situations where you link to an untrusted source.
An Example of nofollow
<a href="http://www.example.com" title="Example"
rel="nofollow">Example Link</a>
5. Rel="canonical"
Often, two or more copies of the exact same content appear on
your website under dierent URLs. For example, the following
URLs can all refer to a single homepage:
http://www.example.com/
http://www.example.com/default.asp
http://example.com/
http://example.com/default.asp
http://Example.com/Default.asp
To search engines, these appear as five separate pages. Because
the content is identical on each page, this can cause the search
engines to devalue the content and its potential rankings.
The canonical tag solves this problem by telling search robots
which page is the singular, authoritative version that should count
in web results.
one.
Key Features
Geographic Target - If a given site targets users in a particular
location, webmasters can provide Google with information that will
help determine how that site appears in its country-specific search
results, and also improve Google search results for geographic
queries.
Preferred Domain - The preferred domain is the one that a
webmaster would like used to index their site's pages. If a
webmaster specifies a preferred domain as
http://www.example.com and Google finds a link to that site that is
formatted as http://example.com, Google will treat that link as if it
were pointing at http://www.example.com.
URL Parameters - You can indicate to Google information about
each parameter on your site, such as "sort=price" and
"sessionid=2". This helps Google crawl your site more eciently.
Crawl Rate - The crawl rate aects the speed (but not the
frequency) of Googlebot's requests during the crawl process.
Malware - Google will inform you if it has found any malware on
your site. Malware creates a bad user experience, and hurts your
rankings.
Crawl Errors - If Googlebot encounters significant errors while
crawling your site, such as 404s, it will report these.
HTML Suggestions - Google looks for search engine-unfriendly
HTML elements such as issues with meta descriptions and title
tags.
Sign Up
Site Configuration
This important section allows you to submit sitemaps, test
robots.txt files, adjust sitelinks, and submit change of address
requests when you move your website from one domain to
another. This area also contains the Settings and URL parameters
sections discussed in the previous column.
+1 Metrics
When users share your content on Google+ with the +1 button,
this activity is often annotated in search results. Watch this
illuminating video on Google+ to understand why this is important.
In this section, Google Webmaster Tools reports the eect of +1
sharing on your site's performance in search results.
Labs
The Labs section of Webmaster Tools contains reports that Google
considers still in the experimental stage, but which can
nonethelsss be useful to webmasters. One of the most important
of these reports is Site Performance, which indicates how fast or
slow your site loads for visitors.
Key Features
Sites Overview - This interface provides a single overview of all
your websites' performance in Bing powered search results.
Metrics at a glance include clicks, impressions, pages indexed,
and number of pages crawled for each site.
Crawl Stats - Here you can view reports on how many pages of
your site Bing has crawled and discover any errors encountered.
Like Google Webmaster Tools, you can also submit sitemaps to
help Bing to discover and prioritize your content.
Index - This section allows webmasters to view and help control
how Bing indexes their web pages. Again, similar to settings in
Google Webmaster Tools, here you can explore how your content
is organized within Bing, submit URLs, remove URLs from search
results, explore inbound links, and adjust parameter settings.
Trac - The trac summary in Bing Webmaster Center reports
impressions and click-through data by combining data from both
Bing and Yahoo! search results. Reports here show average
position as well as cost estimates if you were to buy ads targeting
each keyword.
Sign Up
Features
Identify Powerful Links - Open Site Explorer sorts all of your
inbound links by their metrics that help you determine which links
are most important.
Find the Strongest Linking Domains - This tool shows you the
strongest domains linking to your domain.
Analyze Link Anchor Text Distribution - Open Site Explorer
shows you the distribution of the text people used when linking to
you.
Head to Head Comparison View - This feature allows you to
compare two websites to see why one is outranking the other.
Social Share Metrics - Measure Facebook Shares, Likes, Tweets,
and +1's for any URL.
Learn more
Search engines have only recently started providing better tools to help webmasters improve their
search results. This is a big step forward in SEO and the webmaster/search engine relationship.
That said, the engines can only go so far to help webmasters. It is true today, and will likely be true
in the future, that the ultimate responsibility for SEO lies with marketers and webmasters.
It is for this reason that learning SEO for yourself is so important.
Meta Tags
Keyword Stuffing
Once upon a time, meta tags (in particular, the meta keywords tag)
were an important part of the SEO process. You would include the
keywords you wanted your site to rank for, and when users typed
in those terms, your page could come up in a query. This process
was quickly spammed to death, and was eventually dropped by all
the major engines as an important ranking signal.
Ever see a page that just looks spammy? Perhaps something like:
Other tags, in particular the title tag and meta description tag
(covered previously in this guide), are crucial for quality SEO.
Additionally, the meta robots tag is an important tool for controlling
crawler access. So, while understanding the functions of meta tags
is important, they're no longer the central focus of SEO.
At Google, advertisers spending tens of millions of dollars each month have noted that even they
cannot get special access or consideration from the search quality or web spam teams. So long
as the search engines maintain this separation, the notion that paid search bolsters organic
results should remain a myth.
2. Smarter Engines
Search engines have done a remarkable job identifying scalable, intelligent
methodologies for fighting spam manipulation, making it dramatically more
difficult to adversely affect their intended algorithms. Metrics like Moz's TrustRank,
statistical analysis, and historical data, have all driven down the value of search
spam and made white hat SEO tactics (those that don't violate the search engines'
guidelines) far more attractive.
More recently, Google's Panda update introduced sophisticated machine learning
algorithms to combat spam and other low-value pages, and the search engines
continue to innovate and raise the bar for delivering quality results.
We obviously don't recommend employing spam tactics. But to assist the large
number of SEOs who seek help when their sites get penalized, banned, or flagged, it
is worthwhile to review some of the factors the engines use to identify spam. For
additional details about spam from the engines, see Google's Webmaster Guidelines
and Bing's Webmaster FAQs (PDF).
The important thing to remember is this: manipulative techniques generally won't
help you, and they often result in search engines imposing penalties on your site.
Keyword Stuffing
One of the most obvious and unfortunate spamming techniques, keyword stung, involves
littering keyword terms or phrases repetitively on a page in order to make it appear more
relevant to the search engines. As discussed above, this strategy is almost certainly
ineectual.
Scanning a page for stued keywords is not terribly challenging, and the engines' algorithms
are all up to the task. You can read more about this practice, and Google's views on the
subject, in a blog post from the head of their web spam team: SEO Tip: Avoid Keyword
Stung.
Manipulative Linking
One of the most popular forms of web spam, manipulative link acquisition, attempts to exploit
the search engines' use of link popularity in their ranking algorithms to artificially improve
visibility. This is one of the most dicult forms of spamming for the search engines to
overcome because it can come in so many forms. A few of the many ways manipulative links
can appear include:
Reciprocal link exchange programs: Sites create link pages that point back and forth to one
another in an attempt to inflate link popularity. The engines are very good at spotting and
devaluing these as they fit a very particular pattern.
Link schemes: These include "link farms" and "link networks" where fake or low-value
websites are built or maintained purely as link sources to artificially inflate popularity. The
engines combat these by detecting connections between site registrations, link overlap, and
other methods targeted at common link scheme tactics.
Paid links: Those seeking to earn higher rankings buy links from sites and pages willing to
place a link in exchange for money. These sometimes evolve into larger networks of link
buyers and sellers, and although the engines work hard to stop them (Google in particular has
taken dramatic actions), they persist in providing value to many buyers and sellers (more on
that perspective).
Low quality directory links: These are a frequent source of manipulation for many in the SEO
field. A large number of pay-for-placement web directories exist to serve this market and pass
themselves o as legitimate, with varying degrees of success. Google often takes action
against these sites by removing the PageRank score from the toolbar (or reducing it
dramatically), but won't do this in all cases.
There are many more manipulative link building tactics that the search engines have identified.
In most cases, they have found algorithmic methods for reducing their impact. As new spam
systems emerge, engineers will continue to fight them with targeted algorithms, human
reviews, and the collection of spam reports from webmasters and SEOs.
Cloaking
A basic tenet of search engine guidelines is to show the same content to the engine's crawlers
that you'd show to a human visitor. This means, among other things, not to hide text in the
HTML code of your website that a normal visitor can't see.
When this guideline is broken, the engines call it "cloaking" and take action to prevent these
pages from ranking in their results. Cloaking can be accomplished in any number of ways and
for a variety of reasons, both positive and negative. In some cases, the engines may let
practices that are technically cloaking pass because they contribute to a positive user
experience. For more on the subject of cloaking and the levels of risk associated with various
tactics, see our article on White Hat Cloaking.
filtered types of pages are thin aliate content, duplicate content, and dynamically-generated
content pages that provide very little unique text or value. The engines are against including
these pages and use a variety of content and link analysis algorithms to screen out low value
pages.
Google's 2011 Panda update took aggressive steps to reduce low quality content across the
web, and Google continues to iterate on this process.
Linking Practices
Just as with individual pages, the engines can monitor the kinds of links and quality of
referrals sent to a website. Sites that are clearly engaging in the manipulative activities
described above on a consistent or seriously impacting way may see their search trac suer,
or even have their sites banned from the index. You can read about some examples of this
from past posts, including Widgetbait Gone Wild and the more recent coverage of the JC
Penney Google penalty.
Trustworthiness
Content Value
Once youve ruled out the list below, follow the flowchart beneath for more specific
advice.
Errors
Errors on your site that may have inhibited or prevented crawling. Google's Webmaster Tools
is a good, free place to start.
Changes
Changes to your site or pages that may have changed the way search engines view your
content. (on-page changes, internal link structure changes, content moves, etc.).
Similarity
Check for sites that share similar backlink profiles, and see if theyve also lost rankings. When
the engines update ranking algorithms, link valuation and importance can shift, causing
ranking movements.
Duplicate Content
Modern websites are rife with duplicate content problems, especially when they scale to large
size. Check out this post on duplicate content to identify common problems.
While this charts process wont work for every situation, the logic has proven reliable in helping us identify spam penalties and
mistaken flagging for spam by the engines, and separating those from basic ranking drops. This page from Google (and the
embedded YouTube video) may also provide value on this topic.
teams.
your request.
backlog.
Be aware that with the search engines, lifting a penalty is not their obligation or responsibility.
Legally, they have the right to include or reject any site or page. Inclusion is a privilege, not a right;
be cautious and don't apply SEO techniques that you're skeptical about, or you might find yourself
in a rough spot.
They say that if you can measure it, then you can improve it. In search
engine optimization, measurement is critical to success. Professional SEOs
track data about rankings, referrals, links, and more to help analyze their
SEO strategy and create road maps for success.
Analytics Software
The right rools for the job
Moz Analytics
Omniture
Fireclick
Mint
Sawmill Analytics
Clicktale
Coremetrics
Unica NetInsight
Additional Reading:
How to Choose a Web Analytics Tool: A Radical Alternative
from Avinash Kaushik way back in 2006 (but still a relevant and
quality piece)
No matter which analytics software you select, we strongly recommend testing different versions of
pages on your site and making conversion rate improvements based on the results. Testing pages
on your site can be as simple as using a free tool to test two versions of a page header or as complex
as using an expensive multivariate software to simultaneously test hundreds of variants of a page.
Google Trends
At google.com/trends, you can research keyword search volume
and popularity over time. Log in to your Google account to get
richer data, including specific numbers instead of simple trend
lines.
Bing IP Query
Fluctuation
In search engine page and link count numbers
The numbers reported in "site:" and "link:" queries are rarely
precise, so we caution against getting too worried about large
fluctuations unless they are accompanied by trac drops. For
example, on any given day, Yahoo! reports between 800,000 and
2,000,000 links to the moz.com domain. Obviously, we don't gain
or lose hundreds of thousands of links each day, so this variability
provides little guidance about our actual link growth or shrinkage.
If drops in links or pages indexed coincide with trac referral
drops from the search engines, you may be experiencing a real
loss of link juice (check to see if important links that were
previously sending trac/rankings boosts still exist) or a loss of
indexation due to penalties, hacking, or malware. A thorough
analysis using your own web analytics and Google's Webmaster
Tools can help to identify potential problems.
Falling
Search traffic from a single engine
1.
2.
3.
Falling
Search traffic from multiple engines
Chances are good that you've done something on your site to block crawlers or stop
indexation. This could be something in the robots.txt or meta robots tags, a problem with
hosting/uptime, a DNS resolution issue, or a number of other technical breakdowns. Talk to
your system administrator, developers, or hosting provider, and carefully review your
Webmaster Tools accounts and analytics to help determine potential causes.
Individual
Ranking fluctuations
Gaining or losing rankings for one or more terms or phrases happens millions of times a day
to millions of pages and is generally nothing to be concerned about. Ranking algorithms
fluctuate, competitors gain and lose links (and on-page optimization tactics), and search
engines even flux between indexes (and may sometimes even make mistakes in their
crawling, inclusion, or ranking processes). When a dramatic rankings decrease occurs, you
might want to carefully review on-page elements for any signs of over-optimization or violation
of guidelines (cloaking, keyword stung, etc.) and check to see if links have recently been
gained or lost. Note that with sudden spikes in rankings for new content, a temporary period
of high visibility followed by a dramatic drop is common; in the SEO field, we refer to this as
the "freshness boost".
Positive
Increases in link metrics without rankings increases
Many site owners assume that when they've done some classic SEOon-page optimization,
link acquisition, etc.they can expect instant results. This, sadly, is not the case. Particularly
for new sites, pages, and content that have heavy competition, ranking improvements take
time. Even earning lots of great links is not a sure recipe to instantly reach the top. Remember
that the engines not only have to crawl all those pages where you've acquired links, but also
index and process them. So the metrics and rankings you're seeking may be days or even
weeks behind the progress you've made.
Contributors
We would like to extend a very heartfelt thank you to all of the people who contributed to this guide, including Urban Influence, Linda Jenkinson, Tom
Critchlow, Will Critchlow, Dr. Pete, Hamlet Batista, chuckallied, lorisa, Optomo, identity, Pat Sexton, SeoCatfish, David LaFerney, Kimber, g1smd, Steph
Woods, robbothan, RandyP, bookworm seo, Rafi Kaufman, Sam Niccolls, Danny Dover, Cyrus Shepard, Sha Menz, Casey Henry, Lisa Wildwood, Jeremy
Modjeska, and Rand Fishkin.