Session Intro: Provides a checklist and workflow for diagnosing your
web sites for SEO obstacles using freely available diagnostic tools.
Vanessa Fox, Features Editor at
Search Engine Land is moderating this session and speakers include David
Golightly, User Experience Developer Lead at
Zillow, Jonathan Hochman, President of
Hochman Consultants and Chris Silver
Smith, Lead Strategist at
GravityStream / Netconcepts.
Vanessa starts off with the question of "what really matters?" Accessibility,
discoverability and conversions are the big stuff that matters. The place to
really start in investigating potential problems is the search results
themselves. Is problem related to indexing, ranking or crawling? Identify
problem first.
Chris is then introduced as first speaker. Diagnosing problems involves a wide
range of criteria. Most issues are basic and easy to diagnose - things like mis-using
robots.txt tag, inadvertently blocking spiders and the like. First question to
ask is "are pages actually indexed?" If not, there is a problem! Does site URLs
have sessionIDs and if so, are they doing anything to resolve that? Google
Webmaster Central is useful to get a bird's eye view of what Googlebot sees when
visiting your site. The title tag and meta tag data they show can be quite
revealing and interesting.
Chris uses www.web-snifer.net to check
server header status codes to ensure pages are reporting the proper codes. He
uses www.seebot.org to view web pages like a
search engine would. You can also use a Firefox Developer toolbar as well.
Firefox Link Counter Extension provides useful link data. Chris makes reference
to SEOmoz's SEO Toolbox as a good source of tools for diagnosing problems. One
such tool shows other sites that are resting on your IP address. Use Google Sets
to ID your competitors. You can also see sites that Google thinks are related to
yours.
Jonathan is up next who will dive into some diagnostic stuff. Using a NoScripts
add-on, Jonathan can turn scripts off and discover problems. He shows a problem
with the SMX site when scripts are turned off as well as Gillette. He recommends
the Googlebar (not Toolbar) which is a Firefox plug-in. It has a one-click
button that shows you Google's cache. Another Firefox add-on he mentions is Live
HTTP Headers which shows header status codes.
With regards to rich media applications, you need to be able to feed the bots
content they can understand. Replace html content with rich media content by
manipulating Document Object Model (DOM). For Silverlight, create SEO-friendly
insertion code or better yet, bug Microsoft to provide a solution.
Xenu's Link Sleuth will crawl href links just as a search engine bot would. It
makes it easy to find and fix broken links. It will also help you to create a
site map. Firefox Web Developer Add-on has multiple functions that are valuable.
Watch out for problems with frames, iframes, Flash and Silverlight. Each object
is treated as a separate thing and not as a part of the host page. Ajax as well
can be problematic and it may use iFrames frequently.
Finally David is up to show some problems they found at Zillow. One problem they
had with old database is that it was not highly configurable for multiple data
sets. They also wanted it to be responsive to a wide range of user options. One
problem they had was that many of the functions Zillow had were dependant on
Java Script. With JS turned off, users (and search engines) were not able to use
site. As of 2/08 only 200,000 out of 80,000,000 homes were actually indexed.
They also did not rank well.
They also improved navigation to help bots find pages. Using breadcrumbs, they
help bots find very interior pages. Ajax on the top, not the bottom. In other
words, AJAX should be built on top of functioning web page and not other way
around. SEO should work in concert with great UX.
Q&A:
Here is a recap of "some" (not all) of the questions that were asked and
answered.
- Is there any automated tools that check to see that redirects are working
correctly after they have been set up?
Jonathan recommends Xenu's Link Sleuth and Vanessa recommends using Google Webmaster Central.
- What are nofollows for?
Chris answers to control flow of PageRank (PageRank sculpting) and to stop flow of link juice to undesirable pages. Also to control blog comment spam.
- Does private WHOIS put you at disadvantage?
Chris says probably not. Vanessa says so long as you are not spamming, should not be a problem.
- To optimize or not optimize pop-ups?
Do you want it as separate page? It may not be good entry page. If JavaScript, is not going to be indexable anyway.
Session coverage by David Wallace - CEO and Founder SearchRank.