Click anywhere to close

Posted: Mar 8, 2015

One step in SEO efforts that is often overlooked is that of comparing what you think your web page looks like against what the search engines see.  There are several scenarios where this is crucial to the success of your organic search growth.

Case 1:  If your site is very heavy Flash based or Javascript based.  Search engines are unable to see what happens through those assets

If your site is very heavy Javascript based or heavy Flash based, you may be disappointed to learn how your site actually gets interpreted by the search engines.  Let's look at an example. is a website that shows off photos and allows people to vote on the ones they like the most.  Here is a screen image of what it looks like to a standard user: Homepage

  However the Google Search engine (and others) interpret the website and see it as the following: Homepage No Javascript

Actually, if you wanted to get technical, the search engine actually does not interpret the colors and spacing... but the above visual example is what I wanted to convey. The above example shows the site running with JavaScrpt disabled.  The problem in this case was that the site relied on JavaScript to display it's primary bulk of information.  Since google does not interpret javascript, Google basically got a blank page to index.  Since there was nothing to index, the page received no traffic. How do you fix this then?  Well there are a couple routes and none of them are easy. The most straight forward way is to display the base information before javascript starts rendering.  If all of you content is retrieved via AJAX, this might mean that you need to have a server side scripting language, such as PHP, prepare the initial batch of data and present it on your site. Another way around this error is to have a secondary page that displays the information needed for search engines.  Then when users land on this page you can redirect them to where they should be going.  I actually don't like (and have stopped recommending this method) because at times the search engines will interpret it as a bait and switch method.  I.E. you are having Google indexing one thing and displaying another. Page caching also can work to your advantage here to combat the javascript indexing issues.   You can have the outputs of a page saved in cache so that when the googlebots crawl your page, they get the cached version of the page.   In order to do this effectively and with a large site, you may need a crawler of your own to automatically update the caches on a regular basis.

Case 2:  If you site has a lot of it's primary content behind a login or members only section.

Another common scenario is when you have specialized content that may be hidden behind a members only area.  The plain and unfortunate truth is that the googlebots do not see this content and cannot index it. One of the best ways to get around this is to offer a preview of each piece of content.  If your site has specialized articles then you may offer the first paragraph or two in an open and free format.  Google really only needs the first ~300 words or so to determine the prominent keywords.  If you do choose this route, determine what the keywords are that you are targeting for each article.  Then ensure that you play those up on your preview pages.  Include the keywords in the title, description, image assets, H1 tags and other spots on your page.
Copyright © 2017 Dimbal Software - All Rights Reserved
Dashboard | Software Modules | Support | Blog | Article Index

Links | NPS | Polls | SEO Scans | Test Plan | Triage