Love it or hate it, we’ve all seen it. Flash sites built for beauty and not spiders have often been the source of grief and stress within the SEO community as we try and explain why it was a dumb idea to do it that way. Ya, that’s harsh but so it trying to force a use to guess where they need to mouse over to find a menu.
I have a site I love to use as the perfect example of how flash lets you down. This chocolate company has some of the most yummy looking chocolate treats. Problem is, Google doesn’t think it has much to do with chocolate. Take a look at the Sweet Thoughts chocolate site which looks like this to Google and results in the Cadbury UK site along with many others to outrank it for a search for chocolate. In fact, the spidered flash is there but just not ranking. So the new spidering of flash content by Google should be a RankSaver for sites like this but I would caution everyone for the moment against complacency and encourage adoption of existing best practices.
From July 1st, Google started spidering the content of flash files to help to unlock some of the last places on the web left unlookedat. However, rather than looking normal, these sites are clearly marked with [FLASH] next to the search result. Yahoo! have had the technology made available to them as well but have not yet committed to a timeline to implement this enhanced spidering. Who knows what Live will do – maybe Ms. Dewey will simply take her crop and beat the offending sites in to HTML submission. Or run over them on her motorcycle. Mocking them might work as well
Google’s help files specifically state that In general, search engines are text based. This means that in order to be crawled and indexed, your content needs to be in text format. This doesn’t mean that you can’t include images, Flash files, videos, and other rich media content on your site; it just means that any content you embed in these files should also be available in text format or it won’t be accessible to search engines. So while this is changing for flash, I would still continue to create a mirror site for those with things like old browsers and flash turned off.
A lot of the Flash content has no text basis to it since it has been created out of images and possibly links. This spidering change will have no impact on the ranking of these sites until it combines image recognition with flash spidering. Sometimes, flash files are better left hidden but it may help with the ability of Googlebot to access and possibly rank deeper pages previously inaccessible.
Some flash content has multiple text-based pages buried within the program. So lots of text, no context. If this causes one page to become authoritative for multiple keywords, it could cause issues with ranking. With the chocolate shop example, it could cause all items to be lumped on a single page, diluting a stronger authority if similar products were grouped logically together with unique textual content. So really, to catch chocolate lovers like myself that will want to include a different version of the content which breaks up text through a set of logical, focused pages to help Google rank each page for what it is authoritative for.
While duplicate content will be less of a problem with Flash since it’s all lumped together, the underlying code itself would need to be clearly understood by a robot. So you’ll suddenly need very clean, very well structured and very well written code. Since sometimes code which executes well is not necessarily written well, this may also pose potential spidering issues.
With the potential for so many problems, best practice remains the same. Include alternate, text based information for search engines in order to help them spider, understand and rank your site. Especially if you sell chocolate