page header image


Featured Article


SEO – The Death of Keywords – Part 1
Are their days numbered? 

The series of 2011 Panda updates by Google enhanced the discrimination of its search engine technology. This was done in an effort to enhance the quality of results returned for users’ searches by filtering out low quality sites, regurgitated rubbish, scraped content and wordy but ultimately low quality content, from the users’ perspectives. The ultimate objective was, of course, to ensure that the user community maintained a high regard for the Google search enterprise, and its associated advertising business model. As an aside of course, this also means that Google is reaching deeply into knowledge of individual users’ browsing habits – not just superficially as had been the case.

 The Panda enhancements reportedly included: assessment of user residence time on a site and their inter-page navigation on that site; users’ readiness to share the results with friends and colleagues directly, and through social networking; their preparedness to comment (on Web 2.0 sites).

Some time ago I wrote on whether Google was a neural network (insert reference). It’s not there yet, but the learning capability is increasing in leaps and bounds.

Whilst the emphasis of the model appears to have shifted from content-analytical to consumer-analytical, I cannot believe that the lexical analysis routines which Google has employed for several years now, have been dumped.

I believe that Google still harbours a vision of being able to take a piece of content and assess its value.

Now ‘value’ is a subjective measure with no inherent scale – it is a relative concept. But what if the content was ‘measured’ relative to, say, Wikipedia or other relevant authority site?

If Wikipedia (or equivalent) became the ultimate reference baseline, then what is the point of anyone writing anything else (other than to disagree or to extend the content)? Of course, it would be to venture an opinion or disagree with Wikipedia.

Anyway, to get to my main point: Up until 2011, the idea of keywords has worked to both Google’s and commercial interests. For Google it has meant that relatively accurate search results could be obtained, whilst for content creators, the concept has been, mostly, profitable.

As the lexical analysis technology proceeds however (and companies such as Booklamp have irons in this fire, too), then there will surely come a time when the concept of ‘keywords’ would be meaningless. Google will decide for itself what the content is about, and measure its ‘absolute’ quality against Wiki-likes, supplemented by the user community grading – both automated and invisible by Google itself, and by direct user engagement in a site.

Fine, you say, but that’s not the whole story! No, it’s not. That’s why this is Part 1.

The next articles deal with:

-> Google Page 2 – Will That Be The New Battleground?
-> Why Adwords Will Have To Change
-> Why Review And Comparison Sites Will Grow…And Grow

© 2011 Phil Marks

Interested in a great thriller with espionage, cyberwarfare and digital invasions? Check out Gate of Tears.

Back to top