Google Knol – very good but very failing?


Google Knol, the Googley competition for Wikipedia, was announced with some fanfare and really seemed like a great idea.    The ‘knol’ stands for “Knowledge”, and articles are written by people who verify their identities and presumably have some knowledge of the topic.    Community ratings are used to filter good from bad knol posts, presumably leaving the best topical coverage at the top of the knol heap.

However as with many Google innovations outside of pure keyword search knol appears to be making gaining little traction with the internet community.     I say this because I rarely see the sited linked to or referenced by blogs or websites and also from my own knol page for “Beijing” which as the top “Beijing” and “Beijing China”  listing you’d think would have seen fairly big traffic over the past months which included the Beijing Olympics.   Yet in about six months that page has only seen 249 total views – that is less than many of my blog posts would see in just a few days here at Joe Duck.

So what’s up with the decisions people make about using one resource over another?    Like Wikipedia Google Knol is an excellent resource.   Reading my Beijing page, for example, would give you some quick and helpful insights into “must see” attractions there.   It’s no travel guide but it would prove a lot more helpful than many sites that outrank it at Google for the term “Beijing”.    Google appears to have relegated their own knol listings to obscure rankings – perhaps because linkage is very low given the low use of knol.    Like many Google search innovations knol appears bound to the dustbin of obscurity as Wikipedia continues to dominate the rankings for many terms (as they should – it’s generally the best coverage although generally very weak for travel because they fail to capture commercial info adequately).

My simple explanation would be that we are prisoners of habit and have trouble managing the plethora of information resources that lie – literally – at our fingertips.   We all have yet to understand much about how the internet works, and how inadequate a picture one gets if they simply stick to a keyword search and hope for the best.

Steve Fossett ID, cash, jacket found near Mammoth Lakes, California. Hoax or real?


Hikers and searchers from the Mammoth Lakes Calfornia are reporting they have found an FAA ID for Steve Fossett along with a jacket and some crumpled money.     I’m confused as to why this is only coming out now after two days since the Fossett story was international news for many weeks as thousands participated in a huge search for the missing aviator.     He was lost after flying from Nevada on September 3, 2007.

Fossett was declared dead February 15, 2008.   If this find is not a hoax it implies Fossett may have been alive after the crash.   However it seems very odd that a survival expert of Fossett’s caliber could have survived the crash and wandered around leaving only a few items rather than marking a large area with rocks and markings to signal the aircraft he knew would come looking for him.

The Fossett mystery continues.

More from Mammoth Lakes News

Danger DataFossett Flight ?

Matt Cutts from Google


Matt Cutts at the Google Dance
Originally uploaded by JoeDuck

It’s always great to get a chance to talk to Matt Cutts at search conferences though I didn’t have any good complicated search questions to bug him about this year. Matt is one of the early Google folks and arguably the most knowledgeable search expert in the world since he’s one of the few people who knows the Google algorithm inside out. Matt’s actually listed on the key Google search patent.

Today I noticed that Matt’s post about Google Chrome is near the top at Techmeme after some early reports suggested Google was going to nab all the info people created via use of the Chrome browser. Although I do not worry about Google stealing the content I create using their tools I was surprised in the discussion at Matt’s blog to see how people probably do not understand how much of your data from searches, emails, and other online tools is analyzed by search engines, ISPs, and probably at least a few government agencies. I wrote over there:

Well, I’m sure folks like Marshall knew that Google was not out to steal content. What people should be as concerned about is how the Chrome datastream will be processed now and over time, and how open will it be to examination by companies for advertising purposes ? Personally I’m OK with that but I think many people are not, and the lack of transparency in this area bothers me.

Somebody even suggested I was foolish to think they’d use Chrome data to target advertising, to which I replied:

Josh – you are naive to assume Google does so little with the search term data they explicitly say they have the right to collect. In Gmail, for example, some portion of your header is read by Google (probably just the title and not the content) so that ads can be targeted to you on those topics. Google Toolbar collects a lot of information and my understanding this helps target PPC advertisements though I’m not sure about that. As i noted I’m personally OK with this level of snooping, but I believe Google should make it much clearer what they do with the data they collect and probably also have options so users can delete any information they created – including their search streams – as they see fit.

SES, SEO, Blogs


Blogging the conference has been a great way to test some ideas about blog ranking and watch Google struggle to bring the most relevant content into the main search (they’ve done pretty well with blog search, not so well with regular search which will have all the blog content listed in a week or so, basically too late to be all that helpful to users).  More importantly the stuff from *this year* will probably be ranked above the SES San Jose 2009 information that is likely what people using that term are searching for effective next year.   I’d think they could simply increase the value of ‘freshness’ for listings tagged as events related.

I had a nice discussion about this “events” ranking challenge with Jonathan from Google at the party. The problem is that to combat spam Google does not push out blog content immediately, meaning that if you search for “SES San Jose”, especially a month ago or so, you would have been likely to get old, dated content rather than the current SES page you’d normally want to find. This appears related to linking issues (newer has fewer), but also I think the regular engine is allergic to new content, which is why you’ll often find the most relevant Google stuff at the blog search if it’s a topic that is covered heavily by blogs such as SES or CES Las Vegas where I noted the same issues of “stale content” in the main search with “great content” in the blog search.

I remain convinced that some of the challenges faced in ranking could be solved by a combination of more algorithmic transparency from Google combined with greater accountability by publishers who’d agree to provide a lot more information about their companies so that Google can get a good handle on the credibility of the online landscape. This webmaster ID is happening now in several ways but I’d think it could be scaled up to include pretty much everybody publishing anything online (ie if you don’t register you’ll be subjected to higher scrutiny).

More Tech Posts moved here

SES Site Clinics


SES Site Clinics
Originally uploaded by JoeDuck

A good team including Dave Naylor, Greg Boser, Dani Horowitz —- on stage is evaluating some websites that have volunteered to be reviewed:

http://www.LangAntiques.com

Hmm – not sure on the advice here but they are suggesting Lang avoid full URLs to make server handling more efficient, and avoid putting unrelated content at the site (use a different domain) to be more competitive in that specific niche. ie Generalizing can be helpful in some cases but make it harder to rank for very competitive terms.

http://www.GoldCoastInformation.com.au

s-l-o-w is a problem.

www.teleflora.com

Greg: Narrow title tags to be more focused. Focus on a small number of key phrases or words you want to rank well for rather than several.

Dave: Does not like the navigation on home page – thinks it is weak and will possible get site downranked. Title tags suck all the way through. Cross linking is important. Cannot be excessive, but can be used e.g. to link blog and websites advantageously.

Smarthome.com

Debby and Greg: Tables are bad. Use CSS. Separate content from layout. Greg – improve titles. Debby – make title tags *backwards* from the breadcrumbs – ie from most specific to the generic. [e.g. diamond studded shock collars, dog collars, dogs]

Blog advice: One of the most powerful tools you have. Check templates. Avoid the template sponsor links at footers? Edit the post URL to exclude date.

AllergyAsthmaTech.com
Ranking well for Asthma products but no good conversions. Greg – avoid images for important terms – use words or layer image on top of text which is “awesome” according to Dave and Greg. Caveat – be careful overdoing this.

default.asp as home page – avoid multiple home pages.

Question to audience – how many companies use blogs with the site? (very few raise hands). Greg and Dave: Get a blog! Google ranks blogs over product pages.

JustAnswer.com keep bots out of the https. site:JustAnswer.com

Greg – use subdomains but keep in mind search engines are sensitive to host name spam.

SES San Jose Final Day – local search


This is an excellent presentation but I haven’t had enough coffee to take good notes, and unlike some of the other conferences I’ve covered like CES and MIX there isn’t much traffic to the conference posts anyway. Anxious to check other blogs covering this to see if they had much of a traffic spike, which one would really expect from search related news.

I’m lazy so only posting the summary from SES:

Local & More
Special Kelsey Group Presentation: Local 2.0: The Evolution of Local Search
What percentage of online searches are local? If you consider searches that end up having some influence on local buying activity, the opportunity is put into perspective. But there are still large gaps between the point of search and the point of purchase. How are online mapping, shopping engines, and directories starting to fill these gaps with user-generated content, video, or inventory data that funnel searchers towards local businesses? And how can marketers utilize these tools to get local searchers to pick up the phone, schedule appointments, or show up at their stores?

Moderator:
Speakers:

Black Hat / White Hat SEO Session at SES San Jose


Black Hat SEO Session at SES San Jose
Originally uploaded by JoeDuck

The SEO session here at SES San Jose is packed as everybody struggles to get competitive advantage, though I’m concerned that even some fairly advanced SEO folks simply don’t understand how seriously they may damage their clients or sites with a number of techniques that are no longer tolerated by Google.

Jill’s making an excellent point about “incompetent SEO” who, in her words…suck. I think she’s right that there are a lot of very bad folks doing SEO – I’d suggest about 90% or more.

Boser – the sites that rank are the ones that put in the effort. He says that you could generally replace the top sites with any others wihout bothering users.

Boser: buying links remains a key tactic for competive markets.

Naylor: Common mistakes are from within the organization where their link buying goes over the thresholds. Good black hats “know the threshholds” at which the search engine will identify link buying problems.

Disagreement (hey, cool) Jill’s Whalen suggesting good SEO is simply common sense, Todd Friesen suggesting it’s not. Jill says it’s not about tricking the engines.

Boser: Must compete now, so using paid links to get the clients going is both acceptable and needed.


Widget marketing – ?

SEO Champion question: He’s angry but not clear why – some conflict with Greg Boser over SEO stuff.

Boser: Links are about blending in to your niche. Hotels dominated by aggressive search spam for years – a tough industry. Hard in those space to follow the rules and compete.

Bruce Clay: Doing things that are way out of bounds is much riskier for established sites. You can’t afford to burn your own house so consider these risks.

BMW: Burned themselves buying links for short time. Greg – less consequence for the big players. WordPress also got a hand slap for link abuses. [ yes, but if we use a “user quality” metric we’d expect big players to have fewer consequence rather than severe punishment].

Matt Cutts from Google (audience). We take action on a lot of big sites. Some panel argument with Matt here about how fairly Google applies the rules. Boser: Forbes is spamming – why no action?

Naylor: Legal site in UK. Restoration involved identifying the paid links that a company representative had purchased. Google found them in 4 months and banned the site.

Audience question: What about the user – isn’t their interest the best definition of white vs black hat? Todd may not understand what she means here, he’s noting that outright deception is out. But I think her question is more nuanced and the answer is generally yes. Naylor’s correctly noting that users may not care much if they get site A or site B, and that is often true.

Made for Adsense sites: Boser – it’s crap, and includes a lot of blogging content. [This topic is so complicated – I’m always amazed how everybody thinks that they know crappy from good content. I think you need to ask communities about what they want, and trust their judgement. I think Google is moving in that direction and it’s a good thing.  What if somebody has a fabulous site, better than competition, that is made for adsense?   Where is the line?]  Answer – community judgement.

SES Session Description:

Searcher Track
Black Hat, White Hat: Playing Dirty with SEO

Some say that “black hat” search marketers will do anything to gain a top ranking and others argue that even “white hat” marketers who embrace ethical search engine optimization practices are ultimately trying to game the search ranking system. Are white hats being naive? Are black hats failing to see the long-term picture? This session will include an exploration of the latest black and white issues, with lots of time for dialog and discussion.

Moderator:
Matthew Bailey, President, SiteLogic

Speakers:
Greg Boser, President, WebGuerrilla LLC

SES San Jose – SEO Rehab and Intervention


The crowd is starting to pour in now from the big SES car giveaway. I think they gave away three of them, but I was no eligible as a “press” person. Just one more huge sacrifice I have to make for the good of journalism, the internet, and global warming.

Andy has started out with a very generic a presentation, now noting that they no longer take English speaking clients. Europe was easier because so few there were doing SEO. [note to self – get the Mandarin lessons going ASAP]

Greg Boser – excellent point about importance of experimenting with different tools and techniques – you can’t necessarily know which may have positive impacts.

Dave – good points about worrying more about where things are going than your daily pagerank numbers which is only updating every 3 months anyway.

[Interesting. All these guys appear to be a lot less focused on the old style things people tend to worry about like keyword density and pagerank. I think the business leaders are moving strongly to PPC even as social media is making organic optimizing a whole new ballgame. Where are all the social networking / blogging / media optimization strategies?]

OK, not much SEO here, but the Twittering was really interesting. It would be great to integrate messaging into the session so people in and out of audience could ask questions – sort of a hybrid virtual / real space / real time environment…

————————-

Searcher Track
SEO Rehab & Intervention
It’s 2009 and rehab is en vogue from Beverly Hills to Talladega. This session will include 12 steps to finding harmony in a search-dominated world. Confessions from leading experts include how to get off PageRank, vanity news alerts, inbound link and keyword density analysis, and 301’s.

Moderator:
  • Anne Kennedy, Managing Partner & Founder, Beyond Ink
    (Kevin Ryan standing in for her)
Speakers:

SES San Jose – click fraud session


The intro talk by Tom Cuthbert is listing click fraud numbers that, at over 16% for overall and over 25% for content PPC, are dramatically higher than the few percent normally cited by Google. I’m anxious to hear Shuman’s take on this. Excellent slides…. will try to link them up later.

Erin Sheedy-Owen from Yahoo on catching fraudulent clicks. “We err on the side of the advertiser”. You’ll see these in your logs but won’t be charged for them.

Outright fraud vs Low converting vs unwanted clicks. 12-15% of the clicks are filtered and advertiser not charged. Large recent rise in bot fraud. (this was also noted by Cuthbert).

Yahoo goal is to respond to click fraud reports from advertisers within 10 days. Erin’s making the case that advertiser feedback is very important.

Waiting for Internet Advertising Bureau guidelines.

Deborah from Outrider, a Search Marketing agency managing 1.3 billion in search advertising.   [wow – wouldn’t this be approaching 10% of the market if it is annual? – I’m skeptical of this number’s relevance – maybe this is over a long period of time, though still….an impressive data set].

Client opinions vary from huge concern to little.

ClickForensics is their click fraud application.   Interesting – she just mentioned parked domains as a click fraud issue.  Traffic from a highly relevant parked domain name would probably be good.

Matt Greitzer from Razorfish: They manage 300 million per year in paid clicks.

Virtuous cycle of quality clicks [hmmm — IMHO the optimal revenue model for Google and Yahoo is probably not optimal for advertisers.
I don’t think folks understand this well yet and for agencies overbids are money in the bank too, but standards are coming too slowly because search profits will go down as standards go up.   That said, I have a lot of faith in the next speaker’s sincerity ]

Shuman Ghosemajumder from Google:

Google’s Proactive approach:   Filters – automated.  Invalid click reports available to check these out.    Less than 10% filtered this way.

Offline analyis – leads to credits to account.    Click quality adjustments.  Statistical anamolies.

Finally, reactive approach involves investigation and report.  Very, very tiny.  <.02% handled this way.

[yikes – so why were Cuthbert’s numbers so much higher?]

Smart pricing:  Google gets same ROI by adjusting cost per click according to conversion metrics – ie lower performing publishers command less PPC. [But how do they measure the conversions?]

Google Placement reports and other performanc metrics allow you to track your own campaigns with great precision.

Google competes on basis of ROI, so their incentive is to keep it high and kill fraud.