Why Blogs are better than Google.


Today, as I searched for some breaking news and technology insights, I was struck by how much better informed you tend to be after reading a few blogs targeted to a topic (and following related links and sites and ask questions) than when you simply search Google (or Yahoo or MSN or, if you enjoyed the silly and short lived TV campaign, ASK).    

Don’t get me wrong – I like the search engines and I love the way you can quickly winnow through billions of pages down to the handful that are relevant and good for your topic.     But I’m noticing how increasingly I wind up turning to blogs *first* for the best news, links, and insight.    I’m beginning to understand why I’m doing that, and why it’s a big deal.

There are the obvious advantages to blogs over websites.   They are fresher (ie recent and new content) – especially compared to Google searches that often yield so much old content.   They usually offer some community components so you feel like you are “where the action is” on topics.   This is usually true for major blogs.  TechCrunch is a key watering hole for startups, HuffingtonPost.com for liberal political folks, etc.

However these advantages are secondary to the fact that as blogs mature they offer an excellent “human powered search engine” for your niche of interest, and as we all know humans still beat out computers in terms of understanding what information is most relevant to our inquiry when it is a broad field of interest.

Again, the TechCrunch Technology blog is a great example of this.  A search in Google for “startups” or “technology news” or “venture capital” will give some good results, but even a careful study of those results won’t give you nearly the insight you’ll get from a one hour session at TechCrunch.     Even a Silicon Valley startup new arrival – or distant silicon startup wannabe, could sound like a veteran if they simply kept up with the parade of posts from Mike Arrington and his clever crowd at TechCrunch.

I think this blog advantage breaks down as you move into very specific topics, but it’s going way up as an advantage in the study of general topics as blogs explode and gather traction and community.     Of course there are caveats to this.   Learning in any form takes time, and you would never simply stuble into a blog about a topic without checking other blogs and sites related to that.   But my point is that once you find “the key blogs” about a topic, even if it is a contentious one, you’ll find through those blogs links, references, breaking news, and a community of other interested parties.   This complex, interactive, cross referenced community experience is how humans learn best, and the internet is making that type of learning exponentially easier to obtain.     

San Jose Mercury News – A Cautionary Tale from Business Week


There is a great summary at Business Week of the  remarkable rise and pending fall of Silicon Valley’s newspaper – the San Jose Mercury News.     They note that in many ways the Mercury News saw it all coming, but still failed to position itself to profit from the migration of offline info to online info.  

Although the article does not make this point, to me the failure supports the idea that paradigm shifts do not come from old systems evolving into new ones even when the old systems “get it”, rather they come from new folks thinking out of the old boxes and building the next generation of innovative solutions basically from scratch.  

Obviously new technology rests on the shoulders of old technology, but it seems reasonable to assume that the next big things are not going to come from the previous big things, they are going to spring up from the harsh, quirky, and shifting sands of technology and innovation.     I would suggest that IBM might be an exception to this notion but clearly Microsoft, then Yahoo and Google, now YouTube, Myspace and Facebook all fit this model of major changes coming more from scratch than from a slow simmering of existing ideas.     This also helps explain the challenges of Venture Capitalism in finding “the next big thing”, which may right now only be known by the glimmer in a college kid’s eye.

If so, who is next?

TechMeme Secrets


TechMeme has rapidly become one of the key techno watering holes in the blogosphere thanks to how it helps sift through tech blogs and posts to winnow out those getting maximum buzz.     Todd recently suggested he thinks a lot of SEO firms  are gaming techmeme, especially by post plants from A list bloggers – the implication is that they are paid for this.

He’s largely wrong about this and has given *way* too much credence to the always inflated claims of SEO companies (his inspiration for the post was a small SEO firm in Honolulu).

Of course Tech blogging, and most of the web for that matter, and much of the offline world, have been “damaged” with respect to objective quality content by various tactics that come about as the inevitable result of content monetizing.

But take a look at the prominent TechMeme posts tonight – it’s clear that these are generally spawned from sincere interests and not “planted” as part of advanced SEO tactics. Do any plants happen? A few, but in SEO you have to balance the chance you’ll “sneak in” a good plant against the greater chance that you’ll permanently tarnish the blogger’s reputation cause a scandal (Wal Mart’s Edelman fiasco), or simply spend a lot of time and money for a marginal result. The best SEO strategies rely more than ever on getting legitimate content and placements.

Now, Robert Scoble has a great video post today that is a lot more interesting because he’s trying to reverse engineer TechMeme, something a lot of people in tech are interested in for several reasons.   Robert also manages to feed the new little Scoble during his impromptu 2am advanced tech blogging lecture, which is really a fun statement about how far social networking and life/work integration has come in the past few years.

I hope Gabe responds to Robert to clear up some confusion though he may want to keep the TechMeme algorithm top secret, following in Google’s footsteps.

Some key points by Robert as he speculated about the TechMeme algorithm:
Tech blog database of perhaps 10,000 blogs.
Blog rankings (see TechMeme leaderboard) used to reflect their authority and thus “weight” the power of outbound links from  those blogs.
Reciprocal linking is not as heavily weighted as one directional outbound linking.

Robert suggests an experiment to test some of his ideas and I hope he does it, though Gabe may simply shut down that post or (if he wants to mess with his TechMeme folks) manually override the algorithm so it does funny things that lead to wrong conclusions.  Scoble’s Breeeport experiment was fun a few years ago, and this stuff can be a great way to bring more transparency to the mysteries of content ranking.

China redirects searches to Baidu? OR NOT!


TechCrunch is reporting today that China is redirecting internet searches from Google, Yahoo, and MSN and I assume all other engines – to Chinese search engine Baidu.   However I can’t find anything but little anecdotal posts to support this.   Looks to me like some videos and blogs have been affected, but that the big search engine issues may have related to a temporarily problem or testing of DNS stuff.

They suggest this may relate to the recent award given to the Dalai Lama I’d guess China is spending a lot of time thinking and experimenting with ways to maximize their search revenues, and this redirection, if it really did happen as dramatically as some suggest, would probably be testing ways to gather data on how well Baidu monetizes search compared to the agreements they have with other players.

Wait – here’s a blogger in Beijing, China saying that he’s getting to places TechCrunch says have been sent to Baidu, like Google.   Not sure what’s up …

Is this a false alarm?   I think so, though it might be another example of how China’s centralized socialist economy can create power and monopoly conditions the most ruthless old style US capitalists could only dream about.    Increasingly control of the online landscape is control of the business landscape, and as China’s massive economic expansion continues it will be very interesting to see how the China wields her power.

Note – I just edited this post quite a bit thanks to the new info.  Still dunno what’s going on.

Google and Wikipedia combine to “bomb” NYC.


Update – below was “fixed” with Wiki’s correction and Google’s refreshed index. Looks like the bogus snippet lasted about 1-2 days at Google – probably even less at Wikipedia because they have people reviewing the edits.
————-
Search credibility is still a challenge for Google and Wikipedia as today’s second result for the query “New York City” indicates:

————

New York City– Wikipedia, the free encyclopedia

new york city has just been hit with a nuclear bomb and it has destroyed half of the cityand has left thousands dead. george bush says the people involved
en.wikipedia.org/wiki/New_York_City – 315k – CachedSimilar pages

——— 

This is a very clear example of the challenges of information systems that have no human intervention in the routine editing process (Google) , or have defective human intervention (Wikipedia).    What happened here was a malicious change of the NY City page at Wikipedia followed by Google’s spidering of the bogus content.   I’m hardly a naive user but during my search tonight for NYC info I did a double take on this Google query result and quickly had to reason out that it was bogus.    Wikipedia’s been fixed and this will probably go away within days when Google refreshes it’s listing, but you can sure see how things can get out of hand fast online. 

A recent study suggested Wikipedia and Brittanica were about equally authoritative, and I do think this is an exception to the normal super quality at Wikipedia.

TechMeme on … TechMeme


Gabe Rivera’s TechMeme is a favorite info destination for thousands of technology enthusiasts and news junkies all over the world.   However Bobbie Johnson over at the Guardian  has taken a mini-swipe at TechMeme, suggesting that the relative low traffic from the site shows that people are overrating TechMeme’s importance in the scheme of things technological.    

Robert Scoble, in response, has a  great  summary of how he gets very different traffic depending on the source.     Although he does not focus on the *topic*, clearly that matters a lot as well.    Scoble’s blog is influential enough that it would often send more traffic to a linked site than TechMeme.     I’ll have to check my own stats to be sure but I the times I’ve had links from “A list” bloggers like Robert Scoble, Jeremy Zawodny or Matt Cutts  it has sent more traffic than my frequent links at TechMeme – though I have never had a post be a “headliner” at TechMeme.    In fact I don’t think Gabe’s algorithm would allow his “second tier” sites to have featured posts.   My (wild) guess is that TechMeme has at least two lists of blogs/sites, and only sites and blogs on his top tier list can have the posts featured prominently – others are relegated to the comments section even if it’s a more detailed, more linked, or better post.  

I think this “small stable of premier technology sites” may be a potential defect at TechMeme that keeps the Tech echo chamber very loud but not very diverse, though Gabe may have learned that this helps keep irrelevant posts out of the mix.     At SES I was talking with Matt Cutts about how the TechMeme algorithm might work, and how it might be applied to a broader set of blogs as a ranking mechanism.  

Nick Carr jumped on this TechMeme traffic bashing bandwagon suggesting “juicelessness” which is a  cleverly coined phrase but  misleading because TechMeme clearly reaches a lot of key technology folks and that’s a juicier kind of audience in terms of advertising and influence than, say, 16 year old Diggers.  

So, I’m sticking with TechMeme and redict it’ll get bigger and better. 

Social media frenzy may kill high quality content. Somebody fix this!


The news last month that Microsoft may wind up offering Facebook $500,000,000 for a 5% stake is great news … for the tiny number of Facebook insiders who stand to gain from this move which would effectively value the social media giant at about $10,000,000,000.    For the millions of Facebook folks like me who provide the content and faces that drive Facebook it means … um … more advertising.   

Gee, thanks Facebook.   

When people wake up they may start to realize that we’ve got a potential crisis as small numbers of “info intermediators” like Google and Facebook scoop up the lion’s share of the online ocean of cash while the “info creators” are distinctly second class citizens in the big show.   Small time web publishers and mom and pops are in this group.  So are major newspapers like the New York Times and Washington Post and most other print outlets who tend to make relatively little online despite offering much of the web’s best content to date, especially now that the foolish paywalls of some newspaper outlets like NYT are coming down.   Having no paywall will allow them to make more, but it’s not clear to me they’ll make enough to keep all that high quality content coming.  

Print and newspapers are  hurting and that is going to continue.   That’s OK as long as websites and blogs continue to provide great insight and breaking news, but it’s about time the big players in the online world start working *a lot harder* to feed the hands that are feeding them.  It’s about time they realize that the best web ecosystem encourages high quality content and not just socializing for the sake of hanging online with friends.

Yes, it is true that revenue sharing programs like Google adsense give publishers a nice share of revenues that come directly from activity at their websites.  However lost in this debate is the fact that *most* of Google’s money  (and virtually all of Myspaces), goes into the pocket of Google and Fox (owners of Myspace).   This is because most of the cash comes from searches done at Google.com rather than publishing affiliate sites, and Google keeps all that despite the fact it’s generated *indirectly* from the ocean of content Google has categorized.  Sure Google should make *a lot* from categorizing *your content* so effectively, but should they make 100%?   You can argue this arrangement is fine if the big players turn around and do things with that money that make the internet ecosystem thrive and grow in ways it could not without their involvement.  I think that argument was far more valid a few years ago than it is now.  Literally thousands of  startups are dying off as the Youtubes and Facebooks – built squarely on the shoulders of other people’s content  – scoop up the super gigantic big money.    It is not a problem that startups die – in fact it’s a good part of the ruthless evolution of things – but it’s problematic when the lion’s share of online resources from the work of so many are redistributed to so few.    Not because this is “unfair”,  but because this type of  inequity does not lead to optimal system efficiency and growth.

Social media in all its various and sundry forms is a wonderful development.  Finally we see clearly that people, not computers, will be at the heart of future online developments – probably for some time into the future.    Facebook users are now leading the innovation in this area, though Alice at NYT thinks this could lead to unintended consequences.

To protect this new socially charged online environment from the ravages of our silly, stupid and prurient human interests we’ll need better incentives than the big players currently offer to quality content producers.   Those incentives will ultimately shape the quality of online content for years to come.

Google’s Constitutional Amendment: The Right to Rank as you see fit


Some of the most lively debate and controversy at search conferences surrounds the issue of Google ranking rights.   At Search Engine Strategies in San Jose the most interesting (and confrontational) session involved Michael Gray taking Matt Cutts to task on Google’s aggressive stand on commercially driven linking.    

The stakes of the “right to rank” question may become even higher in the context of a recent Microsoft v Google case, where MS is suggesting in their court brief against the Google Doubleclick merger that the merger will create something like monopoly conditions in the online advertising space because (according to Microsoft’s sources) Google+Doubleclick serve more than half the world’s online advertising.  

Although I don’t think MS is attacking Google ranking methods directly here it’ll be interesting to see if Google claims that since their algorithm does not rank the free “organic” listings on a commercial basis the suit has less merit than it would if they *did* favor sites in the organic listings.   

This would, of course, beg the key point that Google’s ranking power is now so high that it can make or break companies – offline as well as online – depending on how they rank in the organic “free” listings.   This confers on Google an obligation that IMHO they still do not take seriously enough – the obligation to minimize the collateral damage and maximize the correct rankings using, if necessary, more human intervention.     In short I’m saying that until the results are *so good* that only highly subjective opinions are coming into play Google needs to do *more* than is currently done, based on the principle that “with great wealth comes great responsibility”.    Ironically I think Google’s success has to a large extent insulated them from the growing criticism in the webmaster community.   Some of that criticism is self serving, e.g. spammers who are unhappy their tactics now fail, but much of the criticism is coming from users and newly minted webmasters or mom and pops who are frustrated because they can’t seem to get ranked properly for even the most obvious queries.   Google blames the spammers for this, but it’s a dynamic process and more transparency from Google – perhaps with stronger forms of site and webmaster ID for “official” or clearly white hat sites – could go a long way to solving the transparency problems.

Over at Matt Cutts’ blog he makes this point about a recent ASK court case decision in favor of a search engine’s right to rank as they see fit.  This point lies at the heart of the right to rank debate:

 Again, it makes sense that search engines get to decide how to rank/remove content in their own index…

I replied over there:

Matt …hmmm….wouldn’t you agree that this has some clear limits?   What would you call crossing the line on this freedom to rank however you see fit?
*
If Google pulled what Yahoo did some time ago and essentially forced sites to pay for inclusion or be excluded would that fall within the sensical realm?  
*
MSN is claiming (somewhat ironically and hypocritically, but correctly) that Google’s ad power is becoming close enough to a monopoly that remedies are in order.  Historically there has been trouble when a single company or country controlled more than half a resource – why no problem here?      

—– end reply —–