Yahoo Shareholder non-meeting


Today Yahoo Shareholders are meeting in San Jose.   Or maybe we should say non-meeting since there are apparentely mostly empty chairs and uneaten pastry in a venue that was to hold 1000.

With shares now trading about $19 you’d think shareholders would be out in force with torches and pitchforks, but Yahoo management – at enormous cost to shareholders and the company – has kept the corporate raiders and Microsoft at bay partly by granting a newly sheepish Carl Icahn a seat on the board and two more seats.     Icahn noted last week that enough large shareholders were sticking with the current board, making it impossible for him to take over the company.     His plan was fairly simple – buy a lot of Yahoo and then sell the company to Microsoft at a huge profit.    As a shareholder I remain  *totally* confused as to why large shareholders were unwilling to support this move – the obvious choice in terms of maximizing shareholder value with minimum risk.

However with challenges come opportunities.  Yahoo at $19 is looking pretty ripe right now given that Microsoft offered $31 just months ago when Yahoo’s prospects were not significantly different than they are right now.    Either MS is horribly miscalculating Yahoo’s value, or the Market is underestimating that value.     Clearly the current board is convinced there is a lot more value, and in this at least I would agree with them.

It’ll be interesting to see how the rank and file Yahoo folks are feeling at SES San Jose in a few weeks.   SES is the biggest search conference of the year in the heart of Silicon Valley, and hundreds of Yahoo folks will be there.  It will be interesting to get a feel for the current morale challenges at the company.

Disclosure:   Long on YHOO.  Considering buying more.

Knol Knows Ranking


Google’s new knol project features articles on any topic by anybody who cares to call themselves an expert.   The concept is really intriguing as it relates to bringing higher authority information online that will be vetted both by the author and by the community at large.

Search marketing guru Aaron Wall is one of the sharpest fellows in the SEO business and he’s been testing knol and has concerns about the results both in terms of outranking original material and copyright.

Aaron on knol

The SMX search folks – I think it was Danny Sullivan – have also been testing knol and also are suggesting knol material is ranking very well.     Google says they are not giving their own site preferential treatment, but I’m guessing what they mean is that they are applying the same rules to knol as other sites.    If, for example, those rules that are *applied equally to all sites* happen to include a high value for incoming links from Google sites or other knol pages, the knol content effectively has a big advantage, technically without any special treatment.

In terms of search strategy I think the rule here is to …. write some knol page for topics of interest to you or topics for which you want to improve your rank.      I expect knol to be a huge topic at the upcoming search conference  – SES San Jose.

Later:  Time to check out this knol thing by creating some content myself.    Here are listings I created for:

Beijing, China |     Blogs and Blogging

Extremism in the defense of the algorithm is no vice?


WordPress surfing led me to another interesting sob story from a penalized webmaster and my reply got so long it deserved to become a post:

Marshall Sponder wrote:

Take Know More Media’s case – you have 100+ blogs and 2+ years of content – that’s easy, 30,000 to 50,000 blog posts and Google, with just one or two paid links that pass PageRank, is going to throw the entire blog network out of it’s index over that?

Yep, it appears that’s it – that’s the reason.  But is it fair?  No.
Strictly from a users point of view I think it is very hard to justify technical penalties on good content.    Few users know or care what “hidden text” is, so if a mom and pop webmaster uses this tactic and Google deletes the otherwise informative, relevant website it is hard to argue that users are served well.    Even if a black hat SEO created a site filled with illegal tricks but also full of highly relevant quality content I think Google’s case against including that site is weak.  As a user I want  *quality content* and I don’t care about the site’s technical construction.    Where Google is simply banning sites for using spammy tactics I’d agree with Marshall that to be faithful to user centricism they really have to take it a step further and look at the content they are excluding.   Even if the content contains paid linking and other violations if it unique, quality content Google cannot exclude it without violating their stated “prime directive” of providing the best for the users.

However, Google has to manage about one trillion URLs, so obviously they need shortcuts in ranking and one of them is a page from AZ Senator Barry Goldwater’s playbook when – many years ago – he tried to justify an escalation of the Vietnam war, perhaps to nuclear level.   Google’s coin of the famous Goldwater phrase would be: “Extremism in the defense of the algorithm is no vice”.

I don’t think penalties are generally *fair* or *user friendly*, but I’m willing to concede they may be necessary for Google to function as profitably as they do since it would take a lot of human intervention to help every mom and pop determine what’s wrong with their sites.

However, I feel Google continues to fail in their obligation to communicate more effectively with penalized sites although I think they are s-l-o-w-l-y  catching on to the fact that most webmasters of penalized sites remain unclear as to why the site has been penalized or downranked.    Removal offers you a shot at “reinclusion” and (very rarely) possible webmaster tools staff feedback.   Downranking is algorithmic and Google will not generally offer any advice to help downranked sites.     In this case you generally want to re-read the webmaster guidelines and experiment with different approaches in an effort to boost rankings.

My view is that as many thin content database sites have flowed online Google is holding online material to a higher standard of quality, especially if it’s at a new website.    This helps explain why you can find well ranked pages that are inferior to pages at a new website.

There is a solution to all of this in my opinion, which is for Google to include a lot more community input and feedback into the process than they currently appear to do.    I’d guess the recent discussions to aquire DIGG may have been in part to gain more community feedback tools and data.     Historically Google has been brilliant at using algorithms to determine ranking and advertising, but has fallen short of brilliance in their ruthlessness in dealing with website practices they don’t like, leaving a lot of collateral damage – especially related to sites involved in “paid linking” and variations on that complex theme.

At SES San Jose 2009 I’ll hope to get to ask Matt Cutts more about this in person.   Matt is Google’s top spam cop and always very open to conversations about ranking and search.    In fact the best event of the conference is the Google Party where engineers are on hand to discuss search related issues – including complex ranking technicalities that are sometimes brought to Google’s attention as part of the search conference circuit.

Microsoft BrowseRank to compete with Google PageRank


CNET profiles a new paper showcasing a Microsoft effort to enhance search by looking at *user behavior* as well as the old standby standards that all the major search engines use such as links in to the page, the content of the page, titles of the page, and several others.

Google’s initial brilliance was recognizing that the link relationships on the web gave you great insight into the best websites. Google correctly noted that sites with many links to them, especially for a particular keyword, were more likely to match a users interest for that keyword. Although many factors have been included in Google ranking for years, pagerank was arguably the most important breakthrough. Initially the system tried to be an online analogy to academic citation. Google’s Larry Page reasoned that websites with more incoming links would tend to be better, and that those incoming links themselves should also be weighted according to the importance of the site from which they came.

The system started to show severe signs of wear as search marketeers as well as mom and pop businesses began to “game” the pagerank system, creating spurious incoming links from bogus sites and buying links from high rank websites.

Enter Microsoft “BrowseRank”, which will arguably be harder to game because it will monitor the behavior of millions of users, looking for relationships between sites, pages, length of time on page, and more. It’s a good idea of course but arguably it is Google that has *by far* the best data set to manage this type of approach. So even if Microsoft’s system starts to deliver results superior to Google’s one can expect Google to kick their own efforts into gear.

As with all search innovation the users shoud be the big winners. Search remains good but not great, and competition in this space will only serve to make everybody better….right?

Google Ranking Needs a Spanking


Over at the Google blog today Amit Singhal has post 1 of 2 that promises an introduction to Google ranking.  As usual I’m disappointed in the way Google maintains what to me is a pretense of transparency while using some very ruthless and mysterious tactics to downrank sites they claim don’t meet quality guidelines.   Google (correctly) sees themselves as warring with spammers for control of the web but (incorrectly) thinks transparency is the wrong approach in this fight.

There were some rumblings last year of contacting webmasters directly about site problems but my understanding is that this would represent only a tiny fraction of total sites under penalty.    Of course, due to so little transparency in this area we can’t know the real numbers.

I’ll hope Amit’s second post is a LOT more specific, because I think he’s already practicing the kind oblique speak that is becoming commonplace when many from Google talk about ranking:

Amit:
No discussion of Google’s ranking would be complete without asking the common – but misguided! 🙂 – question: “Does Google manually edit its results?” Let me just answer that with our third philosophy: no manual intervention.

That statement is false, and he should not say it.   He does try to clarify later in the post:

I should add, however, that there are clear written policies for websites recommended by Google, and we do take action on sites that are in violation of our policies or for a small number of other reasons (e.g. legal requirements, child porn, viruses/malware, etc).

Action?  Yes, of course he means the *manual intervention* he said above does not happen.  Google has a right to pull sites out of the rankings, though it is annoying how much they talk about NOT manually intervening when they do it.    Because of no transparency nobody outside of Google knows how often they manually intervene.    Amit makes  it sound like it’s only for horrors like child porn or malware, but note that the use of inappropriate “SEO” tactics such as “hidden text” can get you removed and even banned from the Google index.   Unfortunately for small sites – e.g. “Aunt Sally’s House of Knitting website”  Aunt Sally may have no idea her webmaster is using these tactics.   How often does this happen?    My guess is that hundreds of thousands of legitimate sites are ranked very improperly due to technical penalties, but due to no transparency (and probably no measure of this at Google) nobody knows.

The big Google problem is that the policies for algorithmic downranking are not *clear enough*.  Many SEO companies prey on this lack of transparency, ironically often using Google’s mystique to lure unsuspecting businesses into expensive “optimization” schemes that don’t work or can get them seriously penalized.

Part of Google’s search algorithm philosphy is that they don’t share details because spammers would exploit them before honest people.   Although a weak case can be made for this idea, a better one is that in  non-transparent systems dishonest folks will do *better* because they invest more energy into finding the loopholes.    For example inbound linking, a very hot SEO topic last year at SES San Jose, has complex rules nobody understands outside of Google.    For example linking between sites in an information network can be advantageous or it can be penalized depending on whether Google (rather than the community or webmaster) sees the practice as manipulative of the algorithm or user-friendly and thus allowable.

Amit – a clear policy is one where the webmaster will know, rather than guess, what they are doing to annoy the Google algorithm or the manual intervention folks.

There is a pretty good source for information about how to approach site architecture for optimal ranking and it is to read Matt Cutts’ SEO related posts here.

Although Matt won’t give out much about the algorithmic penalties that create much of the Google confusion and frustration for established websites, if you follow Google’s guidelines and Matt’s posts on SEO you are unlikely to have serious problems with ranking.     Of course unless you work to optimize a new website you will have the *standard problems* with ranking since your competition is probably doing basic SEO on their site.   I’d argue (along with many SEO folks) that the best way to enter things this late in the game and hope for good ranks is with a topical *blog* to support your website.   Start with several posts about your general area of business, using a lot of the terminology people would use to find your website, and add posts regularly.

I’ll be covering the SES San Jose Search Conference and expect to hear a lot more debate about the issue of transparency, blogging, and SEO.

Google on SEO


Search Engine Optimization is at the same time a simple concept (help the search engines find and rank your pages) and a very complex one (proper use of redirection when changing domain names, Google downranking, duplicate content and hundreds more topics that are covered online in many places and also at conferences like the SES Conference Series, Webmasterworld PubCon, or the SMX Conferences.  

Arguably the best source for basic SEO information is Matt Cutts’ blog, and he always has great summaries of the conferences at which he gives talks.    Here’s a great post from Matt today after Danny Sullivan’s SMX Seattle Conference.   Google has added some information to their famous (and infamous) webmaster Guidelines, which should be read by every webmaster as they are the best *basic* information about how to structure a site to be ranked properly.   You’ll also want to read Matt’s SEO posts which offer a lot more specifics and technical advice.  

Although several years ago you would *also* have been well advised to read up on some of the tricks of the trade such as various schemes for keyword optimization, I would argue that for most webmasters tricks are more likely to be counterproductive than productive.   This is a really rich topic because there remain many techniques that fall into a sort of gray area of optimization where ranks are affected, but crossing the Google draws between acceptable techniques and unacceptable can lead to severe penalties.   Since Google does not draw a clear objective line we have the ongoing gray area of optimization. 

Many SEO techniques relate to *linking* strategies and *keyword optimization*.     It is an area where I believe Google has in many ways fueled the rise of the very content they hate by making the rules too vague and (more importantly) allowed adsense advertising on pages that don’t meet reasonable web quality standards.   Early in the game I was often frustrated when I would improve on a bad page only to have it drop in ranks due to algorithmic quirks.   I soon decided to leave crappy but high ranked pages alone, fearing they’d be downranked if I changed them.  This in turn caused problems as Google tightened up quality standards. Google is great about transparency in several areas, but algorithmic search penalties are not one of them.

I should also say there are some exceptionally good SEO folks out there who always have amazing advice when I bump into them at conferences.    David Naylor and Aaron Wall, and Todd Malicoat all have remarkable insight into the complexities of Google ranking as does Vanessa Fox who used to work for Google and Danny Sullivan who runs the SMX series of SEO Conferences.    My general advice about SEO is to do it yourself or in-house, but there are a handful of people like this who know the game so well that the normal rules about avoiding SEO folks do not apply.

Blog Revolution Note XXIV


At SoundBiteBlog I stumbled (or rather twitter-comment-followed) an excellent post about how much the poisonous / ranting writing styles of many blogs help them succeed.   The author wonders if nice blogs can finish first …

The short answer is “sure”.  A good example is Matt Cutts at Google who rarely has a bad word to say about anybody at his blog yet has one of the most read technical resources on the internet for Google search issues.   Fred Wilson’s A VC is also a blog with heavy readership and a friendly tone.    Marc Andreessen at blog.pmarca.com  is another and there are many, many more.

However I think the key blogging success issue is ranking, and there are many ranking problems in blogging paradise.  Blogs that rank well will be read more often and in turn will confer more rank via linking, so the  *linking style* of most of the old timer blogs  has really inhibited the broader conversation.   The best posts about any given topic are rarely by A list blogs anymore but these posts are rarely seen because the ranking structure favors older, more linked blogs over those with less Google authority.   

The old authority models work much better for websites – where high ranks for a general category make sense  – than for blogging where authors tend to cover a lot of topics.    TechCrunch will appear with a higher rank than almost any other blog if a technology topic is covered even if their coverage is weak, wrong, or misguided.    A thoughtful and well researched post about a critical topic is unlikely to surface if it is written by an “outsider” and escapes the RSS feed of somebody prominent, or sometimes even if linking to that post is seen by the “A lister” as giving a potential competitor too much free juice.   Note how “up and coming” tech blogs like Mathew Ingram link generously while most A list blog writers – who are now often hired writers, paid to be seen as a key breaking source of news – are far less likely to  cite other blogs.    Ironically I think success has really diminished some formerly great blogs.    John Battelle is one of the most thoughtful writers on the web but now he’s way too busy with Federated Media to keep Searchblog as lively as it once was.  

Google and other aggregators (like TechMeme) in part use metrics similar to Google pagerank to define TechCrunch as more reliable because they have more incoming links, more history on the topic, and more commenting activity.   This is not a *bad* way to rank sites but it tends to miss many high quality, reflective articles from sources who do not actively work the system. 

Solutions?  I still think a blog revolution is needed more than ever to re-align quality writing and new bloggers with the current problematic ranking systems. 

In terms of the ranking algorithms I’m not sure how to fix things, though I think Gabe should use more manual intervention to surface good stuff rather than just have TechCrunch dominate TechMeme even when their coverage is spotty and weird.   I’m increasingly skeptical that TechMeme is surfacing the best articles on a topic – rather it seems to give too much authority to a handful of prominent but superficial stories.    As others link and discuss those stories we have only the echo of a smart conversation.  

I don’t spend enough time searching Technorati to know if they are missing the mark or not, but I like the fact they are very inclusive.   However like Google and I think Techmeme, Technorati has trouble surfacing content that is highly relevant and high quality but not “authoritative”.

For their part, Google needs to do more to bring blog content into the web search results.   Last year at SES Matt Cutts was explaining to me that they are doing more of this than ever and I’m sympathetic to the fact that fresh content into the SERPS will lead to spamming problems, but I’m finding that I often get more relevant results from a blog search at Google than a regular search.   This is more the case for breaking news or recent events but it has even happened for research topics where the blog search has led me to expertise I don’t find in the web listings.

SES San Jose


I’ll be covering the big Search Engine Strategy Conference in San Jose – SES San Jose – again this year. SES San Jose is August 18-22 and is always a fire hose of interesting information, though the highlight is often the enormous party at the Googleplex with a buffet dinner and “meet the engineers” session.

This is the world’s top Search Engine Conference started many years ago under the guidance of Danny Sullivan who is now working his own search conference series, SMX.

SES San Jose official website

Click Fraud Class Action against Miva / Lycos: Good idea, but payoff and motives questionable.


Update upon closer examination of the terms:   Holy crap, BatClickMan, this action is pretty bogus unless you are on the legal team.   Here’s the deal:  Lawyers get a bunch of cash from MIVA while the defrauded customers get 50% off future purchases of clicks from MIVA.      Given that MIVA clicks are generally of  questionable value and positive ROI is tough even with PPC campaigns at Google where they do much better job making sure clicks are legitimate and relevant, this is almost a worthless payment for the defrauded folks unless they have accounts with MIVA now and are spending huge amounts AND are getting some good  ROI.

I won’t even be bothering with this nonsense which appears more like the legal firm looking to nab a few million for an interesting case rather than much if any justice getting done. 

As a MIVA advertiser I just got the email announcing a class action lawsuit against Miva/Lycos that alleges:

… MIVA and Lycos breached their contracts with class members, unjustly enriched themselves, and engaged in a civil conspiracy by failing to adequately detect and stop “click fraud” or other invalid or improper clicks on online advertisements.  MIVA and Lycos deny Plaintiffs’ allegations and contend that all payments they have received from class members for online advertising were legally and properly charged …

I’m surprised there have been so few of these lawsuits because there has been and still is a staggering amount of click fraud, and despite some crackdowns all the advertising places are essentially misleading people about the extent of the fraud.    Part of the reason the wrath has been lower than one might expect is that you generally can get pay per click refunds from search engines  for many types of complaints and I assume they have done a lot of crediting of major ad accounts if fraud was discovered or even suspected.

Of course this may not be worth the trouble as the payout is in … wait for it … more MIVA clicks!    Ha – I guess this could be called the “one fraudulent click deserves another” class action?

Under the settlement, MIVA will establish a settlement fund of $3,936,812.00 on behalf of MIVA and Lycos, of which a portion will be used to pay class counsel’s fees and costs, and the remainder will be available to class members in the form of advertising credits that may be applied to up to 50% of the cost of future online advertising purchased from MIVA.  To receive credits, you must submit a valid and timely claim form.  Credits will be awarded on a pro rata basis, taking into account the amount that you paid to MIVA and/or Lycos for ads that you believe in good faith to have been result of click fraud and the total amount of credits available.  For example, if the amounts that you paid to MIVA for the affected ads were 1% of the combined online advertising revenues of MIVA between January 1, 2000 and September 30, 2007 and Lycos between September 23, 2002 and March 30, 2006, you would be eligible to receive 1% of the total available credits.  You must certify in your claim form the percentage of your ads you believe were the result of “click fraud.” Credits must be used within one year of issuance and may be used only for advertising on the MIVA Media US Network.

Here’s the online claim form and a lot more information:  www.PayPerClickSettlement.com

Google’s reinclusion nightmare


John Honeck has an excellent piece about the challenges with Google’s site reinclusion process, a virtual nightmare of inconsistency and confusion.     I’ve seen the benefits and pitfalls of good and bad Google rankings and indexing at many sites, and “inconsistency” is the only clear pattern.    On the one hand I don’t have enough information to fully “blame” Google for the problems.  They have their hands full deleting junk or deceptive sites created by extremely sophisticated spamming operations around the globe, but as I noted over at John’s blog:

This is an *excellent* set of observations, and with all due respect to my pal Matt I’ve always been totally unmoved by Google’s suggestion that making the reinclusion and webmaster information process more transparent would somehow jeopardize Google’s ability to kill spammers.

In fact from my observations over the years I think the lack of transparency, along with initally vague webmaster guidelines (now fixed), have caused many if not most of the spam problems as both spammers and regular web folks vie to push the limits of the rules while staying in Google’s good graces. The big problem now is the profound inconsistency in the way sites are indexed, and the fact that it’s very difficult for webmasters to get much feedback from Google.  Google would be well advised to consider better automated or customer pays routines to examine websites for problems and allow reinclusion, because the frustration is building more than they realize in the webmaster and small business community.