Google Mobilepocalypse Update

A day after the alleged major update, I thought it would make sense to highlight where we are at in the cycle.

Yesterday Google suggested their fear messaging caused 4.7% of webmasters to move over to mobile friendly design since the update was originally announced a few months ago.

The 4.7% of the websites Google pushed to go mobile friendly likely include some sites which would have been mobile friendly anyhow by virtue of being new sites on hosted platforms with responsive designs. But for the rest of the sites, was the shift worth it?

That is a tough question.

It is too early to tell.

  • Google still hasn't put much weight on it in the rankings yet.
  • Mobile traffic is typically worth far less than desktop traffic for most websites.
  • Time which was spent on mobile friendly conversion could have been spent on other forms of marketing.
  • Some sites which became mobile friendly took a significant revenue hit in doing so by switching out long running effective ad placements with mobile responsive units which may not have performed as well.

The problem with going early is you eat the expense upfront, while the rewards are still unknown.

  • Many people who jumped on the "secured everywhere" bandwagon last year saw broken security certificate issues and broken plugins which were hard to fix. And the upfront cost wasn't the only expense, as many AdSense publishers saw less relevant ads, lower ad CTR, and a sharp drop in AdSense earnings after going secured.
  • Those who spent the money to integrate Google Checkout to get AdWords discounts had to spend again to remove it when Google stopped supporting it.
  • TV makers who were early to integrate Google's YouTube API (which allowed ad free streaming) will now have to deal with a rash of customer complaints as Google sunsets the old API to make way to be able to sell an ad free subscription service.

If you are spending your own time & money and you believe in what you are doing and the longevity of a project then it doesn't matter too much if the rewards come slowly or never come. A sense of purpose & a sense of pride in your work is a form of payment.

However, if you are spending a client's money & you ring a 5 alarm fire to rush to make some technical change & then see no upside after the much hyped announcement, that erodes client trust. If there is no upside and a huge drop in revenue, then the consultant looks like a clueless idiot burning money for the sake of it doing various make work projects.

A few years ago a Google rep stated Panda would be folded into the regular algorithms. Then recently we were told it was a near realtime. Then we were told it was something where updates needed to be manually pushed out & it is something Google hasn't done in 4 months. If we trusted Google & conveyed any of these messages to clients, once again we looked like idiots. If we choose to invest client money based on the cycles and advice we are given, quite often that is a money incinerator.

Imagine dropping $30,000 on a link cleanup project where you remove links which were helping your Bing rankings but the Google update "coming soon" takes over a year to show up.

Invest money to lower your current income while you're waiting for Godat.

Good times!

So after Google made a big show of this pending mobile update by pre-announcing it, speaking about it at multiple conferences, comparing it to Panda and Penguin & stating it would have a bigger impact, sending out millions of warning messages via Webmaster Tools, etc etc etc .. when the big day came, did Google make the people who trusted them & invested in their advice look good?

Not so much.

Ayima recently launched a SERP flux pulse tracker tool which shows desktop and mobile flux side-by-side.

As you can see, nothing happened.

So far, no rewards. Maybe they will come. Though here is a hypothetical example where it could be very much NOT worth it for some publishers to go mobile friendly...

  • a webmaster managing an affiliate site converts it to a mobile responsive design
  • but user conversions on mobile devices in some verticals are unlikely, due to it being a pain in the ass to enter credit card info and so on ...
  • well ... person makes their site mobile friendly
  • that leads their mobile version of their site to rank better in Google
  • that leads to a greater share of their overall organic Google search traffic coming from mobile devices
  • their engagement metrics on mobile are somewhat weak, particularly when compared against desktop users, as is the case for many websites
  • their lower aggregate engagement metrics could create a signal which lead an edge case site into a false positive panda penalty
  • that then lowers their desktop search rankings
  • which lowers their desktop search traffic
  • which lowers their desktop search revenues
  • ...worse yet, ...
  • those affiliate cookies they dropped on mobile devices don't count for them when the user later converts on a desktop device

Any form of penalty (even a false positive) can become self-reinforcing. And many of the things which seem like they might help could cause harm.

Did you jump the gun or wait and see?

Consensus Bias

The Truth About Subjective Truths

A few months ago there was an article in New Scientist about Google's research paper on potentially ranking sites based on how factual their content is. The idea is generally and genuinely absurd.

For a search engine to be driven primarily by group think (see unity100's posts here) is the death of diversity.

Less Diversity, More Consolidation

The problem is rarely attributed to Google, but as ecosystem diversity has declined (and entire segments of the ecosystem are unprofitable to service), more people are writing things like: "The market for helping small businesses maintain a home online isn’t one with growing profits – or, for the most part, any profits. It’s one that’s heading for a bloody period of consolidation."

As companies grow in power the power gets monetized. If you can manipulate people without appearing to do so you can make a lot of money.

We Just Listen to the Data (Ish)

As Google sucks up more data, aggregates intent, and scrapes-n-displaces the ecosystem they get air cover for some of their gray area behaviors by claiming things are driven by the data & putting the user first.

Those "data" and altruism claims from Google recently fell flat on their face when the Wall Street Journal published a number of articles about a leaked FTC document.

That PDF has all sorts of goodies in it about things like blocking competition, signing a low margin deal with AOL to keep monopoly marketshare (while also noting the general philosophy outside of a few key deals was to squeeze down on partners), scraping content and ratings from competing sites, Google force inserting itself in certain verticals anytime select competitors ranked in the organic result set, etc.

As damning as the above evidence is, more will soon be brought to light as the EU ramps up their formal statement of objection, as Google is less politically connected in Europe than they are in the United States:

"On Nov. 6, 2012, the night of Mr. Obama’s re-election, Mr. Schmidt was personally overseeing a voter-turnout software system for Mr. Obama. A few weeks later, Ms. Shelton and a senior antitrust lawyer at Google went to the White House to meet with one of Mr. Obama’s technology advisers. ... By the end of the month, the FTC had decided not to file an antitrust lawsuit against the company, according to the agency’s internal emails."

What is wild about the above leaked FTC document is it goes to great lengths to show an anti-competitive pattern of conduct toward the larger players in the ecosystem. Even if you ignore the distasteful political aspects of the FTC non-decision, the other potential out was:

"The distinction between harm to competitors and harm to competition is an important one: according to the modern interpretation of antitrust law, even if a business hurts individual competitors, it isn’t seen as breaking antitrust law unless it has also hurt the competitive process—that is, that it has taken actions that, for instance, raised prices or reduced choices, over all, for consumers." - Vauhini Vara

Part of the reason the data set was incomplete on that front was for the most part only larger ecosystem players were consulted. Google engineers have went on record stating they aim to break people's spirits in a game of psychological warfare. If that doesn't hinder consumer choice, what does?

When the EU published their statement of objections Google's response showed charts with the growth of Amazon and eBay as proof of a healthy ecosystem.

The market has been consolidated down into a few big winners which are still growing, but that in and of itself does not indicate a healthy nor neutral overall ecosystem.

The long tail of smaller e-commerce sites which have been scrubbed from the search results is nowhere to be seen in such charts / graphs / metrics.

The other obvious "untruth" hidden in the above Google chart is there is no way product searches on Google.com are included in Google's aggregate metrics. They are only counting some subset of them which click through a second vertical ad type while ignoring Google's broader impact via the combination of PLAs along with text-based AdWords ads and the knowledge graph, or even the recently rolled out rich product answer results.

Who could look at the following search result (during anti-trust competitive review no less) and say "yeah, that looks totally reasonable?"

Google has allegedly spent the last couple years removing "visual clutter" from the search results & yet they manage to product SERPs looking like that - so long as the eye candy leads to clicks monetized directly by Google or other Google hosted pages.

The Search Results Become a Closed App Store

Search was an integral piece of the web which (in the past) put small companies on a level playing field with larger players.

That it no longer is.

"What kind of a system do you have when existing, large players are given a head start and other advantages over insurgents? I don’t know. But I do know it’s not the Internet." - Dave Pell

The above quote was about app stores, but it certainly parallels a rater system which enforces the broken window fallacy against smaller players while looking the other way on larger players, unless they are in a specific vertical Google itself decides to enter.

"That actually proves my point that they use Raters to rate search results. aka: it *is* operated manually in many (how high?) cases. There is a growing body of consensus that a major portion of Googles current "algo" consists of thousands of raters that score results for ranking purposes. The "algorithm" by machine, on the majority of results seen by a high percentage of people, is almost non-existent." ... "what is being implied by the FTC is that Googles criteria was: GoogleBot +10 all Yelp content (strip mine all Yelp reviews to build their database). GoogleSerps -10 all yelp content (downgrade them in the rankings and claim they aren't showing serps in serps). That is anticompetitive criteria that was manually set." - Brett Tabke

The remote rater guides were even more explicitly anti-competitive than what was detailed in the FTC report. For instance, requiring hotel affiliate sites rated as spam even if they are helpful, for no reason other than being affiliate sites.

Is Brand the Answer?

About 3 years ago I wrote a blog post about how branding plays into SEO & why it might peak. As much as I have been accused of having a cynical view, the biggest problem with my post was it was naively optimistic. I presumed Google's consolidation of markets would end up leading Google to alter their ranking approach when they were unable to overcome the established consensus bias which was subsidizing their competitors. The problem with my presumption is Google's reliance on "data" was a chimera. When convenient (and profitable) data is discarded on an as need basis.

Or, put another way, the visual layout of the search result page trumps the underlying ranking algorithms.

Google has still highly disintermediated brand value, but they did it via vertical search, larger AdWords ad units & allowing competitive bidding on trademark terms.

If Not Illegal, then Scraping is Certainly Morally Deplorable...

As Google scraped Yelp & TripAdvisor reviews & gave them an ultimatum, Google was also scraping Amazon sales rank data and using it to power Google Shopping product rankings.

Around this same time Google pushed through a black PR smear job of Bing for doing a similar, lesser offense to Google on rare, made-up longtail searches which were not used by the general public.

While Google was outright stealing third party content and putting it front & center on core keyword searches, they had to use "about 100 “synthetic queries”—queries that you would never expect a user to type" to smear Bing & even numerous of these queries did not show the alleged signal.

Here are some representative views of that incident:

  • "We look forward to competing with genuinely new search algorithms out there—algorithms built on core innovation, and not on recycled search results from a competitor. So to all the users out there looking for the most authentic, relevant search results, we encourage you to come directly to Google. And to those who have asked what we want out of all this, the answer is simple: we'd like for this practice to stop." - Google's Amit Singhal
  • “It’s cheating to me because we work incredibly hard and have done so for years but they just get there based on our hard work. I don’t know how else to call it but plain and simple cheating. Another analogy is that it’s like running a marathon and carrying someone else on your back, who jumps off just before the finish line.” Amit Singhal, more explicitly.
  • "One comment that I’ve heard is that “it’s whiny for Google to complain about this.” I agree that’s a risk, but at the same time I think it’s important to go on the record about this." - Matt Cutts
  • "I’ve got some sympathy for Google’s view that Bing is doing something it shouldn’t." - Danny Sullivan

What is so crazy about the above quotes is Google engineers knew at the time what Google was doing with Google's scraping. I mentioned that contrast shortly after the above PR fiasco happened:

when popular vertical websites (that have invested a decade and millions of Dollars into building a community) complain about Google disintermediating them by scraping their reviews, Google responds by telling those webmasters to go pound sand & that if they don't want Google scraping them then they should just block Googlebot & kill their search rankings

Learning the Rules of the Road

If you get a sense "the rules" are arbitrary, hypocritical & selectively enforced - you may be on to something:

  • "The bizrate/nextag/epinions pages are decently good results. They are usually well-format[t]ed, rarely broken, load quickly and usually on-topic. Raters tend to like them" ... which is why ... "Google repeatedly changed the instructions for raters until raters assessed Google's services favorably"
  • and while claimping down on those services ("business models to avoid") ... "Google elected to show its product search OneBox “regardless of the quality” of that result and despite “pretty terribly embarrassing failures” "
  • and since Google knew their offerings were vastly inferior, “most of us on geo [Google Local] think we won't win unless we can inject a lot more of local directly into google results” ... thus they added "a 'concurring sites' signal to bias ourselves toward triggering [display of a Google local service] when a local-oriented aggregator site (i.e. Citysearch) shows up in the web results”"

Google's justification for not being transparent is "spammer" would take advantage of transparency to put inferior results front and center - the exact same thing Google does when it benefits the bottom line!

Around the same time Google hard-codes the self-promotion of their own vertical offerings, they may choose to ban competing business models through "quality" score updates and other similar changes:

The following types of websites are likely to merit low landing page quality scores and may be difficult to advertise affordably. In addition, it's important for advertisers of these types of websites to adhere to our landing page quality guidelines regarding unique content.

  • eBook sites that show frequent ads
  • 'Get rich quick' sites
  • Comparison shopping sites
  • Travel aggregators
  • Affiliates that don't comply with our affiliate guidelines

The anti-competitive conspiracy theory is no longer conspiracy, nor theory.

Key points highlighted by the European Commission:

  • Google systematically positions and prominently displays its comparison shopping service in its general search results pages, irrespective of its merits. This conduct started in 2008.
  • Google does not apply to its own comparison shopping service the system of penalties, which it applies to other comparison shopping services on the basis of defined parameters, and which can lead to the lowering of the rank in which they appear in Google's general search results pages.
  • Froogle, Google's first comparison shopping service, did not benefit from any favourable treatment, and performed poorly.
  • As a result of Google's systematic favouring of its subsequent comparison shopping services "Google Product Search" and "Google Shopping", both experienced higher rates of growth, to the detriment of rival comparison shopping services.
  • Google's conduct has a negative impact on consumers and innovation. It means that users do not necessarily see the most relevant comparison shopping results in response to their queries, and that incentives to innovate from rivals are lowered as they know that however good their product, they will not benefit from the same prominence as Google's product.

Overcoming Consensus Bias

Consensus bias is set to an absurdly high level to block out competition, slow innovation, and make the search ecosystem easier to police. This acts as a tax on newer and lesser-known players and a subsidy toward larger players.

Eventually that subsidy would be a problem to Google if the algorithm was the only thing that matters, however if the entire result set itself can be displaced then that subsidy doesn't really matter, as it can be retracted overnight.

Whenever Google has a competing offering ready, they put it up top even if they are embarrassed by it and 100% certain it is a vastly inferior option to other options in the marketplace.

That is how Google reinforces, then manages to overcome consensus bias.

How do you overcome consensus bias?