Hidden AdSense Ads & Hidden Search Results

Ads as content...works well for some. You know AdSense is out of hand when premade sites are selling on eBay.

Debt Consolidation:
Fun to hear a guy whine about his 2 sites in his sig file not outranking Forbes for debt consolidation
http://forums.seochat.com/t54393/s.html
because Forbes has an ad page. :(

I bet his debt consolidation lead generation sites are informational in nature :)

Some are discussing creating links wars
http://forums.seochat.com/t54378/s.html
as if that will help them rank. Good luck knocking out Forbes.com.

If Google dials up their weighting on large authority sites before Christmas maybe the solution is to buy ad pages on some of them. I bet there are some great underpriced ad links and advertisement pages if people would look hard enough.

Coolness:
Link to Jim Boykin's new tool...still a bad tool name though, IMHO.

Controlling Data & Helping Consumers Make it Smarter

Part 4 of my recent ongoing article...

Dynamic Site Advantages:

If you use a weblog or any other type of dynamic site, as content ages you create a large quantity of pages which can rank for a variety of terms in many engines. The site archive systems mean that posts not only get their own pages, but can also be organized by date and category. This creates what is considered to be legitimate keyword driftnet content bank.

People can also subscribe to the feeds to remind themselves when to come back and read your new information. Many people who read feeds also write sites with feeds, and can provide you with additional link popularity and another channel to acquire readers from.

Most people who subscribe to what you have to say will usually be people who agree with many of your points. This means that when they talk about you or mention your site you are:

  • likely to be presented to additional like minded people with similar biases to your own
  • in a positive manner
  • from a voice readers likely trust.

If people disagree with you and still subscribe to your feed then there is a great chance they will frequently want to say how wrong you are, maybe even linking through to your site.

Ultra Targeted Content:

Not all ideas need a whole article to explain. By publishing your thoughts with one topic per post it makes it easier for you to refer back to your own content in the future. It also makes it easier for others to point at / link to / reference it.

Ultra targeted content will also stand a good chance of ranking high for it's keyword theme since it is so well targeted.

Consumer Feedback & Product Catalogs:

For a long time creating pages by keyword phrase permutation was a functional SEO strategy, but Google does not want to display hollow product databases in their regular search results. Creating industrial strength spam works well for some, but as time passes the hollow databases need to get better at remixing sources and integrating user data.

If there is commercial value for a term Google believes Froogle & AdWords work well. It seems to be almost a yearly process that Google dials up the rankings on authority sites right around the Christmas shopping season. This forces merchants to need to buy in to the vertical shopping sites, buy AdWords, or spend Christmas out in the cold.

Allowing user feedback and interaction makes your content more original than most competing sites. It also adds value to the consumer experience & makes it easier to link at your site. Both of which make Google far more likely to want to include your site in the result set. Tim O'Reilly states Data is the Next Intel Inside:

While we've argued that business advantage via controlling software API's is much more difficult in the age of the internet, control of key data sources is not, especially if those data sources are expensive to create or amenable to increasing returns via network effects.

Google is just a giant feedback network, learning to better understand the relationships between queries, links, and content. If you own the smartest and richest consumer feedback network in your vertical you will only continue to gain profit, customers, and leverage, at least up until someone creates a better feedback network that displaces the current one.

My Lawyer Filed a Motion for Summary Judgement in the Anti Free Speech vs Blogger Case

Recently my lawyer filed a motion for summary judgement in the Traffic Power vs Seo Book.com case.

Ariel Stern, my lawyer, also spoke with Max Spilka, who told him that Traffic Power recently switched lawyers.

One wonders why they would switch lawyers so late in the case if they had a real case. Even if they aim to waste bundles of cash I think they well know truth is not on their side.

I hope other bloggers find the paperwork from my case useful. :)

Naming a Website & Redefining Language

Seth offers his new rules for naming, but unfortunately I think some of them are no good.

This:

This means that having the perfect domain name is nice, but it's WAY more important to have a name that works in technorati and yahoo and google when someone is seeking you out.

Sort of a built-in SEO strategy.

is debatable in it's presentation, but this:

So, that was the first task. Find a name that came up with close to zero Google matches.

is absolutely unnecissary.

The concept of needing nearly zero competition to rank is beyond me. If you are creating something of quality over 99% of the competing pages for any phrase are going to be of zero significance.

If your product or service is truely remarkable you should be able to redefine the meaning of language. That is what remarkable people & companies do.

By looking at the number of hits for a word you are just looking at the number of pages that have that term in it. Want a better glance at the competition? Search for the number of pages which have the word in the anchor text and the page title (that tool will be made better & open source...it is still very beta). Even that number does not matter much though.

You really only need to look at the top few results, because those are the ones you will be competing against if you are trying to own the meaning of a word or phrase.

When Nandini named a directory Web Atlas that was a bad call because there are authoritative well established .edu & .gov domains in that keyword space.

When I created Myriad Search, I did not look through the competition at all (in large part because I wanted to create Myriad Search for link popularity and personal use more than for it to spread widely).

In spite of spending under $1 avertising Myriad Search, in the first month the site already ranks at #11 or #12 in Google for the word "myriad" (out of 28,000,000+ results).

You don't really establish a cult status until after many people are talking about your product. People do not search for your product until they heard about it elsewhere.

You shouldn't think of your site starting from zero and every page that has the term you want to rank for as competition. You should compare the quality of your idea to the quality of the top ranking ideas and see if you think it is possible to outrank them based on that.

Also notice how Seth's post title sounding authoritative is more likely to get comments. New rules for naming sounds much more definitive than naming tips & ideas.

Dynamic Sites & Social Feedback

Part 3 in a series... let me know what you think :)

Blog Software is a Simple CMS

Some of the conversations stemming from my article series starting with Why Bloggers Hate SEO's & Why SEO's Should Love Bloggers have stated that blogs are just a simple CMS. The one catch is they are social in nature.

I have probably read about a couple hundred books, and have only emailed about 5 book authors to tell them how great their books were. Most of the book authors quickly replied to my emails to say thanks. This tells me that they must understand the value of having fans (Seth Godin surely fits in that group) or they are not as inundated with email as I sometimes am.

Compare the books, which take months to write, to most blogs. On blogs I have left hundreds or thousands of comments. Across my various blogs I have got thousands of feedback posts others have left. One blog is almost nothing but a framework for people to leave their comments, and yet they still do!

Some people have stated that blogs are a fad that will die out. They may be right, but if they die out it will only be if other software emerges which does a better job of social integration, as some of the current tools are lacking on many fronts.

Static Content & the Game of Margins

Some old estabished static sites may long live on, but both directly and indirectly the web is becoming more of a read write medium. Margins will require content to become more social.

In spite of years of branding and content creation even the most well known publishers are caught playing the margins, selling ad space aggressively, and push the blame onto their advertisers.

Creating content is a game of margins. If you use a static website, and update it's content to keep it current, you are writing over your old work, which means:

  • you are throwing away it's historical record
  • you are creating less pages (which means less chances to pull in visitors) , as each page is another search lottery ticket
  • it is likely going to be harder for an audience to find the new content
  • it is less likely people will reference the new content, since they do not know what URLs are changing when
  • it is less likely people will reference the old content, since it may eventually change
  • many people will not want to reread the parts they already read
  • as your content size grows it means you are forced to worry about keeping it up to date while still trying to keep up with the news and the shifting marketplace

Add all of those things together, and a business model which would wildly succeed could easily become a complete failure.

The static site this article is on generally sucked until my blog became popular. In spite of the effort writing this aritcle, my average blog post will probably be read many times more than this article is.

Who is a Static Site For?

When you first learn about a topic it may be useful to create a large site about the topics you are learning, just as a way of forcing you to learn it all. Even in doing that, so long as you map out the general hierarchy ahead of time, there is no reason to avoid creating the site using a dynamicly driven database. Eventually when I have enough time this site will likely be shifted to a dynamic format.

The only people who can really afford to get away with using purely static sites are:

  • those who have other dynamic sites which help build their credibility & authority
  • those who are creating a site out of boredom or for a personal hobby
  • those who are not trying to profit or spread ideas
  • those who are known as the authority on their topic (who can do well in spite of the shortfalls in their publishing methods)
  • amazing writers who write so well that they can do well in spite of their publishing format
  • those who were first runners or are in niche fields with few competitors
  • those who are gurus in fields that change slow
  • those who run tons of sites and want to make them scalable (although it is even easier to do this with dynamic sites)

In almost all the above cases I can point to examples of how using dynamic sites could save time or be more profitable.

Example of a Sucky Static Site:

Not too long ago I created a site called Link Hounds to give away free link building tools on. I find the tools exceptionally useful, but the site failed to take off for a number of reasons.

  • API Limitations: when I first announced the site people used it beyond the API limits and it did not work. I should ask the engines for increased limits.
  • Lack of Incentive to Syndicate: in part to make up for the API limitations I gave away the source code and referenced tool mirrors, but some who mirrored the tool did not want to share it with others. Also Yahoo! requires that sites have DOM XML support if you use PHP4 to program the tool. I should have had my friend program in PHP5.
  • Crap Design: While the site design was not bad for free, it obviously is not something stellar.
  • Open Source & SEO: Are generally not concepts which are paired together. I think it will take a bit of time for people to get used to it. An open source website recently asked me to write an article, so that may help a bit.
  • Perception of Value: People think they get what they pay for. In spite of the fact that some of my software is similar to (and in some ways better than) stuff that sells for $150 or more, some people think the software is worthless because it is free. Similar software with strong affiliate marketing is seen by many more people:
  • Boring / Static: If I started working a bit harder at link building and placed a blog offering a bunch of creative link tips on that site I suspect it would garner many more links.

As it sits, there is little reason for people to remember to go back to the Link Hounds site, so they rarely do.

Sites that are dynamic in nature which make it easy to give feedback will fair far better.

Will Unfollowed Redirect Links Ever Count as Votes?

I have not put much effort into following most directory type sites that use redirect links (especailly if they are not ranked well in the related search results), but will engines eventually count many weird links as votes if they notice that the users click on a link and like what is on the other end of the series of redirects? Will those links ever count as much or more than static links that never get clicked on?

Jim Boykin On History of SEO & the Top Results

The History of SEO:

2005 SEO - Yahoo and MSN, pound with lots of links at once and keep pounding with anything you can get for backlinks with a focused backlink text campaign. With Google, the older the site the better, slow and steady link building with a large variety of backlink text wins (notice it’s the opposite of yahoo and msn).

Google’s Top 10 Choices for Search Results:

I think that for most searches, the top 10 will consist mostly of these types of pages. I think Google does this on purpose to show a variety of Types of pages to the user.

If you’re targeting a phrase, you should start by figuring what type of result your site will be, and what it’s role is in the top 10, and who you’re "real" and "direct" competitor is and what it will take to replace them.

Great posts Jim.

also good comments here Reputable SEO Companies

Why SEO's Should Love Bloggers

Part 2 of an ongoing series, the future mini articles will shift away from blogs and into other areas, at the end there will hopefully be a point to all these :) if not well then sorry ;)

Blog Blog Blog Blah Blah Blah

I run an SEO related blog which sells a guide to doing SEO, and yet despite the chronic hate toward SEO many authoritative bloggers recently linked through to my site because I was sued by an SEO company for blog posts & comments. As of writing this I am unsure of the specifics of what made my site worth suing, although those lack of specifics pissed off many people.

Most likely Traffic Power thought they could scare me silent, and since I was a blogger with a few good blogging friends that story backfired rather badly for them. It is an easy story to link at, and many people did. Adam Penenberg painted a rather accurate picture of the situation. The story spread far and quick. There was much syndication of the story that my site started ranking for the word sued.

Traffic Power Sucks.com was sued along with me, and yet they got minimal coverage because: they did not want to talk much, and more importantly, they had a static site. Method of publishing plays a hugely important role in whether or not ideas will spread, and how quick they spread.

Smarter Content

Sometimes what makes you / your site comment worthy is what others do there, and how people react to that (just look at Threadwatch to see how important the comments can be). Allowing others to add content to my site allowed them to make the content smarter and more complete. It also was the exact reason why the lawsuit became so comment worthy.

People wanted to save the right to comment on blogs without needing to worry about others cutting off their feedback loops. It is a large part of the reason some think blog comment spam and trackback spam is so nasty: feedback about an idea is sometimes worth far more than the original idea.

Ease of Link Acquisition

By giving people something to talk about and reminding them to regularly visit your site it is much easier to build linkage data. It also is easier to reference old stories that once again become relevant as more news emerges.

The viral behavior of blog posts in a large social network benefits those who can figure out what stories would spread & why people would want to spread them. Arbitrarily answering questions like "How much is my blog worth?" is an easy way to get links.

Someone created a blog called anti-blog to say how lame blogs are. As soon as I found it I made a quick mention of how I thought they were a bunch of lamers. They quickly linked back saying how dumb I am. Easier and quicker than a link exchange, and that link is much more likely to be up in a year than most link trades, which usually turn out to be junk.

Echo Chamber

When you have a regular site and are stuck asking for links one at a time it is an arduous task. Blogs have an echo chamber effect. After stories are above radar they spread without effort, and sometimes how stories spread makes them linkworthy.

Examples:

  • SEO Inc cease and desist letter, as Danny Sullivan states:

    That last thread we actually pulled from our forums back in mid-April. No, not because of a cease-and-desist letter or any message. Instead, our forums have a policy about public spam reporting. We don't allow it, unless a site is incredibly well-known or the issue has become discussed in a variety of public forums. Ironically, with the many blog comments now about the cease-and-desist, the thread that previously was pulled now qualifies for restoration.

  • Google onsite dentist blog is a hoax
  • MC Hammer visits Google - how hard would you normally have to work to get authoritative topically related links from sites with a quality level as high as SEW?

I was not trying to pick on Danny with those examples. I used his site as the example because he is the most authoritative voice on search, has a journalism background, and a long history of spotting the future trends in search before they emerge.

Everyone likes to have a bit of fun. The often informal nature of blogs make it easier to reference somewhat random topics, especially if you get to be the crazy frog. Having a blog lets you tap the flow of linkage data from other related sites, for serious or fun stuff.

Hard to Reproduce

When you do link exchanges most of the sites that exchange with you will gladly exchange with your competitors. When your site garners linkage data from authoritative sites that are not heavily directly interested in making money or search rankings it is hard for competitors to reproduce your linkage data. In fact, if they prod too heavily on that front they stand a good chance of damaging their brand value and credibility.

Quality of Links

When you get links from within the content of an actively read channel typically

  • the individual archive pages have few links on each page
  • the links are the type that drive direct traffic. If search engines bias relevancy based on user data and link activity more then these types of links will become more powerful
  • the Google Sandbox concept really does not matter much if all the high ranking active channels are referencing you anyway
  • many links in social networks lead to secondary links

Poisoning the Keyword Databases for Self Promo

I just got an email from Andy Mindel of WordTracker, about the SEO software Matt Cutts thought was rubbish. Andy said the sales text was something likeso:

RankAttack technology does not submit your site to the search engines... rather it creates a persona of "popularity" around your site in the eyes of the search engines. The purpose in RankAttack technology is simple: get the search engines attention and make them want to list your website under the keywords you desire'

I can't see search term co-citation being a trusted source of data unless it is from well estabished search history accounts and/or there are also a number of news stories about the topic and/or new web pages on trusted sites about the topic.

If temporal effects of increased search volume are used to allow sites to gain link popularity at quicker rate then odds are pretty good search engines would also look at the number of news stories and unique sites posting about the topic.

I suppose you can write a number of press releases and the like, but it is going to be hard to get mainstream news coverage for most websites, and without it then I can't see any value in poisoning the keyword research tools in your keyword space (unless you are doing it to screw with competitors keyword research ability or marketing your site through spamming keyword suggestion tools - as many SEO companies have done).

Andy states that this type of search spam is poisoning the keyword databases, but WordTracker has worked hard to filter out most of it:

Unfortunately, this approach is skewing the popular keyword databases such as (our own) Wordtracker, KeywordDiscovery, Overture suggestion tool and the Google keyword tool.

However, we have improved our spam filter and 99% of these skewed terms have now been removed from the Wordtracker database.

Google Giving in to Wall Street?

Google to report proforma earnings, not just net

Google Inc. has agreed to meet Wall Street halfway in how it reports quarterly results, seeking to dispel confusion created by a strict adherence to accounting standards, the company said on Thursday.

The Web search leader said it would present its third-quarter results next Thursday, October 20, in net terms but also, for the first time, in operating terms, excluding the after-tax effect of expensing employee stock options.

Who could blame Wall Street for wanting to know the cost of stock options when Google is giving away huge ones.

With all the economic uncertainty that has been floating through the economy I am betting Google has a soft quarter compared to all the home run quarters they have been announcing.

I am guessing the stock may go down to $250 to $275, at which point it might be a good time to buy Yahoo! if they go down in tandem. I say all this while not having the money or guts to short Google's stock ;)

Pages