Visitors to Yahoo's Music Unlimited will pay $6.99 a month for access to Yahoo's 1-million-song library. That's less than half what Napster and Real Networks' Rhapsody charge for similar services that permit the transfer of songs to portable music players. source
"It was not a hacking or a security issue," said Krane. He said the problem was related to the DNS, or Domain Name System, though Krane did not elaborate. The DNS translates domain names for computers.
"Google's global properties were unavailable for a short period of time," Krane said. "We've remedied the problem and access to Google has been restored worldwide."
Thought I doubt the tool has much use, I love how smart the marketing is. They show a time meter of how much time the tool saved to make it seem as though it is providing an exceptionally useful service for users. To me it just looks like an excuse for Google to try to collect more data.
In my last post there was the following comment
I agree with the premises of privacy, and of not giving people too much information, I just don't know how Google would hurt any individual smart search marketer's business model.
There are many ways Google can hurt many people. One thing you have to worry about with some of these helpful plugins is how often will you see screens like this?
If you rely on internet marketing and do not have a strong brand you can count on Google swallowing more and more of your margins as their network will allow the richest & most socially active businesses to learn from and control the search results.
Most people, even in SEO / SEM, donâ€™t seem to be entirely clear about what data mining actually is about. A lot of fuzzy concepts abound, but only a few people seem to realize the commercial potential inherent in owning the worldâ€™s largest database of trackable and verifiable user behavior.
Take AdWords: a great revenue stream for Google, true â€“ but offhand Iâ€™d estimate that the overall value of the data generated from that venture alone probably beats the AdWords revenues by factor 6 or more if properly processed, analyzed, calibrated and marketed.
The difference is that now, the CTR of the ad copy itself is factored in, instead of it being solely the CTR of the keyword. Which only makes sense, IMO, given that it is the quality of the keyword and the particular ad it brings up that defines relevance, for a given search. source
I always like smaller conferences because when they get too big you (as in me) feel lost in the shuffle. It only costs $100 to attend this one. Smart move by JupiterMedia, as this will surely prevent others from having an easy entry into this market space.
The $199 per month Urchin On Demand also now includes report profiles for up to fifty individual websites (Urchin's previous offering included reporting for only one site). The price includes up to 100,000 pageviews per month. Users can add one million more pageviews for only $99 more per month.
In addition to the reduced price and increased number of profiles, Urchin On Demand is now able to import -pay--per-click costs directly from Google AdWords accounts.
Many smart search marketers probably are not willing to be paid to give Google all their data. As time passes you can be sure that Google will drop their costs further as they try to kill off the business models of everthing between them and ad dollars.
Look, Fwht, & Mama are dropping like it's not hot. InfoSpace (which does search & mobile technology) recently lost about 30% of their market value as well
When I was new to SEO I did a bunch of on the page analysis to try to figure out exactly what other people are doing. The problem is that it gets you focused on things that do not matter. A site may end up ranking high at the sacrifice of conversion.
As search algorithms advance basic link analysis tools, at least for Google, are starting to become what keyword density tools are: a waste of time.
Link analysis software was cool, especially when Google used to show all of the PageRank 4 and above links, back when their search relevancy algorithm was a bit more dependant on raw PageRank.
Now Google only shows a limited random set of backlinks, and the other search engines also limit the search depth to 1,000 results, which makes it hard to do useful analysis with the various link analysis tools on the market.
If it were quick and easy to query a database deeply (deeper than 1,000) then the link analysis tools would be much more useful. None of them currently on the market really make that a quick and easy process.
To keep improving the results, you find more variables for the algorithm-creating machine to use, and you add to your store of human-ranked pages for it to "learn" from. What you don't do is bother understanding the actual algorithm -- it was constructed by a machine and is way too complex for anyone to keep in their head.
Psychologists have shown repeatedly that when you give people a system to optimize, all you have to do is secretly introduce a delay between their actions and the results of their actions, and they will go bonkers. In fact, in a very simple (single variable!) model in which people are trying to control the temperature in a virtual refridgerator, you can get some of the same irrational responses you see in these forums
and the first post here by Captain CaveMan (which incidentally is the name of an awesome cartoon character) does as well:
Without giving away the store, I don't know how else to say it. There is no sandbox. People speak of it as though it were some simple 'thing' that stops new sites from being seen. That has simply never been true. What was true was that in its early days, some of the algo elements and related filters were so tight that only a very few new sites got past them (some accidentally; some methodically). Over time that changed; more sites started getting out, presumably as G worked to surface more new, higher quality sites.
There is no sandbox. There is only a serious of rotating algo's and related filters, that make it far harder for sites launched after spring of '04 to be widely seen in the SERP's. Not impossible. Harder. And certainly not as hard now as was true seven months ago. This has been hashed and rehashed so many times that it's hard to understand why it's still confusing.
If you can only see a few of the variables and overexert effort to satisfy those variables you may end up tripping filters and not satisfying other criteria.
AdWords Spying: GoogSpy looks scrapes hundreds of thousands of searches from Google to determine who is bidding on what terms. The idea is killer, but the implementation is a bit lacking. Link found from ThreadWatch.
So I have been getting some of the Gmail feeds and ads recently. Hopefully I answered this question correctly or you the reader will call me dumb...
Bad Call #1:
Here is an example thread
Question from Search Marketing Info
Which internet search engine was co-founded by a math major who chose the name to imply a vast reach ?
Thanks in advance,
Google was a mispelling for Googol, which means a 1 with a million
zeros behind it.
Larry Page founded it and Sergey Brin was his co founder.
and here was Google's contextually targeted Gmail ad:
Head Gasket Blown? - www.rxauto.com - Repair It Yourself Guaranteed ThermaGasket The Mechanics Choice
That is data stored on Google's servers and that is the best that they can target it? When you couple that in with all the AdSense spam sites and click fraud it really makes you wonder why Google assumes anybody would want that traffic.
Bad Call #2:
One of the default feeds was Engaget. Presumably because they run AdSense? Don't get me wrong here, its cool to help smaller publishers, but if you put Engaget in there you should put Gizmodo there also unless you want people to quetion you motives.
Placing random off target off topic crap I don't want in my email is being evil. At least the old Hotmail dating ads would occassionally show pictures of cute girls ;)
I know that I can unsubscribe from feeds, but I shouldn't have to opt out. Maybe off the start you could just promote Google News, Froogle, and your other portal pieces up there?
Bad Call #3:
Google actually places feeds in your spam folder. How stupid is it to place contextually relevant feeds near stuff that was deemed as being unwanted useless junk? What better way is there to turn users off?
Another thing that is really weird is most (maybe all) of the spam feeds were for spam recipies:
Spam Vegetable Strudel - Bake 20 minutes or until golden, serve with soy sauce
Savory Spam Crescents - Bake 12-15 minutes or until golden brown
French Fry Spam Casserole - Bake 30-40 minutes
They may place the spam ads in there to try to push the cute and innocent corporate culture, but I don't buy it.
After bogusly adding the Google Toolbar Autolink feature which directed B&N customers to Amazon many people started to become increasingly suspicious of potential hidden business partnerships. Is Google partnered with Hormel Foods now too?
I feared this post was reading as though I got it from Google's PR firm, so I felt I should include this... Google Blowout Quarter Update:
There's a blurb in the Wall Street Journal today that explains how Google's reported bottom line is being gamed by their own options program. It seems since they backloaded the options expense onto last year's earnings statement, this quarter's results will be ARTIFICIALLY BOOSTED almost 100%, even though it has absolutely nothing to do with their actual profitability as a business. Keep that in mind when they announce earnings tomorrow.