This site has one for $300, which is the same price as a paid submission to the Yahoo! Directory, but it can probably help you get many many many more links. ;)
Lots of quality features, including:
Autosubmit to unlimited sites
Use browser simulate system for anonymous Your requests
Use random proxies
Use random User agents
Use random Referer sites
Log file for requests
Save bad urls
Save successful submited urls
Check the proxies
Save bad proxies
URL extractor extracts URLs for specific terms with most popular search engines using fast technology - the 1000 URLs You can extract up to 20 seconds.
They also sell 3 lists of 10,000 blogs each at $100 a pop. A friend of mine who is a big time blog spammer stated that the crawl was the hardest part of blog spamming.
Now I do not know a lot about blog spam other than I delete lots of the shit. I have not tried out the blogspam tool as using it is negative 1000 karma points, but if you give it a try please let me know what you think of it.
I also think there are many valuable techniques to the art of effective blog spam. Some people probably are better at getting their spam to stick than others are. Its all about relevancy and providing useful content. hehehe :)
[update: A mate of mine has slightly better in house software but said that this second piece of software is solid spam framework. Again, I have not tried it though.]
The Term Extraction service provides a list of significant words or phrases extracted from a larger content. It is one of the technologies used in Y!Q.
Google Blogoscoped created a free auto linker tool, which makes adding on topic outbound links exceptionally easy. Am betting some people creating fake blogs probably enjoy the offering.
Part of Google's strong brand is PageRank, which now is of little use AND rarely updated. With all of these other good ideas Yahoo! Search is coming out with I am a bit surprised they are not providing and heavily promoting a regularly updated connectivity measurement service. Whatever happened to WebRank?
Black Hat: PPC Techniques. First time I have seen an article about Black Hat PPC. Good stuff Mikkel.
I do understand the reasons as they want to know what pages they send users to but why on earth does Google have to reset my hard earned high CTR if all I change is an added tracking parameter? In any case, I am not going to pay for it!
The simple solution is to set-up some kind of layer between you and them so the URL you use is actually not the one that shows the content.
Content, Content, Content:
A couple newerish (is that a word) products aim to help people grab or create loads of content. I have yet to use any of these.
Article Equalizer - pulls articles from various content sources. I think he also created traffic equalizer, rss equalizer, and many other equalizer products. Wonder if he will eventually release the ultimate equalier suite. ;)
Jim Boykin creates another free SEO tool. The reports take about 3 minutes to make. It tells unique linking domains, unique linking IP address, and unique linking C blocks (as well as how many links from each unique linking domain).
Since links from sites on the same C blocks are more likely to be owned by the same person some search relevancy algorithms may lower the weight given by links hosted on the same C block.
Just logged into the Digital Point Keyword Tracker. In the past it only worked for Google, but I just saw that they added Yahoo! and MSN as options.
It may not be a good idea to make it too easy for search engines to cross connect too many of your sites, but if you like the whole keyword tracking concept this is probably the best free web based one on the market.
Yahoo! and MSN limit the number of API queries from a given IP address each day. You need to place a script on your server to interface with tracking Yahoo! backlinks and Yahoo! or MSN rankings.
The tool can also track the number of Google backlinks and PageRank, although generally the data given by Google is rather useless.
When using it don't forget to use the full URLs including the http:// part. Why is Hub Finder Useful / Important?
Pages and sites which link to common related resources often exist in the same topical community.
Well themed pages have a tendency to rank well in search results since they use many similar variations to describe simarly related sites, products, and services.
Pages which rank well and are topically related may drive direct traffic which converts well.
As search advances more search technologies will likely place greater weight on links which come from pages and / or sites of the same theme.
One of the fundamental flaws with PageRank is that it looks at the web on the whole. Looking for thematically related links makes it harder for a person to manipulate relevancy with links from entirely unrelated powerful sites. By looking for thematically related links it forces websites to be well cited within their community of experts to achieve top rankings.
What does Hub Finder Cost?
Hub Finder is free.
Not only is Hub Finder free to use, but you can download the source code and place it on your site.
By default the tool has a link to this page and this site on it, but you can remove that if you like.
Hub Finder was created in PHP. Your host must support PHP and DOM XML for the tool to work properly.
Problems with Hub Finder:
Some scraper sites tend to scrape thematically similar resources. Some of these may show in the search results. Along with those many of the actual hub pages will be returned.
Hub Finder usually works best if you look for cross referencing backlinks in well developed fields or check backlinks across a good number of sites.
Hub Finder works with the Yahoo! API, which I believe only allows you to query their database up to 5,000 times per day. This is part of the reason why the tool was made to be distributed, so that anyone can host it and so one central host site was not hosting a tool that frequently exceeded its limit.
Additionally other upgrade ideas are listed below.
Suggested Upgrades for Programmers:
Fetch top ranking sites: Currently the tool allows you to enter URLs one at a time, up to a limit of 10 URLs. This tool could be improved by also allowing the option of entering a search term and fetching the backlinks of top ranked sites.
Disclude: Option would be to disclude results from any specified URL. (This could be used to prevent you from sifting through backlinks on sites you already know well. Or sites which clog up the results with hundreds or thousands of rubbish scraper pages).
Paired with: Option would allow webmasters to look at pages which link to topical resources AND link to a specific page or site. An example use of this tool could be informing webmasters that they have a broken or outdated link when one of your competitors moves their site.
Find pages OR sites: On top of searching for common backlink pages the tool could also have another option or section of its results which looked for common root URLs between backlinks.
Grab more details: Tool could grab IP ranges, number of links on the page, and link text.
Multi Engine: the tool could allow people to grab backlinks from multiple engines. MSN seems to be more realistic than Google is in their policy toward SEO tools.
Sort: Tool could allow people to sort the search results by any of the topics it grabbed.
A cool friend made this tool. I have a few other tool ideas. Production depends on:
how quickly I can learn PHP
or how quickly I can find programmers who would like to make a few tools
or how quickly my friend is available to make more tools.
If you would like to host Hub Finder or tinker with the source code you can find a copy of the original source code here. Change the file name to index.php and it should work if your host supports it. When using it don't forget to use the full URLs including the http:// part.
Upgrades & Mirrors:
If you upgrade or mirror the tool feel free to leave a comment below.
Website Dating: My Density shows first and second degree relationships between websites.
The interface could be a bit cooler and smoother (perhaps if they used a bit of AJAX and ensured the text was easier to read on scroll overs if they are showing a ton of data in a small area), but it looks like a cool idea. more info about My Density and even more info here.
I saw this mentioned a while ago, but I think I forgot to link to it. Essentially it allows search results to be reformatted & reappear elsewhere. Robin Good has more about why he feels it is important.
Resource Rate is the brainchild of Chris Ridings from SearchGuild. Essentially it aims to tap an Alexa type data from those of us willing to add a bookmarklet to FireFox to rate threads we view on various SEO forums.
Currently the system is heavily biased toward the newer / smaller / more hip forums ;)
If a few people join from some of the other forums it could help create a rather comprehensive resource. Editors get a share in the AdSense revenues from the site and it requires little effort beyond automatically sending back your feedback.
Buying Links to Ban a Competing Site:
Rumour has it that with the latest Google update a few people have started in on this practice...buying sitewide keyword rich links to help their competitors get blacklisted...surely SEO is going to get a bit more ugly here soon ;)
I do a good bit of directory submissions from time to time. I usually submit to about the same number of directories as that package except I do both free and paid directories. I do not mind paying for links because it means that the directory is more likely to have a functional business model and the links will not go away as quick.
Directory registration is exceptionally effective in MSN and Yahoo! right now.
other than having a different name what the hell does this software do that AdWords Analyzer does not? Wouldn't it have been better to release any additional AdWords related features as an upgrade to the AdWords Analyzer program?