e enjte, 30 gusht 2007

Phishing: Spam that can’t be ignored

Etrade Services.com
Spam that can’t be ignored
By Shaikh Parvez
SEO
August 30 2007,


If you haven’t already heard about phishing, then get ready. Like a lot spam, phishing is a form of unsolicited commercial email. Whereas all spam is not a scam, all attempts at phishing are scams, and the potential losses to corporations and consumers alike is stunning.

Phishing, as the name implies, is when spam is used as means to “fish” for the credentials that are necessary to access and manipulate financial accounts. Invariably, the e-mail will ask the recipient for an account number and the related password using an explanation that their records need updating or a security procedure is being changed that requires confirming an account. Unsuspecting e-mail recipients that supply the information don’t know it, but within hours or even minutes, unauthorized transactions will begin to appear on whatever account was compromised.


If you haven’t already heard about phishing, then get ready. Like a lot spam, phishing is a form of unsolicited commercial email. Whereas all spam is not a scam, all attempts at phishing are scams, and the potential losses to corporations and consumers alike is stunning.

Phishing, as the name implies, is when spam is used as means to “fish” for the credentials that are necessary to access and manipulate financial accounts. Invariably, the e-mail will ask the recipient for an account number and the related password using an explanation that their records need updating or a security procedure is being changed that requires confirming an account. Unsuspecting e-mail recipients that supply the information don’t know it, but within hours or even minutes, unauthorized transactions will begin to appear on whatever account was compromised.

y now, most people know that giving this information away on the Internet is a no-no. With phishing, however, it’s almost impossible to tell that the e-mail is a fraud. Like spam, e-mail from phishers usually contains spoofed FROM or REPLY TO addresses to make the e-mail look as though it came from a legitimate company.

In addition to the spoofed credentials, the e-mail is usually HTML-based. To an undiscerning eye, the e-mail bears the authentic trademarks, logos, graphics, and URLs of the spoofed company. In many cases, the HTML page is coded to retrieve and use the actual graphics of the site being spoofed. Most of the phishing I’ve received pretends to come from PayPal and contains plainly visible URLs that make it look as though clicking on them will take me to PayPal’s domain. Upon quick examination of the HTML tags behind the authentic looking link, the actual URL turns out to be an unrecognizable and cryptic looking IP address rather than an actual page within PayPal’s domain.

PayPal, the payment subsidiary of EBay, is a common target of phishing. If you get one and you’ve never joined PayPal, then you obviously know it’s a fraud. But if you are a PayPal member, as I am, the phisher has at that point broken through the unofficial security-by-obscurity layer that once protected you. It not difficult to see how PayPal members could be victimized by this technique.

e hënë, 27 gusht 2007

Advanced On/Off-Page Optimization for Engines using Semantic Analysis

This discussion will be a "how-to" guide for conducting on-page and off-page optimization while targetting the advances made by search engines in the realm of semantic analysis and on-topic analysis. There are several threads at other boards (SEW & WMW) and some terrific info on these subject at miislita.com, but little of it has reached SEOChat.

Evolution of AI & Search Engines
Basically, search engines are getting smarter - using more and more advanced techniques to analyze pages, sites & links in order to return more relevant results. One of these techniques is that of topic analysis (in many, many forms) that tell a search engine whether a page is focused on a specific topic (term/phrase) based on the other words and phrases on the page (and their format, usage, placement, etc.) and in links & linking pages(not just anchor text). The basic method of retrieval of these "related terms" is the first item I'll focus on, followed by some basic instructions on how to use these terms to improve optmization on and off-page.

Retrieval and Discovery of Related Terms
  1. Search for the target term/phrase at Google and use 100 results per page.
    .
  2. Analyze (either manually, or through a script) the top 100 SERPs and put the text into rows of a table that can be compared and picked apart. note: I will try to have a tool that will do this for you by the end of the month on the site in my signature
    .
  3. Pull out the top 20 occuring phrases/terms of 1, 2, 3 and 4 words in length (don't count stop words - a good stop word list can be found at http://www.princeton.edu/~biolib/instruct/MedSW.html)
    .
  4. Conduct semantic connectivity (C-Index) analysis on each word/phrase in comparison to the target term/phrase

    C-Indices use the following formula to come up with a PPT (parts per thousand) number:

    C=Z/(X+Y-Z)

    Where:
    X = The number of pages containing keyword 1 (your target term/phrase)
    Y = The number of pages containing keyword 2 (the term/phrase you're comparing it against)
    Z = The number of pages containing BOTH keyword 1 & keyword 2

    This is important to understand and use, so I'll create a sample for the phrase 'seattle restaurants' compared to another phrase 'lake union':

    C=Z/(X+Y-Z) which is 14.77=6740/(58100+405000-6740)

    In this equation:
    X = The number of results at Google for a search of "seattle restaurants" (always use quotes for a multi-word phrase) - 58,100
    Y = The number of results at Google for a search of "lake union" - 405,000
    Z = The number of results at Google for a search of "seattle restaurants" "lake union" - 6,740

    The highest C-Index I've ever seen is between norton & antivirus - 140. Commonly, I'd start thinking of a word as semantically connected at around 10ppt and closely related over 25ppt.
    .
  5. A high C-Index means the terms are related. Rank your 10-25 phrases/terms according to C-Index and remove any that are lower than 10ppt. For caution's sake, I often repeat this activity at Yahoo! - BTW, Excel makes this take very easy.
Using Related Terms for On-Page Optimization
Many SEO specialists recommend natural language writing and I could not agree more. Write your text without thinking of SEO at all, the SEO pieces can be added in later. Just remember to base your the topic of your page on the term/phrase you're optimizing for. Once again, I'll use the step-by-step guide:
  1. Write your page naturally, think of marketing and conversion rates, not SEO (but keep the topic on the subject of your keyword).
    .
  2. Go back over your text and see if you can use the related terms/phrases discovered above 1 or more times in the text effectively. If you can't don't worry. Just do your best.
    .
  3. Check the term weight of your target term/phrase using the 2 tf*idf (term frequency inverse document frequency) formulas:

    Classic Normalized Term Weight uses the following equation:
    Wi = tfdi / max tfdi * log (D/dfi)

    Where:
    tfdi = term (or phrase of a given length) frequency in document
    tfdi = maximum frequency of any (same number word) phrase in document
    D = number of documents in the database (when using Google, I estimate at 8.1 billion)
    dfi = number of documents containing the term/phrase (# of results for a search in quotes)

    A second equation, Glasgow Weight, can also be useful (I generally use both when analyzing my own site vs. the competition):
    Wij = log(freqij + 1) / log(lengthj) * log (N/ni)

    Where:
    freqij = frequency of term i (a word or phrase of a given length) in document j
    lengthj = number of unique terms (word or phrase of the same length) in document j
    N = number of documents in database (again, I use 8.1 billion for Google)
    ni - number of documents containing the term (results of a search in quotes)

    Once again, I'll try to have a tool built to do this automatically for a page very soon. In the meantime, it's still worth using, and once again, Excel can come in handy.
    .
  4. Check the term weight of your top related words - they should optimally be lower than your target term, but higher than any other term (of the same word length, not counting stop words). You really do not need to get this exactly right, close really is good enough.
Using Related Terms for Off-Page Optimization
Once you have the list of related terms and the formulas for term weight, you can see where off-page optimization can be done. Simply check the term weight of your target phrase and related phrases at the sites and pages you want to get links from. The more on-topic the pages/sites are to your phrase, the more relevant the link will be. You don't even need the page or site to mention your particular term once, as long as the term weights of your related phrases are high.

I hope this has been of use to everyone. Please give me your honest feedback and I'll try to edit any errors/omissions.

e martë, 21 gusht 2007

We are frequently asked if Submission Services such as "Submit It" or "Register It" are worth using.

We are frequently asked if Submission Services such as "Submit It" or "Register It" are worth using.

Answer:

In a single word... No.

Why? you ask.

Several reasons...

First: Each different Search Engine uses different "relevancy criteria" by which your page will be indexed, while "submission" services are suited for only the one size fits all generic type web pages that typically get buried under a mountain of other web sites found in the various Search Indexes.

As you should know from reading the SE Book, your various web pages should be designed to "fit" each Search Engine individually. True, this is a lot of work but it is worth it because it is still the least expensive way to expose your business to huge numbers of prospective customers. Unlike TV and print advertising, the Internet is the most economical way to reach masses of people who are looking for your services and products! ...but in order to succeed you must use the right approach. Designing duplicate pages to "fit" each of the major search engines is part of the "right" approach.

On the other hand, when a submission service "submits" your page, they are not concerned about how "high" up on a search your page will be indexed. Their only promise is that you will be submitted. In addition, it is not their concern whether the index accepts or rejects your page -- their only job is to submit it.

Second: It is difficult to tell if your page is actually ever submitted. It has been our experience that web pages submitted through such a service have sometimes never shown up at all in the various indexes or catalogs that they were supposed to have been submitted to.

Third: Every expert we know agrees that submitting pages through a service is ineffective. And, it makes no difference how many indexes you are listed on if you can't be found... or if only a few people even use the index.

Remember that 95% of all people who look for indexed web pages search using one or two of the top eight indexes. If you focus your time and effort building pages that are designed to be found by these top eight indexes and then take the time to submit them "in house"... you will get the results that you are looking for; which, we assume, is increased traffic to your site that will translate to profitable income.

If, on the other hand you opt to have your pages submitted by the "services" then the most you can realistically hope for is to be buried somewhere in the middle of thousands of web sites in a few hundred indexes and get lost in the crush of web sites that are continuously flooding the Internet.

Now, of course, that's only our opinion and we could be wrong. If so, we would like to hear from someone (anyone) who has concrete experience that a submissions services can place a web page on various Search Engines in favorably advantageous position. If you know of any way to do this, please write us and we'll report it here. In the meantime, your submissions and page design should be handled "in house" by knowledgeable personnel in order to optimize your Internet marketing efforts.

Let's face it. Developing an effective web site can require a lot of time and a little money. It makes good sense to go the extra mile to insure it is registered equally effectively... and it makes no sense to trust your newly created "sales resource" to an "assembly line" submission process that is notoriously ineffective.

Should we use Submission Services such as "Submit It" or "Register It" ?

We are frequently asked if Submission Services such as "Submit It" or "Register It" are worth using.

Answer:

In a single word... No.

Why? you ask.

Several reasons...

First: Each different Search Engine uses different "relevancy criteria" by which your page will be indexed, while "submission" services are suited for only the one size fits all generic type web pages that typically get buried under a mountain of other web sites found in the various Search Indexes.

As you should know from reading the SE Book, your various web pages should be designed to "fit" each Search Engine individually. True, this is a lot of work but it is worth it because it is still the least expensive way to expose your business to huge numbers of prospective customers. Unlike TV and print advertising, the Internet is the most economical way to reach masses of people who are looking for your services and products! ...but in order to succeed you must use the right approach. Designing duplicate pages to "fit" each of the major search engines is part of the "right" approach.

On the other hand, when a submission service "submits" your page, they are not concerned about how "high" up on a search your page will be indexed. Their only promise is that you will be submitted. In addition, it is not their concern whether the index accepts or rejects your page -- their only job is to submit it.

Second: It is difficult to tell if your page is actually ever submitted. It has been our experience that web pages submitted through such a service have sometimes never shown up at all in the various indexes or catalogs that they were supposed to have been submitted to.

Third: Every expert we know agrees that submitting pages through a service is ineffective. And, it makes no difference how many indexes you are listed on if you can't be found... or if only a few people even use the index.

Remember that 95% of all people who look for indexed web pages search using one or two of the top eight indexes. If you focus your time and effort building pages that are designed to be found by these top eight indexes and then take the time to submit them "in house"... you will get the results that you are looking for; which, we assume, is increased traffic to your site that will translate to profitable income.

If, on the other hand you opt to have your pages submitted by the "services" then the most you can realistically hope for is to be buried somewhere in the middle of thousands of web sites in a few hundred indexes and get lost in the crush of web sites that are continuously flooding the Internet.

Now, of course, that's only our opinion and we could be wrong. If so, we would like to hear from someone (anyone) who has concrete experience that a submissions services can place a web page on various Search Engines in favorably advantageous position. If you know of any way to do this, please write us and we'll report it here. In the meantime, your submissions and page design should be handled "in house" by knowledgeable personnel in order to optimize your Internet marketing efforts.

Let's face it. Developing an effective web site can require a lot of time and a little money. It makes good sense to go the extra mile to insure it is registered equally effectively... and it makes no sense to trust your newly created "sales resource" to an "assembly line" submission process that is notoriously ineffective.

How do I submit my pages to the different engines without having to change them?

Question:

Stephen,

Although there is a lot of good information here, what I really need is a good general strategy.

From the material I get that the algorithms are different from engine to engine. By focusing on the eight engines profiled, I cover 90% of the Internet population. However, submissions to each engine should be individualized to get the advantage and not incur any penalty. What works for one engine may get me kicked off another.

Since submission time varies from 2-8 weeks, I can't just submit a page in a specific optimized format, wait for it to be indexed, and change it for submission to another engine. Since crawlers come back periodically, if the page has been "de-tuned" from that engine and "re-tuned" to optimize another engine, I could get bumped off the first.

How do you approach this issue from an overall strategy perspective? Sorting this all out is very important to me. Would greatly appreciate any help you can offer.

Paul



Answer:

You are correct, you can't just submit a page...

These days, the key to success on the internet is to have several pages. This is very easy to accomplish... (unless you have one of those "free" sites and are limited by your provider -- if so, we can assist you in getting great service for as low as $30 a month -- with upper level domain name and up to 10 megs of space).

s that reflect the keywords that people are likely to search for us under... Here's what we do with our own pages... Our main site(s) are programmed in such a way that they fit reasonably well with all of the engines. In other words, we use keywords & descriptions that work with the engines that use tags and we add keyword density wherever possible without detracting from the content of the page.

In addition, we use our on images and also a reasonable amount of keyword text at the top &/or bottom of the page.... there are no engines that object to these procedures and many recommend them in their FAQ's.

To give ourselves and our clients an added advantage, we also make additional pages that "point" to our site. These pointer pages, or side doors are easy and quick to design. They can also be duplicated easily and adapted for another search engine. Normally, they offer a short description of what we offer and links to the main page(s) on our site... sometimes they have small file byte sized graphics, sometimes just text, and sometimes they are even "redirect" pages.

These"pointer" pages are short and sweet... keyword dense and designed specifically for the particular engine that we are submitting to.

By the way, the engines that respond most favorably to pointer pages are Infoseek, Excite, and Lycos.

Now, keep in mind that once someone finds a page (any page) that is linked to the rest of your site, they have found your site.

By using this strategy, your "main" pages remain "unoffensive" and, if written intelligently in a generic sort of way, themselves have a very good chance of scoring high in a search.

On the other hand, the pointer pages, Meta keyword loaded pages, and keyword dense pages give your site that extra chance of scoring high on the search. In the meantime your main pages remain fixed... you don't need to change those very often -- only the "side doors' into your site need to be changed and updated.

To see an extreme example of what I am talking about, go to Lycos and search "hawaii" ... #2 on the list you will see a simple "re-direct" page that points to a main page on one of our sites. We borrowed this technique from someone who was using it on Lycos for a "sex" site and we decided to try it as an experiment -- it worked.

There are many tasteful ways to accomplish the same result. For instance, You can use a menu page that is keyword dense that points to your other pages on your site.

Just like in print or TV advertising, nothing works better than saturation. Today on the Internet, saturation is what the professionals are using to succeed... at a fraction of the cost of print or TV advertising... however, it does take some work and also some time to create the additional pages... as well as some maintenance -- yet still at a fraction of the effort it would otherwise take to "break bread with the big boys" out there in the international world of commerce.

And, yes -- it can take up to 2-8 weeks to find out if what you are doing is working, that is a reality of the net. (That is also a reality of most print advertising as well)

We recommend that you design your pages (site) according to the guidelines in the book and, when ready, submit them as your best shot -- some of them are very likely to work and whatever does not, will give you the info you need to redesign and resubmit... to a large degree it is still a numbers game even though the dice are now loaded in your favor... And there will never be a time when "trial and error" is not part of the equation.

As to getting kicked off a search engine? ...We have seen some pretty flagrant "tricks" that you'd think would get someone kicked off an engine -- but they did not... Unless you use irrelevant keyword spamdexing combined with 100's of pages intended to saturate every category imaginable (many people have actually done this -- don't!), we don't believe you will have any serious problems with getting kicked out of an index. (exception YAHOO - they only take straight forward pages... no tricks and they will just ignore you if you try)

Your best bet is to always analyze what your competition is doing... search the web using words that you'd expect others to look for you with. See who comes up... analyze their pages and determine what technique(s) caused them to score high.

Then, design entry(s) into your site using their page as a template -- only make yours better -- and use the improved versions as side doors into the rest of your site.

By doing this you have two things going for you...

  1. You know the page you used as a template is acceptable to that particular search engine
  2. You also know that you are emulating what is already working. Don't try to re-invent the wheel.
Thanks for your question, Good luck ...and I trust this will help you

Search Engine Bytes


Notes, Tips, Questions, & Answers...
aka, topics too short for an article, but too important to leave out!

Question Topics
Q&A - What's with Google's minus-30 penalty? ...and how can I undo it? Q&A - How long does Google take to remove a site after receiving a spam report?
Q&A - Are links from international versions of .edu sites also good for rankings? Q&A - What's the advantage of a Google Sitemap over a regular HTML sitemap?
Q&A - Are programs that automatically generate articles safe to use on my site? Q&A - Is there any way a new site can compete with an old one?


Q&A - What's with Google's minus-30 penalty? ...and how can I undo it?

  • We currently rank #31 for a huge number of queries related to our site. Previously we consistently ranked #1 for those exact same queries.

    Our own view is that Google has applied a 30 place penalty on our site. Have you seen this before? What seems really strange is that our site is displaying Google Sitelinks in position #31—a feature that's typically reserved for pages in position #1. If Google sees our site as sufficiently qualified to receive Sitelinks then surely we should rank #1. We would be very grateful for your thoughts on this.

Answer: You're experiencing a textbook case of a bizarre phenomenon that's cropped up recently in Google: the minus 30 ranking penalty.

Sites slapped with the -30 ranking penalty often appear in Google just as they did prior to being penalized—with Sitelinks and indented double listings intact, except that for every keyword they previously held position #1 for, they now rank at exactly #31.

At first it just seemed like a bizarre coincidence that pages were being dropped exactly 30 spots, but as more and more sites began showing up with this exact penalty it became clear that Google is doing this intentionally.

Sites experiencing this penalty are typically those which Google would consider "good"—meaning they're real sites selling actual products or services that provide value to their users and are in top positions for several keywords; but they've committed some infraction in Google's eyes. In some cases they appear to have duplicated too many pages on their site or maybe they have too many of the wrong kinds of links.

In any case, Google doesn't consider the infraction to be serious enough for a complete ban or major penalty. Instead they seem to be sending the site a message that their act is looking a bit shady and needs some brightening up.

Avoiding the -30 penalty involves doing the same thing that you would do to avoid any other Google penalty—play by the rules, avoid duplicating large parts of your site, and strive to get relevant and legitimate incoming links.

As to undoing a -30 penalty, that can take a while. These penalties are becoming more and more challenging to get lifted. Regardless, here's a check list of strategies to speed up getting that penalty lifted:

  • Make sure your site is super-clean and eliminate anything questionable having to do with your incoming and outgoing link structure. Then file a reinclusion request.

  • Sign up with Google Sitemaps and Google Webmaster Central. This will generally give you good feedback about any problems Google has spidering your site, and can sometimes give you feedback about a penalty.

  • Do a duplicate content check, both of your own site and for people who might be stealing your content. Copyscape offers this service. Their charge is not huge, but can get slightly expensive for a large site. Price is $0.05 per page, last we checked. And take action if you find someone is ripping you off.

  • If all else has failed you can call Google directly at 1-650-330-0100. However, it is still tough to get them to respond directly to penalty issues.

At least the -30 penalty is less ambiguous than Google's other penalties. This one tells us for sure that a site is being penalized while the other penalties are less exact; making it hard to determine if the drop in ranking is the result of a penalty, better optimization by your competition, or Google simply changing their ranking algorithm.

With the -30 penalty, Google is sending a direct message saying, Yes, you have been penalized and it's going to stay that way until you purge your site and its link structure of its over-optimization tactics.

top


Q&A - How long does Google take to remove a site after receiving a spam report?

  • Our main competitor is at the top of Google with about 25 indexed spam doorway pages with deceptive redirects and misleading & repeated words for important keyword search phrases. They have been reported to Google twice since early June, and Google hasn't done anything about it. How long does the process of removing a spam site take?

Answer: Google may never actually get around to following through on that spam report. We've seen people submit spam reports on sites that were complete spam and Google never penalized or banned the sites being reported.

You have to remember that Google almost certainly receives a massive number of spam reports everyday, some legitimate, others just being filed by people trying to harm their competition. Each has to be evaluated, and it can take quite a while before someone actually follows up on your spam report.

And once they do, Google rarely removes a site from their index manually. They instead prefer to try and do it automatically by adjusting their ranking and spam-detection algorithms. Most likely their team has reviewed the pages you reported to them. However, unless the offence is exceptionally bad or high-profile enough to be embarrassing to Google, they will most likely try to address it in their engine's algorithm versus using a manual edit of their search results.

up


Q&A - Are links from international versions of .edu sites also good for rankings?

  • You've written many times about how getting links from educational domains (.edu's) can give you a trusted link that's very good for boosting search rankings. I'm in the UK, and I'm wondering if links from British educational domains (.ac.uk's) can also provide a rankings boost.

Answer: Absolutely. While no one is really sure whether the rankings boost from .edu links comes from Google placing more trust on them because these links are more exclusive and hard to obtain, or whether it's simply an issue of educational domains simply being great links because they tend to be older sites with their own set of really good incoming links, the results are clear: Links from educational domains are great for rankings.

In the US, institutionally accredited universities (and similar postsecondary institutions) can use the top level domain (TLD) .edu. This domain is reserved for use only by schools—only a very small group of people can get their hands on one. This exclusive nature makes it harder to fake links from these domains, and Google does appear to have placed a greater level of trust on them (here's the rules on who can get an .edu domain).

The same applies to educational domain in countries other than the US, although those countries almost always use different TLDs. For example, schools in the UK use .ac.uk. A few other examples include:

  • .ac.in for India
  • .ac.jp for Japan
  • .ac.nz for New Zealand

These types of domains are also restricted in their respective countries. Other countries like Australia (.edu.au) China (.edu.ch) use a variation on the .edu, while countries like Germany and Italy just use the country's regular TLD but modify the domain name to indicate that it belongs to a university (uni-erfurt.de and unibo.it, for example). And a few countries, like Canada, don't give universities any kind of special domain at all.

Still, even in cases where the school doesn't get its own special top level domain (TLD), universities do tend to have old, high-quality sites with excellent inbound links, and are superb places to get a link from whenever you can. Here's some good tips on attracting .edu links to your site.

There's also the .k12 namespace, which uses the format k12..us. This is typically used for US high schools and elementary schools. That extension is not strictly a TLD, and it doesn't have the same restrictions as a .edu, therefore it's questionable whether search engines would assign it the same level of trust.

up


Q&A - What's the advantage of a Google Sitemap over a regular HTML sitemap?

  • Now that we have a Google XML Sitemap telling Google which pages on our site to index and how often to index them, do we still need our regular HTML sitemap that we used to use to make sure search engines could index all our pages?

Answer: Your site should utilize both an onsite HTML sitemap and a Google XML Sitemap. Both are an important part of helping your site get and stay indexed by search engines.

The regular HTML sitemap is just a webpage that links to all the other pages on your website. It makes sure that any search engine crawling your site can easily and quickly find and index all the pages on your site. We cover the HTML sitemap strategy extensively in Chapter Two of our Unfair Advantage book.

The Google XML Sitemap is a format Google designed (and which Yahoo and Microsoft have recently agreed to also use) which allows you to feed a range of information about your site directly to Google, Yahoo and Microsoft. This information includes not just the location of your pages, but also their importance relative to the other pages on your site, how often those pages should be indexed, and more.

In order to maximize your site's crawlability, we recommend you take full advantage of both an onsite HTML sitemap and a Google XML Sitemap. It's also important to realize the important role that good incoming links play in making sure your site stays well indexed. Links are the number one factor in getting search engines to index all your pages on a regular basis.

Furthermore, neither the HTML sitemap nor the Google XML Sitemap play any role in where your pages will rank. They are simply a means to get your pages into a search engine's index. Where your pages rank depends on your incoming links and other optimization factors.

up


Q&A - Are programs that automatically generate articles safe to use on my site?

  • I'm looking for a new way to make money on the web and found a program which will automatically create a site for me, along with large numbers of already-written articles which are supplied with the software.

    The offer is enticing, but I'm suspicious. I've heard of this before and wonder if it's a legitimate way to do business. The program suggests you can set up a site in minutes. The last site I built took a lot of money and about 6 weeks, so their offer seems rather too good to be true. Are programs like this safe to use? Any suggestions or advice will be most appreciated.

Answer: The site building software you're referring too is commonly called autogen software. As the name suggests, it will automatically generate a site in a few minutes. The problem is that such software is commonly used to spam search engines.

While there are ways to use this kind of software in compliance with search engine guidelines, such software has been irresponsibly used to produce so many spam sites that search engines have developed the ability to recognize footprints, or tell-tale signs of an automatically generated site. When detected, such sites are often quickly banished from the search results.

Typically, the software comes with a large database of articles as well as the ability for you to input additional articles from article directories like ezinearticles.com or from your own subscription to a private label rights (PLR ) article database.

Of course, the problem with most article databases is that they're used by so many other people that you just end up duplicating content that is already found elsewhere on the web. To help solve that problem, autogen software typically offers the ability to modify the articles, usually by swapping out some words and replacing them with synonyms or other similar words.

While that can help the duplicate content problem, you're still faced with the issue that search engines detect footprints left by the software. For instance autogen sites tend to reveal themselves by using of similar template code common to most autogenerated sites. That means you'd be well-advised to significantly modify the template interfaces provided by the software to make it harder for search engines to detect that footprint.

As you're probably seeing, it can be a lot of work to "automatically" create a site. And even then you still end up with something of fairly low quality. That's not to say that these programs can't be used to make some pretty significant money if you create a lot of pages and avoid getting it banned by search engines. But doing it right requires a lot of expertise and usually falls into the category of unsustainable SEO/SEM that flies in the face of what Google and the other engines are looking for in terms of maintaining a relevant and useful search index.

However, if that's your cup of tea, then autogen software may work for you. But if you're aiming to create a site that will do well for the long term, then you'll want to focus on more sustainable strategies and avoid autogen software. If you do decide to test it out, be sure you:

  • Don't use it on a site that is important to you.

  • Use unique articles that no one else is using. You can generally find good article writers through Rentacoder.

  • Don't use the default templates provided.

Like the directory software we commented on last month, if a program offers to automate much of the site building or promotion process for you, it's probably the type of software that will get you banned before long. Building a long-term profitable online business means creating a site that customers will find unique and valuable and marketing it by getting people genuinely interested in what you have to offer. And those processes really can't be automated.

up


Q&A - Is there any way a new site can compete with an old one?

  • You've often written about the advantage older sites have when it comes to ranking in Google. Why do older sites tend to do better than newer ones? If older sites are dominating the rankings for a keyword, should I not consider targeting that keyword? And how do I determine how old a site is?

Answer: The age of a domain is an important ranking factor in Google, where older sites generally rank better. You can outrank an older site if you've got the right trusted links and good onpage SEO, but, all things being equal, old sites tend to do better. Since these sites have been around a long time, Google knows they aren't just part of some spammer's short-term "pump-and-dump" ranking strategy, so they're viewed as being more trusted.

Beyond Google's preference for old sites, there's also the fact that older sites have had longer to build links. So if your competitor's domain launched in 1996 and yours started in 2004, they've got an 8-year head start on link building and content creation. That's a significant advantage.

The Wayback Machine will give you an idea of when search engines first started indexing pages from a site. The drawback to the Wayback Machine is that it only goes back to 1996, so if a site came online in 1994 the Wayback Machine will still show its birthday as 1996. However, that doesn't much matter because you're not really concerned with a site's specific birthdate, just with whether it's a old site or a new site. Any site from the 90's is officially old by Internet standards.

To put things in perspective, a site launched in 1996 will not have much advantage over a site launched in 1998 (other than 2 additional years to build content and links). But it will have an often significant advantage over a site launched in 2004.

The other problem with the Wayback Machine is that some sites block its crawler. This means that site won't be listed. In that case you can sometimes use Netcraft.com. For example, here's where you can find the NetCraft birthdate for Google.com.

If that doesn't work, you can always go with the date their whois info reports, such as that provided by domaintools.com:

http://whois.domaintools.com/

However, this isn't a very reliable way to get a site's age, since the domain may have been purchased long ago but never put online. For example, we have several domains we bought in 1996 that were never turned into sites. If we were to create a site on them today, search engines would consider their birthday to be 2006, not 1996.

up

Search Engine Bytes


Notes, Tips, Questions, & Answers...
aka, topics too short for an article, but too important to leave out!

Question Topics
Q&A - Will adding more outbound links to my site hurt my rankings? Q&A - What is the best blogging software for my business?
Q&A - Should I host my company's blog on a subdomain or a subdirectory? Q&A - Will my rankings suffer if I change my pages from .aspx to .php?
Q&A - What's going on with DMOZ? Are they broken? Q&A - How concerned should I be with keyword density?


Q&A - Will adding more outbound links to my site hurt my rankings?

  • I run a popular niche directory and was planning on offering sites that list with me the option to have multiple links from their directory listing to various pages within their site.

    I believe this will make people more interested in being listed with me, but I was wondering if that could hurt my current rankings. I'm currently very well positioned for many keywords and don't want to risk anything.

Answer: You actually could experience some negative effects if you do this. This would come from diluting the PageRank you're spreading around to your site's internal pages. Let's say you have a PR4 page with 10 links on it, one of which links to another site, and the other 9 which link to pages on your own site (homepage, internal navigation, etc...). In that case, you're sending 90% of the link juice back into your own site, while 10% goes offsite.

Now if you add links to that page so that there are now 10 links to another site and 10 pointing to pages on your site, your site is getting just 50% of the linking power of that page returned to it, while the other 50% goes offsite.

Generally, this isn't a huge issue, but it is a good idea to be aware of the way PageRank and link popularity are passed around on your site. The more links pointing out from your site, the less those pages are able to help other pages on your site rank well. On a small scale it's not really an issue, since what's most important is the external links you have pointing to your site.

However, on a large scale like the one you're mentioning you'll likely be adding many hundreds of outgoing links all at once, so there is the potential for a negative impact on rankings. Don't mistake this advice for saying that you should never link to anyone. "Hoarding" PageRank is not productive. Just be aware that suddenly adding large numbers of outgoing links could effect the power of your internal linking strategy.

You might consider implementing this idea on a small scale at first. Outgoing links in themselves are not bad. It's the radical alteration in the way PageRank is distributed on your site that could hurt you. If you do decide to add these outgoing links, be sure to build additional incoming links to your site to help offset any negative impact of adding a large number of outgoing links.

top


Q&A - What is the best blogging software for my business?

  • I've been searching for some information on the best blogging software. I've seen recommendations for Movable Type, Word Press, Blogger and others. If a medium-sized company is looking to host a blog on their own server, what would you recommend today as the best software? I'd like for the blog to be search-engine friendly, of course.

Answer: Movable Type and WordPress are both great blogging platforms. We use each extensively. SearchEngineNews.com is actually run on a version of Movable Type which we had our programmers modify to work as a content management system.

WordPress is nice because it's free and fairly easy to use, plus it comes with a huge number of default templates to allow you to select how you want your blog to look. Movable Type is more complicated and you have to pay for it, but it comes with excellent customer support.

In simplest terms: Choose WordPress if you want something that's free and easy to use. Go with Movable Type if access to customer support is more important.

Do not go with the hosted versions, however. These include TypePad for Movable Type and the version WordPress hosts on their subdomains (like yourcompany.wordpress.com). You want to install the software on your own server to ensure that you maintain complete control of your blog. This becomes especially important as your blog grows. You don't want to have to deal with moving it to your own server at a later time and have to redirect all your incoming links.

Another option is Google's Blogger. This platform used to be quite limited, but has been steadily adding features. Blogger has both a version you host on Google's servers (as a subdomain of blogspot.com) as well as a version that allows you to publish pages directly to your own server.

Blogger is extremely easy to use and can be a viable option for business blogging. At this point, however, we still recommend the more robust and flexible Moveable Type or WordPress platforms.

up


Q&A - Should I host my company's blog on a subdomain or a subdirectory?

  • I read recently that it would be better to use www.xyz.com/blog as the URL for a blog instead of blog.xyz.com. The person said that placing the blog on a subdirectory would be better for search engine rankings as it would immediately inherit some PageRank benefits, which wouldn't happen if the blog was started on a subdomain. Any thoughts on whether this is true? Wouldn't both start at zero PageRank?

Answer: It won't make a difference one way or the other in terms of PageRank. The PageRank of the pages on the blog will depend on what links you have pointing at the blog, not on whether you're using subdirectories or subdomains.

Our recommendation when starting a new blog on a site is to go with the subdomain. Google lists subdomains as separate sites in their search results. So if someone does a search for your domain name, they'll see multiple listings from your site, such as:

www.xyz.com
blog.xyz.com

This allows you to occupy more top spots in the search results. When it comes to things like reputation management, the more top listings you can control on searches for your business or domain name, the better.

If your blog is already established on one of your subdirectories, however, you should leave it there, as the ranking headaches involved in moving your blog and changing all your URLs usually outweighs the advantages of having your blog on a subdomain.

up


Q&A - Will my rankings suffer if I change my pages from .aspx to .php?

  • My site currently uses .NET and I may be changing over soon to PHP. I've been told by my web developer that because PHP is open source there is a large amount of already written code available. Assuming this is all so (and what do you think?) my concern is that changing all the URLs on the site will hurt my traffic. I get about 60% of my traffic from organic search and I've worked too hard to blow all my SEO.

    Can I make this change without hurting my rankings? I've read many times in your publication to use 301 redirects—is that all there is to it?

Answer: You will hurt your search rankings if you change your URLs - there's no two ways about it. A 301 redirect can be used to minimize the damage, but we typically only recommend changing all of a sites URLs in extreme cases, such as when a site has lost their domain in a copyright dispute and is forced to move to a new one.

301s are the best option if you must change your URLs, but it's always best to avoid changing your URLs at all (unless those URLs have serious problems like too many dynamic variables making it hard for them to get indexed).

Anyway, you're not actually moving pages here, just changing extensions, so a 301 redirect is the wrong choice.

However, we can see why your programmer would want to change to PHP if that's the language he or she prefers, although we wouldn't say that there's necessarily anything inherently better about PHP over .NET. But you might get more bang for your buck if your web developer is using a programming language they like and are more productive in.

One option would be to set up your IIS server so that .aspx pages are processed by your PHP engine (typically php.exe). This a simple tweak in the configuration panel of your IIS server and one that can be easily reversed if you decide you want to go back to .NET.

If you take this approach your URLs stay exactly the same (with a .aspx extension)—they'll just be processed by a different programming engine.

To sum up:

  1. Do not change your file extensions. Your rankings will suffer, even with a 301 redirect.
  2. Set .aspx files to be processed by PHP if you want to make your developer happy.
  3. Or just leave things as they are.
up


Q&A - What's going on with DMOZ? Are they broken?

  • Hey guys. I have been trying to suggest a URL to DMOZ for 2 weeks and I keep getting a Service Unavailable message. Do you know what is going on?

Answer: The servers hosting DMOZ crashed on October 20th, 2006. This led to much of their data being wiped out. What's even worse, it turns out no one was really backing up most of that data on a regular basis. This meant that DMOZ had to be reconstructed from various sources like their RDF file used by various DMOZ clones such as Google Directory.

Even though the contents of DMOZ were restored after a few weeks, the panel editors used to edit their categories still needed work, so most editors were prevented from logging in until a few days ago. Thus new submissions weren't being accepted.

DMOZ is back up and running at full force on a new set of servers, and they appear to have learned their lesson about regular backups. However, you should be aware that most sites that were submitted but not yet reviewed have been wiped out. It appears that data is unrecoverable.

So if you've submitted in the past and haven't been listed yet, be aware that your site is probably no longer on the waiting list since the waiting list was deleted. This means you should resubmit your site if you want to get listed.

up


Q&A - How concerned should I be with keyword density?

  • With all the focus on link building, social networking and viral marketing these days, I was wondering if there's still a place in SEO for good old-fashioned keyword density? Is this still something I should be paying attention to?

Answer: We don't spend a lot of time on keyword density because it gets overshadowed by more important ranking factors like quality of incoming links and the trust your site has established by being older and having the right incoming links.

However, keyword density can still play a role, particularly with Yahoo, which is more focused on on-page factors than Google or Microsoft. We generally focus on getting our keywords in the right locations on the page: titles, header tags, internal link anchor text, lists, image alt text (particularly important if the image is also a link), and the actual text of the page. We break down where you should be putting your keywords in the following report:

A Top-Down Approach To Designing High-Ranking Web Pages

Once we have our keywords in the right locations, we try to write as naturally as possible while keeping an eye towards incorporating the keywords we want to rank for. Focusing too much on keyword density results in awkward-reading text, while writing naturally tends to incorporate the keyword variations and modifiers that both help a page rank for a greater range of keywords as well as looking more natural to search engines.

By the way, this topic touches on the concept of latent semantic indexing and how it's used by search engines. Latent semantic indexing is essentially a technology that allows search engines to determine what keyword a page should rank for even if that keyword isn't found anywhere on the page (or in the anchor text of the links pointing at that page).

For instance, a page which contains words like laptop, monitor, keyboard, hardware, compaq, hp, and pc is probably about computers, even if the actual keyword computer doesn't appear anywhere on the page itself.

By using related keywords and variations of a keyword (like computers or computing) you reinforce your primary keyword and increase your ability to rank for it. And, of course, you also boost your page's ranking for those related keywords.

Here's a few good resources for finding related keywords:

We also cover latent semantic indexing here .

Keyword density still plays a role but it's no longer as simple as just using a keyword 4% of the time on a given page. It's now much more about getting the right links, getting your keywords in the anchor text of those links, making your site easily crawlable, putting your keywords in the right places on your page, and writing in a natural-looking way that uses related keywords that reinforce your primary keyword.

Search Engine Bytes


Notes, Tips, Questions, & Answers...
aka, topics too short for an article, but too important to leave out!

Question Topics
Q&A - How do I find out which pages are sending traffic to my competitors? Q&A - Will cloaking my site get me in trouble with the search engines?
Q&A - Does it make sense to optimize my site's meta tags? Q&A - Is it worth my time to list my products in Google Base?
Q&A - What are the rules in regards to interlinking my own sites? Q&A - Is it safe to make a significant change to the topic of my site?


Q&A - How do I find out which pages are sending traffic to my competitors?

  • Is there a way to know how much traffic is being sent to a competitor's domain by specific referring pages? I'm looking for a technique or tool that might help.

Answer: If the competitor is large enough you might be able to get data from Hitwise or comScore. However, they generally don't have detailed data on smaller companies, and accessing their data can be quite pricey. Alexa will give you a rough idea of the level of traffic, but won't tell you where that traffic comes from.

Another option is to study their inbound links using Yahoo's Site Explorer. This won't tell you exactly which pages are sending a specific amount of traffic, but you'll be able to get an idea of where a site's non-search engine traffic is coming from.

Spyfu.com is another very useful site in getting important intelligence on your competition. Spyfu operates their own search spider which crawls search results to see who is ranking for what queries and what terms they're bidding on in Google AdWords. Simply enter a domain name into their search interface and Spyfu will return a range of useful data relative to that site.

By using Spyfu to see what a site is ranking for and what PPC keywords they're bidding on and using Yahoo Site Explorer to see who is linking to a site, you can build a pretty accurate picture of where their traffic is coming from. Cross-correlate that with their Alexa traffic stats and you should be able to get a pretty good general idea of a site's traffic profile.

However, if you want to get exact numbers, you can occasionally find sites who've left their server logs publicly available. Sometimes this is intentional, such as when they use a free stat counter program that publishes the data. Other times the data is made public accidentally when a site forgets to password protect the directory that hosts their server log files.

It's easy to scan a site using Google and see if Google has indexed any of the site's directories which are likely to contain stats files. For example, many sites store their stats in a folder like:

http://www.domain.com/admin/logs/referrers

Here we use domain.com as an example site, but you can see a real world example here:

http://opendocument.xml.org/admin/logs/referrers

In all likelihood, this site intentionally left its stats page public though other sites may do so unintentionally. Some common searches which will uncover similar directories include the following:

site:http://www.domain.com/ inurl:stats
site:http://www.domain.com/ inurl:statistics
site:http://www.domain.com/ inurl:sitestats
site:http://www.domain.com/ inurl:log
site:http://www.domain.com/ inurl:logs
site:http://www.domain.com/ inurl:referrers
site:http://www.domain.com/ inurl:referrer
site:http://www.domain.com/ inurl:referrals
site:http://www.domain.com/ inurl:refer
site:http://www.domain.com/ inurl:admin
site:http://www.domain.com/ "site meter"
site:http://www.domain.com/ inurl:awstats.pl
site:http://www.domain.com/ intitle:"statistics of"

Obviously, replace domain.com with the domain name of the site you're trying to find competitive intelligence on.

While the above examples often will turn up log files for a given site, the more important lesson here is that competitors may be using these techniques against you if you haven't properly password protected those directories on your site containing your sensitive data.

Remember that anything you upload to your site will likely be indexed by Google if you don't restrict Googlebot's access. Try running some of the above searches on your own site to see if Google has turned up anything you don't want your competition to see.

top


Q&A - Will cloaking my site get me in trouble with the search engines?

  • I recently read your UnFair Advantage Book, and I need clarification on one point. In reading about IP Cloaking, it seems most of the SEO world considers it to be "black hat", but you don't and I'm wondering why.

    If cloaking is banned by Google and other search engines, why do you recommend it in your book?

Answer: Everyone has their opinions on controversial issues such as this, and ours really doesn't differ that much from the mainstream. Cloaking is really about how and why you use it.

The reality is that Google lets many sites use cloaking with their blessing. For example:

  • Many newspaper sites let Google's spider in to read their content but require a human to register and log in. The only way they can do that is to use an IP delivery system; aka, cloaking.
  • Sites with content management systems that produce pages with uncrawlable session ID URLs for human visitors show Google versions of those same pages with simple, crawlable URLs. Some call that a content delivery system while others call that cloaking.
  • Advertisers running PPC campaigns use IP-based cloaking to prevent Google from indexing their pages which are only intended for their PPC campaigns.

There's also the multitude of flash- and image-based sites which show Google a text-based version of those otherwise un-indexable pages. This is similar to using the noframes tag to show indexable content to a search engine that can't process frames. Again, the only way they can do that is to use an IP delivery system; cloaking.

There is actually huge amount of legitimate cloaking going on all over the Internet that most people simply don't notice. Based on observing what Google does (not necessarily what they say) we can only conclude that Google is perfectly fine with it because, we suspect, such is not designed to deceive anyone.

Cloaking only becomes a problem for search engines when you're deliberately showing search engines substantially different content than what you are showing your human site visitors. However, rather than dealing with explaining the numerous circumstances where cloaking is permitted, it's far easier for search engines like Google to simply say "Cloaking is bad".

When it comes to SEO strategy, we don't make calls on what's "white hat" or "black hat" or even various shades of gray. What we do is tell you if a working SEO strategy is sustainable or non-sustainable. That is: will it work today, tomorrow, and for many years to come? Our practice is to help you find and employ the most sustainable techniques and strategies—those that will work for your company for the long haul.

In regards to IP cloaking, our advise has always been very accurate. We've been using this strategy in numerous applications since it was introduced in 1996 and have sites with over ten years of problem-free cloaking. To be fair, there are sites that have been banned for cloaking. Those are the ones that did not use it correctly and caused a problem or a potential problem for the search engines. Our goal is to help you learn the difference so you can make an informed, intelligent decision relevant to the nature of your online business, your site and the keyword topics within which it operates.

up


Q&A - Does it make sense to optimize my site's meta tags?

  • I'm having an SEO company help me optimize my site. For the past year they've charged me a small amount per month to submit my site to the search engines every other week and had me change my meta tags to their specifications, including adding a large number of DC metatags.

    Recently, another SEO company told me today that these things were a waste of time. Who's correct?

Answer: First of all, submitting your site to search engines every week is worth nothing. The engines have already found your site. And even if they hadn't, it's better for them to find your site by following a link from another site already in their index.

DC metatag development is a complete waste of time, Major search engines like Google do not index those specific tags. They are meant for places like universities and large companies to help manage their intranet document collections. We know this because we've been on the DC Meta Tag mailing list since its inception. We've also tested this just to make sure the engines aren't using them. They aren't.

In fact, the only meta tags search engines pay any attention to at all are the title tag (some people consider this a meta tag), the description meta tag, and in some rare cases the keywords meta tag.

We also see your current SEO company is simply sticking a keyword list in each of these tags, which is mostly useless. For instance, the meta description tag forms the description for your pages that people see in the search results. Its purpose is to convey a sales-oriented message that will entice people to click on your listing. A list of keywords would not be attractive to someone if they saw your listing in the search results.

What's more, your objective should be to have a unique title and description for each page that focuses on the main keywords that are on the page, not just a list.

Of the three tags just mentioned, the title tag is by far the most important for rankings and is something that you want to experiment with. Even though we aren't impressed with the look of your current page titles, at least they have your important keywords in them.

As a general rule - make sure each page has its own unique title tag and use your primary keywords in your home page's title tag. Then, on the product pages, focus specifically on keywords that include product name, part numbers, etc. Finally, always be sure to use words that people are going to be searching for.

For more on optimizing your title tags, see last year's report: The 7 Essential Title Tag Strategies of High Ranking WebPages in 2006 which contains strategies that are still completely relevant to title tags in 2007.

up


Q&A - Is it worth my time to list my products in Google Base?

  • My techie wrote an editor for Froogle feeds. It allowed me to plant keywords in the description, paragraph, and other tags. However, it wasn't well-synched with some parts of the website and fell into disuse. We are almost finished upgrading the backend of our website and I am debating whether I should spend more money reviving this.

    So, just how important is Froogle and Google Base for e-commerce sites?

Answer: We're currently working on a detailed article about Froogle/Google Base, and we can tell you that submitting a feed is very important. As you may know, Froogle no longer accepts feed submissions; everything is done through Google Base instead. A single Google Base feed can help your site be found in Froogle, Google Base, Google Maps, and the "onebox" that appears for many Google searches. Here's an example search that displays a onebox at the top of the search results (see the small shopping bag icon near the top of the search results):

http://www.google.com/search?q=Canon+PowerShot+SD+800+IS

Until our article is ready, we recommend your programmer spend some time exploring the documentation for creating and optimizing a Google Base feed.

http://base.google.com/support/?hl=en_US

The most important aspect of optimizing your Google Base feed is getting your keywords in the attributes fields of each product listed. Don't "stuff" the keywords in by repeating them over and over. Instead, decide on a set of related keywords you'd like each product to be found for, then work them into your product's title, description and other attributes in a natural and readable way.

Intelligent use of keywords within attributes is the key to optimizing your Google Base feed.

up


Q&A - What are the rules in regards to interlinking my own sites?

  • On the subject of having multiple sites on the same topic that interlink: I realize that if they are on the same IP, you probably don't want to interlink them. But if they are on completely separate IPs, would you want to interlink them as much as possible?

    Also, does the domain name registration have anything to do with Google actually realizing the sites are from the same owners? Ideally, I would like to have hundreds of domains out there that have slightly different content but the same end result: helping me pull in more sales. However, I don't want Google to think I'm creating a mini-net.

    To avoid this, would you suggest different hosting companies? Could I have two domains on the same hosting company and just make sure they are in different class C blocks? This is very confusing to me.

Answer: It depends quite a bit on what your intentions are with your site. If you are attempting to artificially manipulate your search rankings by creating multiple sites and linking them together (the mini-net technique) then you'd better do everything you can to prevent search engines from discovering that all of those sites are owned by the same person.

This would involve hosting them with different web hosts (which means they'll be on different IP addresses and class "C" blocks). But this is not all you'll need to do. You'd also want to register each site with different whois info, use different contact info on your site, and use different design templates.

In short, such a strategy dictates that you hide ownership of the various sites if you're going to be doing a lot of linking between your site. That's assuming, of course, that your goal is to "trick" the search engines into the impression that you have more links than you really do. Clearly, this requires a LOT of work. And, we don't feel it's a sustainable strategy.

Unless you've really covered your tracks well, at some point search engines will find your mini-net due to the likelihood that it will appear as an isolated node—devoid of links from any other sites outside your ownership influence. As such, your site will probably be penalized or banned eventually causing you to either give up or to start all over.

We talk about the different types of mini-nets and why some are risky in this report. We also discuss how to not make your sites appear to be a mini-net if you have multiple sites. It's a risky strategy, and you're much better served building legitimate links that will keep you on top of the search engines for the long term—and without having to worry that one day all of your sites will suddenly vanish from the search results.

up


Q&A - Is it safe to make a significant change to the topic of my site?

  • If we want to use an old domain which was used in another industry and just change it to be about a new industry. Will that be better in terms of SEO success than simply starting a new site? This domain has been around since 2000 and is quite old and trusted, and I'd like to launch a new site on it to help avoid the Google Sandbox. However, the old domain is currently on a very different topic than the one I'm planning to change it to.

Answer: Although old domains do tend to have a strong ranking advantage over new domains, they don't necessarily gain a free pass to the top of the search results. One of the primary considerations is that, when a site is sold to a new owner and the whois info changes, Google will often detect this and erase whatever credit it was giving to the site's incoming links, PageRank and birthdate.

Of course they do this to dissuade people from buying old sites just to cash in on that site's incoming link popularity. But, it's a tricky call because sometimes Google catches the change and sometimes they don't. They've also been know to let bigger sites keep their backlinks and PageRank credits while wiping the credit-ledger clean for smaller sites.

If you're looking at buying an old site—either to make it your business' main site or to link it to your main site—use a tool like domaintools.com's whois history to see who owned that site in the past: http://domain-history.domaintools.com/

If the domain has changed hands many times then it's possible that Google doesn't regard the site's birthdate as the day it was launched, but instead only considers it to be as old as the last time the site changed hands. You can also use the Wayback Machine to see how a site has changed over time.

For instance, take a look at the Wayback record for popular social media site Digg.com: http://web.archive.org/web/*/http://digg.com

According to Wayback, Digg launched in 1998. However, click on the pages that Wayback indexed back then and you'll see the site was actually the homepage for a company called Digg Records: http://web.archive.org/web/19990125094800/http://www.digg.com/

By 2001 it appears Digg Records let the domain drop and it was parked using a generic placeholder: http://web.archive.org/web/20010202015800/http://www.digg.com/

Then, it wasn't until December 2004 that the site started getting spidered regularly and indexed as the actual Digg.com we all know and (maybe) love: http://web.archive.org/web/20041209040106/http://digg.com/ In such cases, Google typically sees the domain birthdate as 2004, not 1998.

The point is that a site which has changed owners or themes may not be as "old" as it looks. Domaintools.com's whois history and Wayback's cached pages are a great way to get a feel for a site's history.

By the way, Digg.com is also a great example of the exception to the rule that new sites can't rank well. Clearly it's true that strong marketing or an innovative concept that takes off in a viral sort of way can overcome the disadvantage of a site's recent birthdate. Other examples of new sites that are able to rank well for extremely competitive queries are real estate finder Zillow.com and Internet video site YouTube.com.

So, domain age is not everything—but it helps. Whenever all else is equal, you'll be ahead of the game with an older domain. However, be sure to factor in the other ranking elements such as the links each has and the current topic of each.

Also be aware that it isn't so unusual for site topics to change gradually over time. But if you shift them radically and too suddenly, you could create rankability problems for the site. So, shifting the topic of an old site isn't a strategy that you can count on to work. It's better if whatever topic changes you're making are gradual.

If you must shift the theme of an old site, then strive to accomplish the change and in ways that Google won't consider it to be a new site. Perhaps you can add new sections while, temporarily, keeping the old sections. Then as your new sections take root in the rankings, gradually do away with the old sections that are no longer needed. Expect this to be a gradual process that can easily take a year or so to complete.