<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	>

<channel>
	<title>Half's SEO Notebook</title>
	<atom:link href="/blog/feed" rel="self" type="application/rss+xml" />
	<link>http://seo4fun.com/blog</link>
	<description>Search Engines &#124; Blogs &#124; Marketting &#124; PHP/MYSQL &#124; CSS</description>
	<pubDate>Wed, 17 Jun 2009 17:22:13 +0000</pubDate>
	<generator>http://wordpress.org/?v=2.5</generator>
	<language>en</language>
			<item>
		<title>Your Obsession to Rank Higher is the Final Nail in Your Coffin</title>
		<link>http://seo4fun.com/blog/2008/05/28/your-obsession-to-rank-higher-is-the-final-nail-in-your-coffin.html</link>
		<comments>http://seo4fun.com/blog/2008/05/28/your-obsession-to-rank-higher-is-the-final-nail-in-your-coffin.html#comments</comments>
		<pubDate>Wed, 28 May 2008 13:03:06 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/?p=443</guid>
		<description><![CDATA[What kind of recording artist says stuff like &#8220;my goal is to make the Billboard Top 10&#8243;? Sure, recording labels may set goals like that, but did Lou Reed ever sit down and write songs just to win a Grammy? If making money is your only goal on the web, I don&#8217;t want you as [...]]]></description>
			<content:encoded><![CDATA[<p>What kind of recording artist says stuff like &#8220;my goal is to make the Billboard Top 10&#8243;? Sure, recording labels may set goals like that, but did Lou Reed ever sit down and write songs just to win a Grammy? If making money is your only goal on the web, I don&#8217;t want you as a client.</p>
<p>I get no kick out of promoting a piece-of-crap-boring-as-hell-cookie-cutter-turnkey-site. I got to be infatuated with a site I&#8217;m promoting. It&#8217;s got to blow me away. Simon Cowell may call you &#8220;karaoke&#8221;, &#8220;cabaret&#8221;, &#8220;cruise ship&#8221;, whatever - I only live once. I don&#8217;t have time to waste promoting mediocrity.</p>
<p>Udi Manber <a href="http://googleblog.blogspot.com/2008/05/introduction-to-google-search-quality.html">recently wrote</a>:</p>
<blockquote><p>[Google&#8217;s] goal is always the same: improve the user experience. This is not the main goal, it is the only goal.</p></blockquote>
<p>Don&#8217;t even think &#8220;if you build it they won&#8217;t just come&#8221; which is just a poor excuse marketers use to swipe more of your dough. If those words don&#8217;t strike a chord in your brain somewhere, nothing I do or say will help you.</p>
<p>Recently, several real estate agents were <a href="http://www.bloodhoundrealty.com/BloodhoundBlog/?p=3062">up in arms about Trulia using widgets</a> and nofollows, accusing Trulia of employing &#8220;aggressive&#8221; SEO (as if that&#8217;s somehow a bad thing). The truth is Trulia is just following the SEO rule book. The real threat is the amount of money and human resources Trulia has at its finger tips. Trulia&#8217;s technology will rapidly evolve. Meanwhile an agent is boxed in by an uninspiring &#8220;SEO-friendly&#8221; template thousands of other people are using and a limited marketing budget. There&#8217;s no competition. Taking collective action against widgets is just delaying the inevitable. Even if realtors stopped linking to Trulia, sites like businessweek.com <span class='nf' title="http://search.yahoo.com/search?p=linkdomain%3Ahttp%3A%2F%2Fwww.trulia.com+site%3Abusinessweek.com">will continue to link in</span>. If you want to talk massive inlinks, how about nearly <span class='nf' title="http://search.yahoo.com/search;_ylt=A0geu9J1UT1I8FUBOetXNyoA?p=linkdomain%3Ahttp%3A%2F%2Fwww.trulia.com+site%3Acnn.com&#038;y=Search&#038;fr=&#038;ei=UTF-8&#038;rd=pref">4,000 dofollow links to trulia from CNN.com</span>?</p>
<p>So far, one and perhaps the only edge agents have over Trulia are comprehensive, up-to-date property listings. If Trulia somehow gains access to that, what then? When will people wake up and realize SEO isn&#8217;t about worrying about SPAM (site positioned above mine) its about pushing their sites to the next level? If Microsoft didn&#8217;t bother creating the XBOX360 what would their market share look like now? If you refuse to evolve, your days are numbered on the web.</p>
<p>Your move. <strong>The clock is ticking.</strong></p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2008/05/28/your-obsession-to-rank-higher-is-the-final-nail-in-your-coffin.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>I Lurve Shari Thurow Too</title>
		<link>http://seo4fun.com/blog/2008/03/07/i-lurve-shari-thurow-too.html</link>
		<comments>http://seo4fun.com/blog/2008/03/07/i-lurve-shari-thurow-too.html#comments</comments>
		<pubDate>Fri, 07 Mar 2008 13:27:17 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<category><![CDATA[internal nofollow]]></category>

		<category><![CDATA[black hat]]></category>

		<category><![CDATA[PPC]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2008/03/07/i-lurve-shari-thurow-too.html</guid>
		<description><![CDATA[Recently, Shari got some flak for the condescending tone of her anti-nofollow post on SEL. But guess what? I agree with her.
Parasites of the Interweb
There are short-sighted people out there always looking for short cuts. They won&#8217;t hesitate to pay you $600/hour to hear you say &#8220;you need to write unique META descriptions.&#8221; Ya know, [...]]]></description>
			<content:encoded><![CDATA[<p>Recently, Shari got some flak for the condescending tone of her <a href="http://searchengineland.com/080306-083414.php">anti-nofollow post on SEL</a>. But guess what? I agree with her.</p>
<h2>Parasites of the Interweb</h2>
<p>There are short-sighted people out there always looking for short cuts. They won&#8217;t hesitate to pay you $600/hour to hear you say &#8220;you need to write unique META descriptions.&#8221; Ya know, easy fixes and promises of big returns. It&#8217;s all about high ROI. Contribute to the community <em>as little as possible</em> and milk it till its drier than <a href="http://en.wikipedia.org/wiki/Erg_Chebbi">Erg Chebbi</a>. It&#8217;s &#8220;you reap 10,000,000 times what you sow&#8221; syndrome. It&#8217;s about living off the web like a parasite. It&#8217;s about putting money in your pocket screw everybody else.</p>
<p>For those people, internal nofollow is an easy sell because its easy to implement. You don&#8217;t have to spend hours writing blog posts. You don&#8217;t have to share valuable ideas with other people. You don&#8217;t have to come up with anything original. You don&#8217;t have to dump your template site someone else spoon fed you and design a truely compelling site from the ground up. All you have to do is spend a couple of hours adding rel=nofollow to your pages.</p>
<p>I&#8217;m not saying short cuts don&#8217;t exist. After making thousands of bucks and 5,000+ visits/day per psuedo-spam domain using a script that took less than two hours to code, and having PPC campaigns that make me $10,000 for every $500 I put in, I know there are short cuts.</p>
<p>Still, there&#8217;s a time and place for every SEO tactic. We&#8217;re always short on time, and given multiple choices, we are forced to choose which path to take. Clients want instant gratification. Don&#8217;t give it to them - unless the client is willing to settle for 80% long term/20% short term strategy. If a client is unwilling to do the right thing, warn your client in advance that what you&#8217;re going to do for him/her is probably going to be a complete waste of your time and hir money. That way, a few months down the road, your client won&#8217;t come back to you and bitch that nothing is happening. The appropriate response in that scenario is &#8220;I told you so.&#8221; Don&#8217;t take the blame for your client&#8217;s bad judgement calls.</p>
<h2>What should you change before you touch internal nofollow?</h2>
<p>First order of the day has always got to be injecting value into your website so that its by far the best in your niche. If it isn&#8217;t the best, forget SEO, forget marketing, forget everything else. Work on improving your site.</p>
<p>How do I know that something I built is going to sell? I know because <strong>I use it every single day.</strong></p>
<p>Second, increase visibility. No, forget &#8220;authority links from high PageRank pages.&#8221; Get noticed on Craiglist, Myspace, whatever, it doesn&#8217;t matter. If your competitor is dominating Google Maps results in your area, for example, you know what you gotta do. If you run a search on Trulia and get a bunch of listings by your competitor, you got work to do.</p>
<p>Many big dogs in my vertical don&#8217;t depend on Google. Sure, they get 80,000+ Google hits/day, but that doesn&#8217;t represent the majority of their traffic. They leech traffic off other sites, they buy traffic, they get on top lists, they submit videos to Youtube &#8212; they do whatever it takes to generate traffic. Is the traffic not converting? Filter it, trade it, sell ads - as long as your site is visible, there&#8217;s money to be made. You fixate on rankings and high TBPR links and guess what? You&#8217;re probably going to end up just scraping by.</p>
<p>Internal nofollow isn&#8217;t evil. But unless you have unlimited amount of time, you gotta prioritize your SEO campaign. Internal nofollow should not be at the top of your todo list.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2008/03/07/i-lurve-shari-thurow-too.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Scraping 101: Extracting Anchor Text with Regexp</title>
		<link>http://seo4fun.com/blog/2008/02/08/scraping-101-extracting-anchor-text-with-regexp.html</link>
		<comments>http://seo4fun.com/blog/2008/02/08/scraping-101-extracting-anchor-text-with-regexp.html#comments</comments>
		<pubDate>Sat, 09 Feb 2008 03:23:18 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Coding]]></category>

		<category><![CDATA[Regexp]]></category>

		<category><![CDATA[Scraping]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2008/02/08/scraping-101-extracting-anchor-text-with-regexp.html</guid>
		<description><![CDATA[There are many ways to skin a cat, but when it comes to scraping websites, I like parsing content with regexp. One of the biggest problems I bumped into when parsing HTML is matching  opening and closing tags.
For example:
(&#60;a [^>]+&#62;)(.*)&#60;/a&#62;
Ok let&#8217;s try that in English:

(&#60;a [^>]+&#62;) matches &#60;a href=&#8221;&#8230;.&#8221;.&#62;.
(.*) *should* match anchor text (I&#8217;ll [...]]]></description>
			<content:encoded><![CDATA[<p>There are many ways to skin a cat, but when it comes to scraping websites, I like parsing content with regexp. One of the biggest problems I bumped into when parsing HTML is matching  opening and closing tags.</p>
<p>For example:</p>
<p><strong>(&lt;a [^>]+&gt;)(.*)&lt;/a&gt;</strong></p>
<p>Ok let&#8217;s try that in English:</p>
<ol>
<li>(&lt;a [^>]+&gt;) matches &lt;a href=&#8221;&#8230;.&#8221;.&gt;.</li>
<li>(.*) *should* match anchor text (I&#8217;ll elaborate on that).</li>
<li>&lt;/a&gt; matches the closing A tag.</li>
</ol>
<p><strong>&lt;a href=&#8221;http://www.searchengineland.com&#8221; rel=&#8221;notpaid&#8221;&gt;search engine land&lt;/a&gt;</strong></p>
<p>will correctly extract the anchor text &#8220;search engine land.&#8221; BUT because (.*) is greedy,</p>
<p><strong>&lt;a href=&#8221;http://www.searchengineland.com&#8221; rel=&#8221;notpaid&#8221;&gt;search engine land&lt;/a&gt; is cool because vanessa fox posts there.&lt;/a&gt;</strong></p>
<p>will incorrectly extract:</p>
<p><strong>search engine land&lt;/a&gt; is cool because vanessa fox posts there.</strong></p>
<p>as anchor text. Hmm..</p>
<p>So how do you fix this? Instead of using a .*, use .*? or other non-greedy modifiers like +?, ??, or {m,n}? (I haven&#8217;t tested the last three, I assume they work).</p>
<p><strong>(&lt;a [^>]+&gt;)(.*?)&lt;/a&gt;</strong> will extract anchor text from web pages.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2008/02/08/scraping-101-extracting-anchor-text-with-regexp.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Top 7 Reasons Why Optimizing Porn Sites is Hard</title>
		<link>http://seo4fun.com/blog/2008/01/16/top-7-reasons-why-optimizing-porn-sites-is-hard.html</link>
		<comments>http://seo4fun.com/blog/2008/01/16/top-7-reasons-why-optimizing-porn-sites-is-hard.html#comments</comments>
		<pubDate>Wed, 16 Jan 2008 19:23:06 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[AdWords]]></category>

		<category><![CDATA[Google Images]]></category>

		<category><![CDATA[Porn SEO]]></category>

		<category><![CDATA[Spammy Linking]]></category>

		<category><![CDATA[Yahoo Video]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2008/01/16/top-7-reasons-why-optimizing-porn-sites-is-hard.html</guid>
		<description><![CDATA[Sebastian recently fired up an experiment so I thought I&#8217;d post this up to give his experiment some link juice. While I&#8217;m at it, I&#8217;ll write a short rant about why optimizing porn sites isn&#8217;t easy.

Digg/reddit/stumbleupon are nearly useless. Of course, there are alternatives and workarounds, but really, life would be easier if digg had [...]]]></description>
			<content:encoded><![CDATA[<p>Sebastian recently <a href="http://sebastians-pamphlets.com/seo-test-do-search-engines-index-password-protected-urls/">fired up an experiment</a> so I thought I&#8217;d post this up to give his experiment some link juice. While I&#8217;m at it, I&#8217;ll write a short rant about why optimizing porn sites isn&#8217;t easy.</p>
<ol>
<li><strong>Digg/reddit/stumbleupon are nearly useless.</strong> Of course, there are alternatives and workarounds, but really, life would be easier if digg had a nsfw section.</li>
<li><strong>Everyone is scared of linking to you.</strong> Since almost everyone uses either affiliate content shared by 1,000 people or Matrix Content that looks all the same, there&#8217;s no such thing as unique content. Mainstream people are scared of linking to porn sites. And adult webmasters will not link to each other ever unless you set up a link exchange. There is no such thing as editoral links in the porn niche (except for sex blogs, but those are for female bloggers; men generally suck at talking on and on about sex). You either have a traffic link, an affiliate link, a reciprocal link, or an internal link. Nobody links out for free.</li>
<li><strong>90% of adult webmasters still don&#8217;t get the concept of one-way links.</strong> Major adult sites like penisbot was built on reciprocal linking done on a mass scale. They accept dozens of sites a day which are required to link to them and they in turn link to those sites. Those webmasters don&#8217;t see massive recip link networks as a link scheme because the links are &#8220;relevant&#8221; - as if Google could tell the difference between a page about [girl+girl action] and [slippery dildos].</li>
<li><strong>There&#8217;s nothing to write about.</strong> Porn is about pics and video - not text. It&#8217;s like optimizing for flash - its a nightmare because you have to spend hours typing bullshit just so you ranks for something. Yeah, you can write paysite reviews, publish chat logs with Brooke Banner or some other hot starlet that does weekly cam shows, post <a href="http://www.youtube.com/watch?v=djOWyCtqayM">barely sfw videos on youtube</a>, and talk about a girlfriend that&#8217;s been annoying you. But on the whole, you&#8217;ll end up writing alot of jibberish and saying stuff you don&#8217;t mean.</li>
<li><strong>No matter who links to you and who you link to, you&#8217;re in a bad neighborhood.</strong> If real estate sites got bitchslapped for excessive reciprocal linking, well, porn sites&#8217; been in hot water for years. Google doesn&#8217;t trust links in adult; even <a href="https://siteexplorer.search.yahoo.com/advsearch?p=http%3A%2F%2Fwww.penisbot.com%2F&#038;bwm=i&#038;bwmo=d&#038;bwmf=s">a site with over 300,000 links</a> is stuck at TBPR 5, a sign that most of those links aren&#8217;t passing much value. But even if each link only pass a trickle of juice, 300k crap links add up to #1 ranking for terms that drive over 80,000 visits/day from Google. This happens because even though you&#8217;re in a bad neighborhood and you got a busload of crap links pointed at you, everyone is also in the same rut, with a lot less links in their profile. So you still end up on top even if you break all the Google rules in the book.
<li><strong>Running AdWord ads is a challenge.</strong> Conversions are phenomenal; but the teen porn flag is easy to trip (different reviewers have different opinions on what is and isn&#8217;t compliant) and one disapproval too many can get your account killed.</li>
<li><strong>A Yahoo! Directory link costs a small fortune.</strong> Apparently Yahoo never heard of a level playing field or they assume, like VISA does, that porn sites make way more money than mainstream sites so adult webmasters can afford to pay more every year.
</ol>
<p>And a few reasons why optimizing porn is easier than mainstream</p>
<ol>
<li><strong>Google image search generates a ton of sales.</strong> I have a few images on this domain but none of them makes me any money. Not so with adult traffic. If you know how to nail top position, you really don&#8217;t even need organic search traffic. One key is to have as many thumbs on a page as possible (and of course the page has to be in the main index - unless Google changed that up). Post a ton of thumbs on your home page (yeah, Google image search is primitive - PageRank is a bigger factor here, since anchor text doesn&#8217;t work). Use huge pics - which helps you rank higher if a surfer filters a search by image size, and use framebreaker JS to prevent people from just looking at one pic then backpedaling to Google porn TGP.</li>
<li><strong>Yahoo Video.</strong> You can generate 600~1000 uniques a day per a set of videos thanks to Yahoo Video. The trick is to make sure traffic converts; otherwise your bandwidth bill may eat into your profits, especially if each vid is big (if each vid is 5mb, 1000 downloads = 5 gigs/day)</li>
<li><strong>There are thousands of long tails you can optimize for.</strong> Model names, celebrity names, paysite names, niches, superniches, DVD titles - an endless stream of keywords you can monetize, some with little to no competition.</li>
</ol>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2008/01/16/top-7-reasons-why-optimizing-porn-sites-is-hard.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>How to Check for Position 6 Penalty</title>
		<link>http://seo4fun.com/blog/2008/01/11/how-to-check-for-position-6-penalty.html</link>
		<comments>http://seo4fun.com/blog/2008/01/11/how-to-check-for-position-6-penalty.html#comments</comments>
		<pubDate>Fri, 11 Jan 2008 05:40:37 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2008/01/11/how-to-check-for-position-6-penalty.html</guid>
		<description><![CDATA[One of my clients is convinced he is smacked with a position 6 penalty. In some cases, he is coming up 7th or 8th, often when sites above him have indented listings. Where would he rank if those indented listings weren&#8217;t there?
Usually, Google pulls up indented listings if the secondary URL is relevant enough to [...]]]></description>
			<content:encoded><![CDATA[<p>One of my clients is convinced he is smacked with a <a href="http://www.seobook.com/google-ranking-6-penalty-filter">position 6 penalty</a>. In some cases, he is coming up 7th or 8th, often when sites above him have indented listings. Where would he rank if those indented listings weren&#8217;t there?</p>
<p>Usually, Google pulls up indented listings if the secondary URL is relevant enough to rank on the same page as the primary URL. As tedster on WMW noted, &#8220;You can see this mechanism at work by changing your preferences to 50 or 100 results per page - that opens up the opportunity for more urls to be clustered.&#8221;</p>
<p>Sandboxsam, a new user on Webmasterworld, noted a trick you can use to filter out those indented listings.</p>
<p>1. Go to Advanced Search on Google<br />
2. Set number of results to 20<br />
3. change num=20 to num=6 in the adress bar</p>
<p>That filters out secondary URLs from search results (unless it ranks 6th or higher).</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2008/01/11/how-to-check-for-position-6-penalty.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Paid Review: Internet Marketing Ninjas</title>
		<link>http://seo4fun.com/blog/2008/01/10/paid-review-internet-marketing-ninjas.html</link>
		<comments>http://seo4fun.com/blog/2008/01/10/paid-review-internet-marketing-ninjas.html#comments</comments>
		<pubDate>Thu, 10 Jan 2008 11:09:17 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2008/01/10/paid-review-internet-marketing-ninjas.html</guid>
		<description><![CDATA[You Can&#8217;t Judge a Book by its Cover
A while back, Fantomaster advertised Brad Callen&#8217;s PPC Arbitrage ebook on Sphinn with a blog post containing an affiliate link, which made some people think&#160;Fantomaster was trying to make a buck by leveraging his reputation on Sphinn. His post went hot,&#160;then a mod later pulled it off the [...]]]></description>
			<content:encoded><![CDATA[<h2>You Can&#8217;t Judge a Book by its Cover</H2></p>
<p><P>A while back, Fantomaster advertised <A href="http://www.netfrontiermarketing.com/adsense-arbitrage-an-overview.html">Brad Callen&#8217;s PPC Arbitrage ebook</A> on Sphinn with a blog post containing an affiliate link, which made some people think&nbsp;Fantomaster was trying to make a buck by leveraging his reputation on Sphinn. His post went hot,&nbsp;then a mod later pulled it off the front page. I got myself into an argument with Ralph after I called the ebook &#8220;crap.&#8221; He said (paraphrasing) &#8220;how the fuck do you know if its crap if you haven&#8217;t read it? You&#8217;re judging the content based on the way its marketed. Don&#8217;t talk till you read the book.&#8221; A day&nbsp;later, I downloaded the ebook. It was well-written, containing how-to information anyone&nbsp;can follow.</P></p>
<p><P>Now I&#8217;m in a similar situation where I&#8217;m asked to review something that I haven&#8217;t dug my teeth into. I don&#8217;t have access to Internet Marketing Ninjas&#8217; members area. I can tell you its a great program because <A href="http://www.shoemoney.com/">Shoe</A> explains step by step how he made his first mill, or that it sucks because the information that you pay for you already know, like how to set up a 301 redirect. But fact is no one can tell you whether or not&nbsp;the program is worth 3K unless they tried it.&nbsp;All I&#8217;m going to tell you is decide for yourself. Jim has his reputation riding on the success of this program - that by itself should tell you something.</P></p>
<h2>Facts About Internet Marketing Ninjas</H2></p>
<p><P><A href="http://www.internetmarketingninjas.com/" rel="nofollow">Internet Marketing Ninjas</A> &nbsp;is&nbsp;a brainchild of <A id=zdmi title="Jim Boykin" href="http://www.webuildpages.com/">Jim Boykin</A>, who recently won Search Engine Journal&#8217;s <A id=sfm7 title="The Best Link Building Blog of 2007 award" href="http://www.searchenginejournal.com/best-link-building-blogs-of-2007-jim-boykins-blog-the-link-spiel/6184/">The Best Link Building Blog of 2007 award</A>. Here&#8217;s the facts:</P></p>
<p><UL><br />
<LI>15 hours worth of video, featuring Aaron Wall, Jill Whalen, Lee Oden, Todd Malicoat, Jeremy Shoemaker, Jim Boykin, Niel Patel, Cameron Olthuis, Bill Slawski, Christine Churchill, and Jim Gilbert</LI><br />
<LI>More authority SEOs will join to contribute their knowledge and&nbsp;more videos will be added to the members area throughout 2008.</LI><br />
<LI>Access to Webuildpages tools that used to be public but are now private, including Quick Backlink Checker, Backlink Anchor Text Analysis, and Strongest Pages Tool.</LI><br />
<LI>Topics include link buying, keyword research, PPC, linkbait, Digg, link building, and affiliate marketing.</LI><br />
<LI>Membership: $3k/year</LI></UL><br />
<P><a href="http://www.internetmarketingninjas.com/" rel="nofollow">Visit the site</a> to see free preview vids.</P></p>
<h2>The Buzz</H2></p>
<p><P>Want more dirt? Check out the buzz. Since the program just launched, none of the articles really&nbsp;give you any specifics. Still, Hobo-web&#8217;s interview is worth a read; SEO book&#8217;s comments contain back and forth arguments about the program&#8217;s high price point; and you also might wanna check out the link to Webmaster radio&#8217;s podcast.</P></p>
<p><A href="http://www.seodisco.com/internet-marketing-ninjas/">http://www.seodisco.com/internet-marketing-ninjas/</A><br />
<BLOCKQUOTE>&#8220;I heard one of the best <em>testimonials</em> of Internet Marketing Ninjas from <A title=Stuntdubl href="http://www.stuntdubl.com/">Todd Malicoat</A>, when we were chillin’ at PubCon. When I asked him if he would advise me to shell out the money for the membership, he enthusiastically said that <A title="Jeremey Shoemaker" href="http://www.shoemoney.com/">Shoemoney</A>’s videos alone are worth the dough!&#8221; - Kid Disco</BLOCKQUOTE><br />
<A href="http://www.semscholar.com/2008/01/03/internet-marketing-ninjas-unleash-the-power/">http://www.semscholar.com/2008/01/03/internet-marketing-ninjas-unleash-the-power/</A><br />
<A href="http://www.hobo-web.co.uk/seo-blog/index.php/seo-ninja-linkbuilding-jim-boykin/">http://www.hobo-web.co.uk/seo-blog/index.php/seo-ninja-linkbuilding-jim-boykin/</A>><br />
<A href="http://www.toprankblog.com/2008/01/internet-marketing-ninja-interview-with-jim-boykin/">http://www.toprankblog.com/2008/01/internet-marketing-ninja-interview-with-jim-boykin/</A><br />
<A href="http://www.seroundtable.com/archives/015863.html">http://www.seroundtable.com/archives/015863.html</A><br />
<BLOCKQUOTE id=c_:n>&#8220;If you&#8217;re in doubt of the price tag, Barry has vouched for it. &#8221; - tamar</BLOCKQUOTE><br />
<A href="http://www.cartoonbarry.com/2008/01/sem_education_videos_internet.html">http://www.cartoonbarry.com/2008/01/sem_education_videos_internet.html</A><br />
<BLOCKQUOTE id=uj9v>&#8220;It is pretty expensive, but it seems to me to be well worth the price tag.&#8221; - Barry</BLOCKQUOTE><br />
<A href="http://sphinn.com/story/21112">http://sphinn.com/story/21112</A><br />
<A href="http://sphinn.com/story/21180">http://sphinn.com/story/21180</A><br />
<A href="http://www.jimboykin.com/internet-marketing-training-seo-tools/">http://www.jimboykin.com/internet-marketing-training-seo-tools/</A><br />
<A href="http://www.marketingpilgrim.com/2008/01/internet-marketing-training-course.html">http://www.marketingpilgrim.com/2008/01/internet-marketing-training-course.html</A><br />
<A href="http://www.netbusinessblog.com/review-internet-marketing-ninjas/">http://www.netbusinessblog.com/review-internet-marketing-ninjas/</A>&nbsp;(paid review)<br />
<A href="http://www.ilovejackdaniels.com/reviewme-reviews/internet-marketing-ninjas/">http://www.ilovejackdaniels.com/reviewme-reviews/internet-marketing-ninjas/</A>&nbsp;(paid review)<br />
<A href="http://www.brucecat.com/internet-marketing-ninjas.html">http://www.brucecat.com/internet-marketing-ninjas.html</A><br />
<BLOCKQUOTE id=jgfe>&#8220;For one year recurring membership fee of $2999.00 you will get access to some of the <STRONG>most amazing SEO tools around</STRONG> plus 15 hours of <STRONG>free</STRONG> video interview &#8221; - Bruce Cat</BLOCKQUOTE><br />
(emphasis mine)<br />
<A href="http://forums.digitalpoint.com/showthread.php?t=632542">http://forums.digitalpoint.com/showthread.php?t=632542</A><br />
<A href="http://raven-seo-tools.com/blog/68/internet-marketing-ninjas-and-their-value-to-seo">http://raven-seo-tools.com/blog/68/internet-marketing-ninjas-and-their-value-to-seo</A><br />
<A href="http://www.webmasterradio.fm/Search-Engine-Optimization/Webcology/Internet-Marketing-Ninjas.htm">http://www.webmasterradio.fm/Search-Engine-Optimization/Webcology/Internet-Marketing-Ninjas.htm</A><br />
<A href="http://www.seobook.com/jim-boykin-launches-internet-marketing-training-tools-combo">http://www.seobook.com/jim-boykin-launches-internet-marketing-training-tools-combo</A><br />
<BLOCKQUOTE id=se8h>&#8220;I agree that Jim could have shown more value upfront with some of the stuff he put in this package. Some of the tools have &#8220;never been released&#8221; next to them without stating what they do. As Jim gets feedback I am sure he will start offering more info about those.&#8221; - Aaron Wall</BLOCKQUOTE><br />
<A href="http://www.stumbleupon.com/url/www.internetmarketingninjas.com/">http://www.stumbleupon.com/url/www.internetmarketingninjas.com/</A><br />
<BLOCKQUOTE id=kt4t>&#8220;Jim is a friend of mine, and I know he always underpromises and overdelivers.&#8221; - Sebastian</BLOCKQUOTE></p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2008/01/10/paid-review-internet-marketing-ninjas.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>YAPAPL: Yet Another Damn Post About Paid Links</title>
		<link>http://seo4fun.com/blog/2007/10/13/yapapl-yet-another-damn-post-about-paid-links.html</link>
		<comments>http://seo4fun.com/blog/2007/10/13/yapapl-yet-another-damn-post-about-paid-links.html#comments</comments>
		<pubDate>Sat, 13 Oct 2007 15:06:11 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/10/13/yapapl-yet-another-damn-post-about-paid-links.html</guid>
		<description><![CDATA[&#8220;I do what I do best, I take scores. You do what you do best, try to stop guys like me.&#8221;
&#8211;Neil McCauley, Heat (1995)
Unless you were living under a rock for the last couple of weeks, you know Google&#8217;s been busy lately in its War Against Paid Links. According to Danny Sullivan, Google is now [...]]]></description>
			<content:encoded><![CDATA[<p>&#8220;I do what I do best, I take scores. You do what you do best, try to stop guys like me.&#8221;</p>
<p>&#8211;Neil McCauley, Heat (1995)</p>
<p>Unless you were living under a rock for the last couple of weeks, you know Google&#8217;s been busy lately in its War Against Paid Links. According to <a href="http://searchengineland.com/071007-173841.php">Danny Sullivan</a>, Google is now bitchslapping the &#8220;guilty high-rollers&#8221; to make an example out them and to put the fear of God into link sellers and link buyers. </p>
<p>While this move triggered a hailstorm of debates on ethics in the SEO community, the reality is this: Google will continue to tighten its security system. Whether Google&#8217;s stock price is $625 or $125, Google will continue to walk that path. Google will never be completely &#8220;hacker-proof.&#8221; Then again, it doesn&#8217;t need to be. Its goal isn&#8217;t to detect all paid links on the face of the Interweb.</p>
<p>See, the cops don&#8217;t work the streets expecting to catch every grocery-robber, rapist, and gangbanger in town (don&#8217;t get your SEO handbook in a twist because I&#8217;m comparing link sellers to criminals; its just a damn example, not a full-fledged analogy).</p>
<p>For example, according to ~1990 stats, only <a href="http://sa.rochester.edu/masa/stats.php">16% of rapes are reported to the police</a>. (Now I can see some people are gonna start asking, &#8220;but how do you define forced sex? If I tell my GF she can&#8217;t have my cream bagle unless she has sex with me, is that considered forced sex? Or what about if I rape someone but she ends up having 10 orgasms and begs me to marry her, steals my phone number and won&#8217;t stop calling me and says if I don&#8217;t marry her she&#8217;ll report me to the police for forced sex? What if she just felt a little uncomfortable for the first two minutes and then started really liking it? There&#8217;s so much gray area around the definition of forced sex maybe I&#8217;m a rapist and don&#8217;t even know it? I mean, some girls say no and then when I stop they say hey, why the hell did you stop? Keep going dammit! Do I really deserve to spend 7 years in jail and pay a $200,000 fine for using a cream bagle as a sex-bait tool? Isn&#8217;t that excessive? If my best friend rapes someone, why do I have to rat on him? Isn&#8217;t it unethical to snitch on a friend? He was drunk and he was horny. He couldn&#8217;t help himself. She was asking for it anyway, with that low cut dress exposing her boobs. She really shouldn&#8217;t dress like that. Yeah, its HER fault! Why do women make such a big deal about sex anyway? The government shouldn&#8217;t tell me what to do; it&#8217;s a free country. I should be able to do whatever I want!&#8221;)</p>
<p>In the United States, 1 out of every 5 women in college is raped (1995 National College Health Risk Behavior Survey). Less than half of those arrested for rape are convicted, 54% of all rape prosecutions end in either dismissal or acquittal, 21% of convicted rapists are never sentenced to jail or prison time, and 24% receive time in local jail which means that they spend an average of less than 11 months behind bars.</p>
<p>So in an ideal world, 649,733 rapes per year might lead to around 600,000 rapists sent to jail every year (give and take a few, considering cases where a group of men rapes the same woman or one man rapes more than one woman).  In reality, only around 103,957 rapes are reported, and even if all of the perpetrators for those rapes are prosecuted, 56,136 rapists are dismissed/acquitted, and 10,042 convicted rapists don&#8217;t see jail time. In the end, we&#8217;re left with 37,779 rapists in jail out of ~600,000 - or a 6% success rate.</p>
<p>If you rape someone, you have a 94% chance of getting away with it.</p>
<p>A cop, unless he/she&#8217;s smoking crack, doesn&#8217;t expect to get from a 6% success rate (or 94% failure rate, however you want to look at it) to 100% success rate overnight. What he expects is to see that 6% inch up to something like 9%. A crackdown on rape will not stop rape from happening. Rape is <a href="http://www.searchenginejournal.com/paid-links-simply-arent-going-anywhere/5817/">simply not going anywhere</a>. And if you rape <a href="http://www.jimboykin.com/buying-links-under-the-radar-so-matt-cant-find-them/">under the radar</a>, there&#8217;s a very low chance of ever spending time in prison. Police efforts may eventually fail. But just because the government cannot stop rape from ever happening doesn&#8217;t mean the government is going to just let it happen.</p>
<p>Google is a bank with a high-tech security system. Any security system, however, has its weaknesses. Matt Cutts and his Anti-Spam squad continue to plug holes in the system. They resort to scare tactics to lower the number of people trying to beat the system. They are trying to go from 6% to 9%.</p>
<p>An <a href="http://www.ericward.com/">SEO professional</a>&#8217;s job is to continue to find ways to bypass that system and &#8220;get the cheese&#8221;, as an open 9-ball hustler pal back in the Big Apple used to say.</p>
<p>But its not my job to whine about the system. It&#8217;s not my job to waste time questioning the legality of that system.</p>
<p>It&#8217;s my job to</p>
<p>: understand the system<br />
: exploit any weaknesses of that system</p>
<p>Identify tactics, draw up a plan, execute, re-evaluate, and make another run. If you&#8217;re in this for profit, everything else is digital vapour.</p>
<p>&#8220;Got. Got. What do we got? *What do we got?* Bon voyage, motherfucker. You were good. I&#8217;m going to the hotel. I&#8217;m going to take a shower. I&#8217;m going to sleep, for a month. &#8221;</p>
<p>&#8211;Vincent Hanna, Heat (1995)</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/10/13/yapapl-yet-another-damn-post-about-paid-links.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Third Level Push (modified Siloing) For Deeper Index Penetration</title>
		<link>http://seo4fun.com/blog/2007/08/22/third-level-push-modified-siloing-for-deeper-index-penetration.html</link>
		<comments>http://seo4fun.com/blog/2007/08/22/third-level-push-modified-siloing-for-deeper-index-penetration.html#comments</comments>
		<pubDate>Thu, 23 Aug 2007 03:32:47 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/08/22/third-level-push-modified-siloing-for-deeper-index-penetration.html</guid>
		<description><![CDATA[Third-Level Push (aka &#8220;siloing&#8221;), according to Dan Thies (who regained my attention after his recent article on Google proxy hacking), helps you get third-tier pages (e.g. article/product detail pages) in the main index and ranking higher by &#8220;taking more of the PageRank from your second tier, and pushing it down into the third tier.&#8221;
Dan explains:
In [...]]]></description>
			<content:encoded><![CDATA[<p>Third-Level Push (aka &#8220;siloing&#8221;), according to <a href="http://www.seofaststart.com/">Dan Thies</a> (who regained my attention after his recent article on <a href="http://www.seofaststart.com/blog/google-proxy-hacking">Google proxy hacking</a>), helps you get third-tier pages (e.g. article/product detail pages) in the main index and ranking higher by &#8220;taking more of the PageRank from your second tier, and pushing it down into the third tier.&#8221;</p>
<p>Dan explains:</p>
<blockquote><p>In most sites, your global navigation links to the entire second tier from every page, including the home page. This causes the second tier pages to accumulate a lot of PageRank, at the expense of your third tier.</p></blockquote>
<p>Makes perfect sense. Sites with slightly low link popularity (home page TBPR 3-4) often have no problem getting the home page and most of the category pages in the main index, but they often can&#8217;t get some of the product detail pages to stick. Why? Often its because of exactly what Dan said: the internal navigation makes the home page and second-level pages PageRank-hogs, leaving the third-level pages high and dry.</p>
<p>Some SEOs call Dan&#8217;s tactic &#8220;siloing&#8221;, and attribute its benefits to better themed internal linking. For example, <a href="http://www.bruceclay.com/newsletter/0505/silo.html">Haylie from Bruce Clay talks about siloing</a>, albeit with a focus on ranking, not index penetration. Siloing, in this case, is done by setting up thematic pyramids via links or directory structure. Just imagine a tree hierarchy, where leaf nodes link up to their parent, then a set of parents link up to their parent, and so on, till you reach the root node.</p>
<p>Dan disagrees: &#8220;At the time we all assumed this had something to do with the topics of the pages not being closely related, but we were wrong.&#8221; According to him, increase in site traffic is due to increase in PageRanks at the third-tier.</p>
<p>So how do you implement Third-Level Push? In brief:</p>
<p>1. Use nofollow to prevent second-level pages from passing PageRank to each other. This forces PageRank downwards to the third-level.</p>
<p>2. Use nofollow on links on third-level pages to second-level pages so that a third-level page passes PageRank to its parent page but not to any other pages in the second-tier.</p>
<p>3. <strong>Tiered Pairing</strong>: To prevent second-level pages from losing too much PageRank, you can link them in pairs: e.g. page A with B, C with D, and so on.</p>
<p>4. <strong>Circular Navigation</strong>: To circulate more PageRank on the leaf level, link them up in circular faction, so page A links to B and C, B links to C and D, etc.</p>
<p>That&#8217;s third-level push in a nutshell.</p>
<p><strong>Note:</strong> Avoid deleting or adding links to do this (like I did); instead, just use nofollow. There&#8217;s no bigger sin than compromising user-experience for the sake of SEO (well, there probably is, but lets not get into that).</p>
<h2>A Third Level Push Implementation for Wordpress</h2>
<p>Does third-level push really work? I decided to use this blog as a guinea pig. But how do I implement third-level push on Wordpress? Sure, you can just Google for a &#8220;SEO Siloing&#8221; Wordpress Plugin, but that&#8217;s no fun. <y solution requires a little hacking, but if you're used to PHP it takes only a second. <strong>Warning to the faint of heart</strong>: don&#8217;t try this at home:</p>
<p>1. open template-functions-category.php<br />
2. find function wp_list_cats($args = &#8221;) declaration.<br />
3. around line 236, under parse_str($args, $r): enter:<br />
<strong><br />
if ( !isset($r[&#8217;nofollow&#8217;]))<br />
		$r[&#8217;nofollow&#8217;] = FALSE;</strong></p>
<p>// That sets the default nofollow value, in case no value is passed.</p>
<p>4. Look for the return list_cats.. line.<br />
5. At the end of the long argument list (after $r[&#8217;heirarchical&#8217;]), type <strong>$r[&#8217;nofollow&#8217;]</strong><br />
6. Find function list_cats(&#8230;..).<br />
5. At the end of the function declaration argument list @line 279, after $hierarchical=FALSE, type: &#8220;<strong>, $nofollow=FALSE</strong>&#8221;<br />
6. Now look for the A HREF echo statement, around line 327.<br />
7. Replace $link = &#8216;&lt;a href=&#8221;&#8216;.get_category_link($category->cat_ID).&#8217;&#8221; &#8216;; with:</p>
<p><strong>if($nofollow==FALSE) $link = &#8216;&lt;a href=&#8221;&#8216;.get_category_link($category->cat_ID).&#8217;&#8221; &#8216;;<br />
			else $link = &#8216;&lt;a href=&#8221;&#8216;.get_category_link($category->cat_ID).&#8217;&#8221; rel=&#8221;nofollow&#8221; &#8216;;</strong></p>
<p>8. Finally, open sidebar.php. Look for the wp_list_cats line for single posts (not the home page), around line # 104, that looks like: wp_list_cats(&#8217;sort_column=name&#038;optioncount=1&#038;hierarchical=0&#8242;);</p>
<p>Replace that with <strong>wp_list_cats(&#8217;sort_column=name&#038;optioncount=1&#038;hierarchical=0&#038;nofollow=TRUE&#8217;);</strong></p>
<p>That&#8217;s it.</p>
<p><strong>Potential negative side effects:</strong> If your blog doesn&#8217;t have a lot of backlinks, your category pages might go supplemental. In that case, try Tiered Pairing.</p>
<p><strong>UPDATE:</strong> Joost apparently <a href="http://www.joostdevalk.nl/wordpress-seo-robots-meta-update/">incorporated my idea into his Robots Meta Plugin</a>. Check it out.</p>
<h2>Does Siloing/Third-Level Push Really Work?</h2>
<p>So what happens to PageRank flow after I implement a third-level push?</p>
<p>Here&#8217;s a before and after:</p>
<p><strong>Before:</strong></p>
<p><img src="/images/seo4fun.jpg" alt="before" /></p>
<p>(Pages in the main index are green. Notice I channeled most of my site&#8217;s PageRanks to only those URLs I want to rank, so I had some supplemental URLS but none of them I cared about.)</p>
<p><strong>After:</strong></p>
<p><img src="/images/after.jpg" alt="after" /></p>
<p>Hmm&#8230;so basically I lost PageRank to some of my unpopular category pages. But where did all that PageRank go? To just a handful of recently-published posts, which had high PageRanks to begin with. So it doesn&#8217;t really look like I gained anything, does it? In fact, it looks to me like a whole bunch of pages might go supplemental.</p>
<p>See, there&#8217;s no point in having pages with too much PageRank (at least for getting pages indexed). You want <em>a moderate amount of PageRank on as many pages as possible.</em></p>
<p>Nah, instead I want something more like this:</p>
<p><img src="/images/smooth-curves.jpg" alt="smooth curves" /></p>
<p>Notice now PageRanks are more evenly spread throughout my site.</p>
<p>(screenshots generated by <a href="/php/pagerankbot.php">PageRankBot</a>).</p>
<h2>Sitewides&#8217; Gotta Go (Modified Third-Level Push)</h2>
<p>The problem was I had other sitewide links besides links to category pages, like &#8220;recent posts&#8221;, &#8220;top posts&#8221;, and links to the home page. Those URLs stole the PageRanks the category pages gave up.</p>
<p>In short, <strong>sitewide links are bad</strong>. So what did I do?</p>
<p>1. Dumped sitewide links: Got rid of sitewide links to recent posts and top posts (better to nofollow them but I was in a rush) and nofollowed links to the blog home page. Used third-level push (nofollowed sitewide links to category pages), except I prevented blog articles from linking back to its parent category page to keep the page from accumulating too much PageRank.</p>
<p>2. Added related posts plugin to circulate PageRanks to internal pages &#8220;randomly&#8221; instead of sitewide.</p>
<blockquote><p>
<strong>Aside:</strong> Some SEOs will tell you you should never, ever, ever nofollow links to your internal pages because it sends a negative quality signal to Google. First, Vanessa Fox, an ex-Googler, confirmed that a reason nofollowing internal links may be a bad idea is because the target urls will still be indexed if other people link to them (not that I take her word blindly as Gospel but hey, I don&#8217;t have time to test everything a Googler says). <strong>My policy is to use nofollow on internal links only when I want to control the amount of PageRanks flowing into a URL but I still want to keep the URL in the main index.</strong> For example, I wouldn&#8217;t nofollow a link to my privacy policy or TOS; I would just use robots.txt disallow (sure, robots.txt doesn&#8217;t guarantee that a URL stays out of the main index, but I don&#8217;t care about that; I just don&#8217;t want internal link juice to flow to my TOS page. But if you really wanted to get rid of a URL from Google&#8217;s index, use META noindex instead of robots.txt).</p></blockquote>
<h2>Conclusion</h2>
<p>Will this setup help me or hurt me? Time will tell. The main problem is that now my site&#8217;s PageRanks are unfocused; my top posts aren&#8217;t getting any special attention. I can modify Wordpress so that <strong>X% of links to top posts are nofollowed instead of nofollowing every single sitewide link</strong>. That way, my most important pages will have the highest PageRanks but they won&#8217;t be PageRank hogs.</p>
<p> I do believe that third-level push can work, as long as you have some kind of tool to make sure PageRanks are actually being pushed down to the third-tier pages. Just by nofollowing links to category pages won&#8217;t guarantee that, though preventing sitewide links from flowing juice will probably do the trick.</p>
<p>Why should you care about this stuff?</p>
<p>As <a href="http://www.searchenginejournal.com/google-pagerank-play-doh/5504/#comment-561206">Matt Cutts recently explained</a> (emphasis mine):</p>
<blockquote><p>You could do a similar post with a bunch of Play-Doh and show how you have a certain amount of Play-Doh (your PageRank), and you choose with your internal linking how to spread that Play-Doh throughout your site. <strong>If a given page has enough PageRank</strong> (reasonable-sized ball of Play-Doh), it can be in our main web index. If it has not-very-much PageRank (tiny ball of Play-Doh), it might be a supplemental result. And if only a miniscule iota of Play-Doh makes it to a page, then we might not get a chance to crawl that page.
</p></blockquote>
<p>The Play-Doh / PageRank metaphor is kinda disturbing, but hey, whatever works.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/08/22/third-level-push-modified-siloing-for-deeper-index-penetration.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>How To Exploit The PageRankBot Tool</title>
		<link>http://seo4fun.com/blog/2007/08/08/how-to-exploit-the-pagerankbot-tool.html</link>
		<comments>http://seo4fun.com/blog/2007/08/08/how-to-exploit-the-pagerankbot-tool.html#comments</comments>
		<pubDate>Wed, 08 Aug 2007 17:09:08 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/08/08/how-to-exploit-the-pagerankbot-tool.html</guid>
		<description><![CDATA[Building a good house means more than buying a pine dining table or 1080p Plasma TV (more &#8220;quality&#8221; content) or telling your friends about the new house you&#8217;re building (marketing). You gotta know how to use hammers, drills and nails too.
If you rather build a good site than worry about supplemental results, why are you [...]]]></description>
			<content:encoded><![CDATA[<p>Building a good house means more than buying a pine dining table or 1080p Plasma TV (more &#8220;quality&#8221; content) or telling your friends about the new house you&#8217;re building (marketing). You gotta know how to use hammers, drills and nails too.</p>
<p>If you rather build a good site than worry about supplemental results, why are you reading SEO blogs? Come on, be honest. When&#8217;s the last time you read an SEO blog that talked in-depth about optimizing a dynamic page for fast page loads or repeating graphic elements on a page to create a sense of unity or using element size and position to establish a visual hierarchy?</p>
<p>Never, right?</p>
<p>But if you&#8217;re a control freak like me, read on. </p>
<p><H2>WTH Does It Do?</h2>
<p>Though some of you guys gave me positive feedback via comments and email about <a href="/php/pagerankbot.php">PageRankBot</a>, I&#8217;m not sure if all of you know exactly what to do with it.</p>
<p>Inspite of the misleading name &#8220;Supplemental Results Detector&#8221;, its not a tool for detecting supplemental results. You have site:www.domain.com/&#038; and site:www.domain.com/* for that. There are also other tools out there (I think Aaron Wall has one and <a href="http://www.sitemost.com.au/supplemental-results.php">sitemost just came out with a new tool</a>).</p>
<p>I don&#8217;t really care how many of my pages are supplemental, but I do care when a page that deserves to rank in the SERP goes supplemental. One way to address that problem is PageRank distribution management. That&#8217;s what I built this tool for.</p>
<h2>Tactics</h2>
<p>First, figure out which pages on your site are important and which pages aren&#8217;t. Ask yourself <em>is this page valuable to my visitors?</em> If the answer is no, the page can go. You might also ask yourself <em>what is this page supposed to rank for?</em> If the answer is &#8220;contact me&#8221; or &#8220;privacy policy&#8221; then ask yourself <em>why the hell would I want traffic for &#8220;privacy policy&#8221; and am I out of my mind thinking I can rank on the first page for &#8220;privacy policy&#8221; alongside Google, Sun, Apple, Adobe, and NY Times?</em></p>
<p>But if your &#8220;contact me&#8221; page contains your email address or IM information and your clients find you by Googling for your contact info, I would keep the page in the main index.</p>
<p>To mark unimportant URLs, multi-select URLs that are unimportant, then Edit > Toggle Importance.</p>
<p><img src="/images/toggle-importance.jpg" alt="" /></p>
<p>Now flag supplemental URLs. Some of you wish the tool does this for you automatically. It doesn&#8217;t. Instead, label URLs returned by site:www.domain.com* command by going to Edit > Mark Page As > Main Index.</p>
<p><img src="/images/mark-as-supp.jpg" alt="" /></p>
<p>You can use the search tool to find URLs. For example, the following image shows a search on seo4fun.com for urls containing the word &#8220;pagerank&#8221;:</p>
<p><img src="/images/search.jpg" alt="supplemental results tool search feature" /></p>
<p>Now go to View > Filters > Hide Marked, which hides all the URLs you just marked. Select all the URLs you see, and then set their status to supplemental.</p>
<h2>Find Your Link Targets</h2>
<p>To manage internal PageRank flow, you add internal links to your site. Decide which page you&#8217;re going to add a link on (link source) and which page you want that link to point to (link target).</p>
<p>To fish out your link &#8220;targets&#8221;, view only supplemental pages and sort them by PageRank (View > Filters > Show Supplementals and then click on the PageRank column). The topmost URL marked &#8220;important&#8221; is your best candidate:</p>
<p>1. The page is important to you (you feel the page deserves to rank in the SERPs).<br />
2. The page is supplemental.<br />
3. The page with the highest PageRank = easiest url to pull back into the main index.</p>
<p><img src="/images/supps-by-pagerank.jpg" alt="Supplemental results sorted by pagerank" /></p>
<p>There&#8217;s your link &#8220;target.&#8221;</p>
<p><strong>Note:</strong> If your site has multiple &#8220;entry points&#8221; (i.e. not all inbounds point to the home page), PageRank flowing into your site from those entry points will change the dynamics of how PageRank is distributed. In that case, take the PageRank values this tool gives you with a grain of salt.</p>
<p>If you&#8217;re anal enough to want to account for IBLs pointing at specific pages, then you can &#8220;add juice&#8221; by going to Tools > Simulate Backlinks. First, set the home page TBPR (use a float, like 4.2 for more accuracy). Go to View > Column Filters > Approximate TBPR. That will show you approximate TBPR numbers translated from raw PageRank numbers. Choose a URL, and adjust as needed using the + and - keys.</p>
<p><img src="/images/add-juice.jpg" alt="add juice" /></p>
<h2>Find Your Link Sources</h2>
<p>There&#8217;s a few ways to figure out your link &#8220;sources.&#8221; One way is to find the page with the most PageRank bleed. (Don&#8217;t believe PageRank bleeds? We&#8217;ll argue about that in another post). Amount of PageRank bleed depends on percentage of outbounds to inbounds and a URL&#8217;s (non-visible) PageRank. For example, a PageRank X URL with two outbound links and two internal links would bleed (X/4)*2 PageRank. Bigger X (increased number/quality of IBLs pointing to a URL) means more PageRank bleed. More internal links means less PageRank bleed, even if the number of outbound links stay the same.</p>
<p>Let&#8217;s not get too obsessed with PageRank bleeds though. You can solidify Google&#8217;s trust in your links by linking out organically. A site that doesn&#8217;t link out needs a strong set of credible, trusted IBLs to &#8220;validate&#8221; with Google (e.g. amazon.com). Consider your outbound links a part of your link profile and a key ingredient in proving to Google that your linking habits are 100% natural with no artificial colors, flavors or sweeteners (yeah, I know that was bad).</p>
<h2>Link from Pages with the Highest PageRank Bleed</h2>
<p>First, limit results to URLs in the main index by going to View > Filter > Main Index Only, so you only link from URLs in the main index. Then sort by Outbound PageRank (click on the &#8220;Outbound PageRank&#8221; column header. If you don&#8217;t see the colum displayed, go to View > Column Filter to activate). The topmost URL with the biggest outbound PageRank is your link &#8220;source.&#8221;</p>
<p><img src="/images/outbound-pagerank.jpg" alt="outbound pagerank" /></p>
<h2>Link from Pages that Flow the Most PageRank</h2>
<p>Another way is to find a page with that flows the most PageRank with each link. Go to View > Filter > Main Index Only. Then click on the &#8220;Increment&#8221; column header, which sorts the result in the order of PageRank flowing per link. The topmost URL with the biggest Increment bar is your link &#8220;source.&#8221;</p>
<p><img src="/images/increment.jpg" alt="pagerank increment" /></p>
<h2>Connect the Dots</h2>
<p>Finally, point a link from your link source to your link target.</p>
<p>If your modification isn&#8217;t sitewide, select the URL you just updated and recrawl that URL only instead of recrawling the entire site to update the site&#8217;s PageRanks.</p>
<p><img src="/images/recrawl.jpg" alt="recrawl url feature" /></p>
<p>You can also try flattening out your site&#8217;s PageRank curve (see the two graphs in my previous post about <a href="/blog/2007/08/01/how-google-failed-to-hide-supplemental-results.html">Google hiding supplemental results</a>).</p>
<h2>Take It Slow</h2>
<p>If your site has enough PageRank, Google should update your pages in the main index every 3-4 days, if not sooner (e.g. if you show up for Google News) - though dramatic on-page edits like rewriting a TITLE tag might make Google sit on a page for a week or two. It should take you no more than 3 days to get a URL out of the supplemental index, as long as you have enough URLs in the main index to play around with.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/08/08/how-to-exploit-the-pagerankbot-tool.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>How Google Failed to Hide Supplemental Results</title>
		<link>http://seo4fun.com/blog/2007/08/01/how-google-failed-to-hide-supplemental-results.html</link>
		<comments>http://seo4fun.com/blog/2007/08/01/how-google-failed-to-hide-supplemental-results.html#comments</comments>
		<pubDate>Wed, 01 Aug 2007 08:48:34 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/08/01/how-google-failed-to-hide-supplemental-results.html</guid>
		<description><![CDATA[If you&#8217;re an SEO with clients that are worried about supplemental results, your job just got a whole lot harder. It&#8217;s like having a patient dying of disease showing no visible symptoms. Not only does he believe he isn&#8217;t sick anymore, but you can&#8217;t tell what he&#8217;s sick of.
First, your clients should know that just [...]]]></description>
			<content:encoded><![CDATA[<p>If you&#8217;re an SEO with clients that are worried about supplemental results, your job just got a whole lot harder. It&#8217;s like having a patient dying of disease showing no visible symptoms. Not only does he believe he isn&#8217;t sick anymore, but you can&#8217;t tell what he&#8217;s sick of.</p>
<p>First, your clients should know that just because they don&#8217;t see the supplemental results label anymore, it doesn&#8217;t mean their worries are over. Their supplemental pages are still supplemental. Google is just trying to hide the fact.</p>
<p>They should also know that just because the label is gone doesn&#8217;t mean you can&#8217;t detect supplemental results. You can, and here&#8217;s how:</p>
<ul>
<li><strong>site:www.domain.com/&#038; hack</strong>, which seems to pull up urls that used to be labeled supplemental. Of course now that <a href="http://searchengineland.com/070731-215828.php">Danny Sullivan blogged about it</a>, that hack probably won&#8217;t last another week. (UPDATE: The hack was covered last week, according to Danny, the same week I pulled almost all my SEO feeds from Google Reader so I&#8217;m not bombarded by SEO news. Bad timing, I guess)</li>
<li><strong>site:www.domain.com/*</strong> shows pages in the main index.</li>
<li><strong>Old cache date.</strong> If s page&#8217;s cache date is old, its a sign that the page may be supplemental. Why? Because Google doesn&#8217;t refresh a supplemental result&#8217;s cache all that often. For example, my blog&#8217;s main urls have cache dates ranging from Jul 25~26, 2007 (today&#8217;s date: Aug 1, 2007) while my old supplemental pages have cache dates as old as Jul 6-7, 2007.</li>
<li><strong>Low-to-none competitive term traffic.</strong> If you&#8217;re not getting Google hits for two-word queries or getting no traffic at all to a specific URL, it may be supplemental.</li>
<li><strong>Uneven PageRank distribution</strong>, which you can control by downloading the <a href="/php/pagerankbot.php">Supplemental Results Detector Tool</a>. See how sugarrae.com and seo4fun.com distributes PageRank?<br />
<img src="/images/sugarrae.jpg" alt="sugarrae pagerank distribution" /><br />
See how sugarrae&#8217;s PageRank distribution is pretty even, so that there isn&#8217;t a huge gap between the home page and the deep pages? The site is 99% supplemental results free. Yeah, its a high TBPR site with only ~100 pages (which means plenty of PageRank to go around for each page) but so is vanessafox.com (TBPR 7 with ~100 pages), which has more supplemental results than sugarrae.com.<br />
<img src="/images/seo4fun.jpg" alt="seo4fun pagerank distribution" /><br />
(orange urls are supplemental)<br />
In contrast, seo4fun.com concentrates PageRank on just a handful of pages while the rest of the site gets very little attention. Consequently, some of the urls near the bottom of the chart with low link popularity are supplemental.
</li>
<li><strong>Low Toolbar PageRank</strong> (0 ~ 3). The toolbar is a weak indicator due to delay but more green generally means less chance of a page being supplemental.</li>
</ul>
<p>TANGENT:</p>
<p>(One interesting tidbit I found in <a href="http://googlewebmastercentral.blogspot.com/2007/07/supplemental-goes-mainstream.html">the recent Google blog post</a> is Matt/Prashanth Koppula saying a url with complicated query strings also might go supplemental. At this point (considering the fact that Dave said stale pages go supplemental as well) it&#8217;s probably safe to assume a myriad of minor factors are involved)</p>
<p>(After reading the post, reasons why Google likes supplemental results are pretty clear:</p>
<p>1. Crawl the web more fully to serve ~1000 results (or maybe Google&#8217;s satisfied with just 10-100) for every possible search query, which means a) a happier user and b) more pages to display AdWords on.<br />
2. Improve efficiency by taking advantage of prioritized crawling: crawl important, frequently updated pages more often while crawl less important, unupdated pages less frequently. Unfortunately, this often means only home page/top-level nav pages get indexed while pages with actual content fails to make it into the main index. I often get frustrated by a search result that lands me on a blog category page with 40+ blog post links instead of the blog post itself.)</p>
<p>Wrapping up:</p>
<p>A site with many pages in the main index receive traffic for competitive two-word queries. Traffic land on not just a handful of pages but on thousands of pages. Googlers promise that, by the end of the summer, supplemental results will generate more traffic and will rank for more terms. We&#8217;ll see. There are a ton of spam pages in the supplemental index, so Google will have to walk a thin line - otherwise odd query terms will be swamped with low PageRank spam while legitimate supplemental results never see the light of day.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/08/01/how-google-failed-to-hide-supplemental-results.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Hey, it Just Sounds to Me Like You Need to Unplug, Man.</title>
		<link>http://seo4fun.com/blog/2007/06/22/hey-it-just-sounds-to-me-like-you-need-to-unplug-man.html</link>
		<comments>http://seo4fun.com/blog/2007/06/22/hey-it-just-sounds-to-me-like-you-need-to-unplug-man.html#comments</comments>
		<pubDate>Fri, 22 Jun 2007 14:50:49 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/06/22/hey-it-just-sounds-to-me-like-you-need-to-unplug-man.html</guid>
		<description><![CDATA[
Image courtesy of What Is the Matrix
If you spend hours a day surfing blogs, you&#8217;re trapped in the Matrix. Being an ex-hardcore MMORPG addict makes me an expert on getting sucked in. In that world, I met thousands of people,  I slew dragons, I saved lives.
Meanwhile, in the real world, I sat in front [...]]]></description>
			<content:encoded><![CDATA[<p><img src="/images/reloaded_01.jpg" alt="Matrix Reloaded Neo and Trinity" /></p>
<p><strong style="font-size:11px;">Image courtesy of <a href="http://whatisthematrix.warnerbros.com/">What Is the Matrix</a></strong></p>
<p>If you spend hours a day surfing blogs, you&#8217;re trapped in the Matrix. Being an ex-hardcore MMORPG addict makes me an expert on getting sucked in. In that world, I met thousands of people,  I slew dragons, I saved lives.</p>
<p>Meanwhile, <em>in the real world</em>, I sat in front of my computer for hours tapping on my keyboard and staring into my monitor.</p>
<blockquote><p>I know why you&#8217;re here, Neo. I know what you&#8217;ve been doing&#8230; why you hardly sleep, why you live alone, and why night after night, you sit by your computer.</p></blockquote>
<p>Living in a <a href="http://www.newswise.com/articles/view/505712/">cacooning world</a>, you give in to your urge to socialize - to connect to other people, to not feel alone, to feel important, to feel wanted.</p>
<p>Like throwing a rock into a lake and seeing ripples on the water or shouting at a mountain to hear the echoes of your voice, you feel a need to confirm your own existence.</p>
<p>But does reading about Google policing paid links help you get your laundry done? Does knowing that Yahoo redesigned its home page take care of your phone bills? Does leaving a comment about why Jason Calanis is wrong pay for your baby&#8217;s diapers?</p>
<blockquote><p>You know, I know this steak doesn&#8217;t exist. I know that when I put it in my mouth, the Matrix is telling my brain that it is juicy and delicious. After nine years, you know what I realize?</p></blockquote>
<p>Ignorance is bliss.</p>
<p>When I wake up in the morning, I instinctively fire up Google Reader to jump-start my brain. Since I&#8217;m my own boss, sometimes, that quick peek turns into hours of reading and commenting and posting and I get no work done.</p>
<p>Here&#8217;s a thought de jour:</p>
<p>You already know enough. </p>
<p>&#8220;Hey, it just sounds to me like you need to unplug, man. You know, get some R and R&#8221; - The Matrix</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/06/22/hey-it-just-sounds-to-me-like-you-need-to-unplug-man.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>JDBC ClassNotFoundException (NetBeans, Classpath, Java)</title>
		<link>http://seo4fun.com/blog/2007/06/16/jdbc-classnotfoundexception-netbeans-classpath-java.html</link>
		<comments>http://seo4fun.com/blog/2007/06/16/jdbc-classnotfoundexception-netbeans-classpath-java.html#comments</comments>
		<pubDate>Sat, 16 Jun 2007 12:48:20 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Coding]]></category>

		<category><![CDATA[Java]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/06/16/jdbc-classnotfoundexception-netbeans-classpath-java.html</guid>
		<description><![CDATA[If you get a java.lang.ClassNotFoundException error when loading a database driver using the statement:
Class.forName({nameOfYourDriverWhateverItIs}).newInstance();
You can either:
Set CLASSPATH in DOS
Go into DOS (Start/Run/cmd.exe):
set CLASSPATH=.;{pathToYourJarFile}
For example, if your jar file is at: C:/Program
Files/java/jdk1.6.0_01/lib/mysql-connector-java-5.0.6-bin.jar,
Type:
set CLASSPATH=.;C:/Program Files/java/jdk1.6.0_01/lib/mysql-connector-java-5.0.6-bin.jar
Now,
javac YourJavaFile.java
java YourJavaFile
That&#8217;s all. But it won&#8217;t work if you&#8217;re trying to run code in Netbeans.
Set Your Project&#8217;s Classpath in Netbeans
If you&#8217;re using [...]]]></description>
			<content:encoded><![CDATA[<p>If you get a java.lang.ClassNotFoundException error when loading a database driver using the statement:</p>
<p>Class.forName({nameOfYourDriverWhateverItIs}).newInstance();</p>
<p>You can either:</p>
<h2>Set CLASSPATH in DOS</h2>
<p>Go into DOS (Start/Run/cmd.exe):</p>
<p>set CLASSPATH=.;{pathToYourJarFile}</p>
<p>For example, if your jar file is at: C:/Program<br />
Files/java/jdk1.6.0_01/lib/mysql-connector-java-5.0.6-bin.jar,</p>
<p>Type:</p>
<p>set CLASSPATH=.;C:/Program Files/java/jdk1.6.0_01/lib/mysql-connector-java-5.0.6-bin.jar</p>
<p>Now,</p>
<p>javac YourJavaFile.java<br />
java YourJavaFile</p>
<p>That&#8217;s all. But <strong>it won&#8217;t work if you&#8217;re trying to run code in Netbeans.</strong></p>
<h2>Set Your Project&#8217;s Classpath in Netbeans</h2>
<p>If you&#8217;re using netbeans, <a href="http://www.netbeans.org/kb/41/using-netbeans/project_setup.html#manageclasspath">set your project&#8217;s classpath</a>.</p>
<p>1. Right-click on your project.<br />
2. Select &#8220;Properties.&#8221;<br />
3. Click &#8220;Libraries.&#8221;<br />
4. Click &#8220;Add JAR/Folder&#8221;<br />
5. Choose your MYSQL driver JAR file.</p>
<p><b>HEY. Did you find this post useful as hell? </b> I know you did. Then link to this page, Stumble it, Del.icio.us it, do whatever you can to pump this page up the Google SERPs so other people looking for this info can find it easier. Thanks.</p>
<h2>Why This Has Anything to Do With SEO</h2>
<p><img style="float:right;margin:7px;" src="/images/java-s.jpg" alt="Java netbeans classnotfoundexception" /></p>
<p>It took me over 30 minutes to find this solution. I was <em>this close</em> to throwing in the towel. So once I figured out the solution, I wrote a page I wish I found at the top of Google&#8217;s search results when I looked for &#8220;jdbc classnotfoundexception.&#8221; That way, other people looking for the same information won&#8217;t be frustrated like I was. Because I&#8217;m working in a long tail space, to get this page to rank high I just optimized the content. But I didn&#8217;t keyword spam or pepper my H tags with related terms. Instead, I optimized for <em>you</em>.</p>
<p>Yep, you.</p>
<p>Ok, so it&#8217;s not perfectly optimized for you if you&#8217;re an RSS subscriber of mine, since you&#8217;re interested in SEO, not Java. But if you found this page through Google, you&#8217;re thanking me now because:</p>
<ul>
<li><em>I give you exactly what you were looking for.</em></li>
<li><em>I get straight to the point.</em> Instead of starting off with a long irrelevant opening paragraph, I explain the problem and give you the answer. <strong>Instant gratification baby</strong>.</li>
<li><em>I keep it short.</em> I use as few words as I can to save you time.</li>
<li><em>I keep it simple.</em> Instead of trying to crack you up with dumb jokes or titilate you with fancy metaphors, I use simple words so even a caveman can &#8220;get it.&#8221;</li>
</ul>
<p>Some people say ranking is about links, not content. Is it better to be tall and rich, or loyal and charismatic? Why are marketers trying to convince you its all about looks? (Yeah, just think about that for a minute :D)</p>
<p>Given two websites with equal visibility, the site that publishes the most compelling content will always win, showering the site with even more visibility. If marketing is all it takes, Paris Hilton would have sold more CDs.</p>
<h2>Related Articles:</h2>
<p>A few articles that both helped me and frustrated me (hint: natural outbound links to relevant, authoritative pages will help you win Google&#8217;s trust):</p>
<ul>
<li><a href="http://java.sun.com/docs/books/tutorial/jdbc/basics/connecting.html">Establishing a Connection (The Java Tutorials)</a></li>
<li><a href="http://dev.mysql.com/doc/refman/5.0/en/connector-j-usagenotes-basic.html#connector-j-examples-connection-drivermanager">Basic JDBC Concepts</a></li>
<li><a href="http://java.sun.com/docs/books/tutorial/jdbc/basics/index.html">Lesson: JDBC Basics</a></li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/06/16/jdbc-classnotfoundexception-netbeans-classpath-java.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Got Supplementals? Accepting PageRank is Only The Beginning</title>
		<link>http://seo4fun.com/blog/2007/06/07/got-supplementals-accepting-pagerank-is-only-the-beginning.html</link>
		<comments>http://seo4fun.com/blog/2007/06/07/got-supplementals-accepting-pagerank-is-only-the-beginning.html#comments</comments>
		<pubDate>Thu, 07 Jun 2007 15:28:34 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/06/07/got-supplementals-accepting-pagerank-is-only-the-beginning.html</guid>
		<description><![CDATA[
Nowadays, supplemental results aren&#8217;t much of a mystery anymore to most people. As Matt Cutts replied to Michael Martinez at Seattle SMX, people know all they need to do is to get more &#8220;quality links.&#8221; The answer is so simple, so easily digestible that I&#8217;m starting to see people answer supplemental threads with just three [...]]]></description>
			<content:encoded><![CDATA[<p><img style="float:left;margin:10px;" src="/images/supplemental-pizza.jpg" alt="supplemental results pizza" /></p>
<p>Nowadays, supplemental results aren&#8217;t much of a mystery anymore to most people. As <a href="http://seo-theory.com/wordpress/2007/06/05/seo-theory-and-smx-advanced-2007/">Matt Cutts replied to Michael Martinez at Seattle SMX</a>, people know all they need to do is to get more &#8220;quality links.&#8221; The answer is so simple, so easily digestible that I&#8217;m starting to see people answer supplemental threads with just three words - <em>PageRank</em> and <em>duplicate content</em>. As if.</p>
<p>What makes you gain weight? <em>Eating too much.</em> Duh. Knowing that doesn&#8217;t help you much, does it? How can you lose weight and keep it off? Knowledge isn&#8217;t power. <strong>Actionable knowledge</strong> is power.</p>
<h2>Understanding Supplemental Results and PageRank Distribution</h2>
<p>You have countless internal-link-based tactics at your disposal to combat supplemental results. In a thread on WMW titled <a href="http://www.webmasterworld.com/google/3352115.htm">Supplemental Page Count Formula?</a>, I summarized:</p>
<blockquote><p>If your site is largely supplemental, it means 1) not enough quality inbound links to your site 2) you have too many pages 3) you link out too much 4) Google may think your IBLs are artificial 5) Cannonical issues are causing PageRank to split.</p></blockquote>
<p>Bouncybunny replies:</p>
<blockquote><p> I&#8217;ve never heard points 2 + 3 being relevant for pages falling into the supplementals. </p></blockquote>
<p>I&#8217;ll leave the discussion about PageRank leaks for another day. As for point 2, I explained (in geek speak):</p>
<blockquote><p>Think of total PageRank X (sum of all inbound PageRank to your domain) split between Y number of pages. Roughly speaking, bigger page count = lower average PageRank per page (depending on your site structure). We know that a page with PageRank below minimum threshold &#8220;goes&#8221; supplemental. With excessively high page count, average falls too low, and you&#8217;ll end up with many pages in the supplemental index. By reducing the number of pages, you slightly increase average PageRank per url. That can result in several supp pages popping back into the main index.</p></blockquote>
<p>As a matter of fact, Shoemoney claims he got rid of some of his supplemental results by following Aaron Wall&#8217;s advice: robots.txt disallow noisy pages.</p>
<h2>Don&#8217;t Cut Up Your Pizza Into 60,000 slices if You Ordered a Small Pie</h2>
<p>Thanks to Andy Beal, <a href="http://www.marketingpilgrim.com/2007/06/smx-video-matt-cutts-explains-how-to-get-out-of-googles-supplemental-index.html">you can hear Matt Cutts</a> say basically the same thing:</p>
<blockquote><p>
If you got 60,000 pages, and you only got &#8220;this much&#8221; PageRank, and you divide it [&#8230;he mumbles], some of them are going to be in the supplemental index. Given &#8220;this many people&#8221; who link to you, we&#8217;re willing to include &#8220;this many&#8221; pages in the main index.</p></blockquote>
<p>The picture below shows you how PageRank is distributed on this domain, assuming an artificial scenario where all inbound links are ignored:</p>
<p><a href="/images/pagerank-distribution-suppl.jpg"><img src="/images/supplemental-results-tiny-t.jpg" alt="supplemental results pagerank distribution" /></a></p>
<p>(click on the thumbnail to see the details)</p>
<blockquote><p><strong>Aside:</strong> I generated this chart using Google Docs, Photoshop, and my <a href="/php/pagerankbot.php">Supplemental Results Detector</a>, which a few of you guys might remember I wrote back in December 2006. It&#8217;s a simple script that emulates Google&#8217;s PageRank iteration. Though it ignores inbound links and uses the original PageRank formula where all PageRanks add up to the total number of pages on a domain instead of 1, it&#8217;s a pretty good indicator of which pages on your site is prone to go supplemental. It&#8217;s a gift-from-above for PageRank horders, but at the same time, it helps you organize your internal links more strategically.</p></blockquote>
<p>Notice:</p>
<ul>
<li><strong>More inbound links = bigger pie.</strong> As you guys link to me more often (hint hint), my pizza gets bigger, which makes all my slices grow. Phatter slices mean fewer supplemental results.</li>
<li><strong>More artificial links = smaller pie.</strong> Reciprocal links, cheap directory links, link injections,easily detectable paid links - all of these work to some extent if you play it like <a href="http://en.wikipedia.org/wiki/Sam_Fisher">Sam Fisher</a>. Uline.com ranks on the first page for &#8220;cardboard boxes&#8221;, Thisisouryear ranks first for &#8220;website directory&#8221;, and customermagnetism ranks 13th for &#8220;search engine optimization&#8221; all thanks in part to paid links. Major adult sites also dominate competitive porn terms using hundreds of thousands of reciprocal links. But having tons of artificial links pointing at your site makes them easier for Google to detect. If Google decides to devalue the PageRanks passed by those links, your large pizza turns into a medium, sometimes causing your site to enter the realm of <a href="http://www.seroundtable.com/archives/013340.html">Google Hell</a>.</li>
<li><strong>More pages = smaller slices.</strong> As I publish more posts, I create more slices, causing all my slices to shrink. What happens to a commercial site that publishes 100,000 new pages in one day? Or what about a blogger that publishes 10 posts a day but gets completely ignored by the linkerati? They often go 99% supplemental because you&#8217;re adding a ton of more slices while the pie stays the same size. More slices mean smaller slices. And if they&#8217;re too small, they &#8220;go supplemental.&#8221; But if I publish something useful, people will link to me, increasing the size of my pie.</li>
<li><strong>Fewer pages = bigger slices.</strong> Conversely, deleting pages causes the size of my other slices to grow. It&#8217;s like a page &#8220;taking one for the team.&#8221;</li>
</ul>
<h2>Move Away from Default Wordpress Setup for Better PageRank Flow</h2>
<p>The chart also gives you an idea of how the default Wordpress template distributes PageRank. The blog home page gets the most love; category pages are second in line; recent posts are third in line. A second page of your category pages (e.g. /category/seo/2/) and old posts have the least internal link juice flowing into them by default.</p>
<p>If you use the Recent Posts Plugin, your recently published post gets love from the blog home page, first page on your category and archive pages, and every single posts page. If you use install a Top Posts Plugin, you can direct extra juice (and traffic) to your favorite posts (see <a href="http://www.jimboykin.com/">Jim Boykin&#8217;s blog</a>, though the way he has it set up is kinda fugly). The Related Posts plugin can help maintain internal linkage between old posts.</p>
<h2>Internal Linking Tactic is Half the Battle</h2>
<p>As Adam Lasnik would probably tell you, the best cure for supplemental results is to create original, compelling content, marketing it, and getting links you deserve. But you can also get mileage out of working with the PageRank you already have. Think of it like tweaking one of your landing pages to improve your CTR. A 2% increase can add up to alot of money. Remember, people used to believe duplicate text caused supplemental results. But <strong>its duplicate urls creating more slices than you need</strong> that&#8217;s partly to blame, as <a href="http://www.vanessafoxnude.com/2007/06/06/buffy-in-duplicate/">Vanessa Fox recently confirmed </a> on her blog:</p>
<blockquote><p>Does having duplicate content cause sites to be placed there? Nope, that’s mostly an indirect effect. If you have pages that are duplicates or very similar, then <strong>your backlinks are likely distributed among those pages</strong>, so your PageRank may be more diluted than if you had one consolidated page that all the backlinks pointed to. And lower PageRank may cause pages to be supplemental.</p></blockquote>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/06/07/got-supplementals-accepting-pagerank-is-only-the-beginning.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>How Will People Find Your Site If Search Engines Didn&#8217;t Exist?</title>
		<link>http://seo4fun.com/blog/2007/05/22/how-will-people-find-your-site-if-search-engines-didnt-exist.html</link>
		<comments>http://seo4fun.com/blog/2007/05/22/how-will-people-find-your-site-if-search-engines-didnt-exist.html#comments</comments>
		<pubDate>Tue, 22 May 2007 15:55:17 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/05/22/how-will-people-find-your-site-if-search-engines-didnt-exist.html</guid>
		<description><![CDATA[Summary (for people who don&#8217;t have time to read blogs all day. We should be building stuff instead of bitching about Google, yeah?): If you depend on Google to survive, your site sucks. If your site gets the same level of traffic from Google a day but your daily visits isn&#8217;t rising, your site sucks. [...]]]></description>
			<content:encoded><![CDATA[<p><strong>Summary</strong> (for people who don&#8217;t have time to read blogs all day. We should be building stuff instead of bitching about Google, yeah?): If you depend on Google to survive, your site sucks. If your site gets the same level of traffic from Google a day but your daily visits isn&#8217;t rising, your site sucks. For big companies, SEO is just an afterthought. Universalstudios.com has a few SEO flaws, but it doesn&#8217;t matter.</p>
<h2>Big Companies Are Already Highly Visible</h2>
<p>In the white hat world, you need to develop a valuable product and then launch a marketing campaign to bring eyeballs to your product. If your a newbie webmaster with a domain name no one&#8217;s ever heard of, you have a long road ahead of you. First, you need to build a product to sell (aka. website). Pour millions of visitors on a piece of crap and your conversion ratios will look worse than plentyoffish and traffic will go right through you like you weren&#8217;t even there. Once your done developing a great product, you need to go on a marketing blitz, because the greatest website in the world will sound like a tree that fell in a forest if no one knows your site exists.</p>
<p>On the other hand, Universal Studios, MGM, McDonalds, IBM, Apple &#8230; these companies existed long before the birth of search engines. These companies have access to mass media (TV news, commercials, billboards, radio, newspapers, magazines, movies). They branded their names permanently into our collective psyche. Even if Google didn&#8217;t exist, people know how to find them on the web.</p>
<h2>Google is Just a Middle Man</h2>
<p>When do you need Google? You need Google when you&#8217;re looking for something but don&#8217;t know where to find it. But as time goes on, you figure out where to find whatever you&#8217;re looking for, so that Google becomes a middle man you no longer need. If you were looking for <a href="http://www.nbc.com/Heroes/bios/niki.shtml" rel="nofollow">Ali Larter</a>&#8217;s pics, for example, <a href="http://images.google.com/images?svnum=10&#038;um=1&#038;hl=en&#038;safe=off&#038;client=firefox-a&#038;rls=org.mozilla%3Aen-US%3Aofficial&#038;hs=2P&#038;q=%22ali+larters%22&#038;btnG=Search+Images">you might first surf Google Images</a>. If you were looking for <a href="http://www.imdb.com/name/nm0005123/" rel="nofollow">a list of movies she starred</a> in, you go to imdb. If you wanted to know <a href="http://en.wikipedia.org/wiki/Asia_Carrera" rel="nofollow">the names of Asia Carrera&#8217;s kids</a> or if she&#8217;s still having financial problems, you either go to <a href="http://www.asiacarrera.com/" rel="nofollow">her official website</a> or go to Wikipedia. If you wanted to buy a DVI adapter or a new PSU, you go to newegg. If you wanted to download MTV videos on YouTube that might disappear by tomorrow. you visit keepvid. If you felt like <a href="http://twitter.com/vanessafox">stalking Vanessa Fox</a>, you&#8217;d go to twitter (Vanessa, don&#8217;t worry. I&#8217;m too busy to stalk anyone :D).</p>
<p>See? If you know where to find what you&#8217;re looking for, you don&#8217;t really need Google, do you? (unless you&#8217;re thinking of sites like Webmasterword with lousy search features)</p>
<h2>For Big Brand Names, SEO is An Afterthought</h2>
<p>For companies like Universal Studios, SEO is almost an afterthought. The bulk of their &#8220;marketing&#8221; that&#8217;s been going on for decades penetrates households that don&#8217;t even own a computer or don&#8217;t have enough money to pay for internet access.</p>
<p>If you already got high visibility, all you really need to do then is build a razzle-dazzle 59 points out of 60 website that <strong>turns your visitors into marketers</strong>.</p>
<h2>Building a Website with Compounding Traffic</h2>
<p>Have you ever lost thousands of daily uniques because your Google ranking suddenly  tanked?</p>
<p>You have? Ok. But what happened to 30,000 people that visited your site last week?</p>
<p>If your site&#8217;s stickier than <a href="http://en.wikipedia.org/wiki/Glue" rel="nofollow">cyanoacrylate</a>, 100 daily uniques from Google would pile up into 1,000 visitors/day after 10 days. Do you see that happening with your site? If your site&#8217;s traffic isn&#8217;t rising, you need to work on content, not SEO. I&#8217;m not just talking about building more pages. I&#8217;m talking about injecting value into your site, making it amazing, mind-blowing, unforgettable. Sure, adding more pages targeting more phrases might increase number of incoming daily hits, but what good does a million pennies in your pocket do you if you have a gaping hole in your pocket?</p>
<p>In investing, you look for high interest rates that compound year after year after year. Investing $4K a year in a Roth IRA at 8% average interest rate, for example, will turn you into a millionare in 30 years all thanks to compounding interest.</p>
<p>Most people know this, yet they often don&#8217;t put much effort into building a website with <strong>compounding traffic</strong>.</p>
<h2>Universal Studios Can&#8217;t Care Less About SEO</h2>
<p>Universal Studios obviously hasn&#8217;t bothered to SEO their official website, universalstudios.com:</p>
<ul>
<li><a href="http://www.universalvod.net/index.html?__source=USMN.GNAV" rel="nofollow">100% Flash pages like this</a> makes some SEOs frown. But see, it&#8217;s a dichotomy: on the one hand, people say SEO is 99.9% about links. They also say all Flash and no text is bad SEO. Makes you wonder if people spreading these ideas are capable of logical thinking.
</li>
<li><a href="http://www.universalstudios.com/" rel="nofollow">Home page redirects</a> to index.php. That&#8217;s not necessarily bad if it was a 301 redirect. But it&#8217;s a 302.</li>
<li>The <a href="http://www.google.com/search?hl=en&#038;safe=off&#038;client=firefox-a&#038;rls=org.mozilla%3Aen-US%3Aofficial&#038;hs=naU&#038;q=site%3Ahttp%3A%2F%2Fwww.universalstudios.com%2F&#038;btnG=Search" rel="nofollow">second site: search result</a> triggers a 404.</li>
<li>According to the home page META keywords, the home page wants to rank for matt damon,ppv, vod, on demand, pay-per-view, jerry springer. Despite the TBPR 7, universalstudios.com ranks for none of those words because the IBL anchor texts are untargeted. According to SEO Digger, the home page does rank for over 637 terms.</li>
</ul>
<p>It doesn&#8217;t matter though. If Google vanished tomorrow, people will still visit universalstudios.com. What will happen to your site if Google vanished tomorrow?</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/05/22/how-will-people-find-your-site-if-search-engines-didnt-exist.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Stop Blowing Money on Text Link Ads</title>
		<link>http://seo4fun.com/blog/2007/05/14/stop-blowing-money-on-text-link-ads.html</link>
		<comments>http://seo4fun.com/blog/2007/05/14/stop-blowing-money-on-text-link-ads.html#comments</comments>
		<pubDate>Mon, 14 May 2007 16:20:03 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/05/14/stop-blowing-money-on-text-link-ads.html</guid>
		<description><![CDATA[Not all paid links (don&#8217;t forget to read the update) are easy to detect, but if you think Google can&#8217;t detect blatant schemes like Text Link Ads you&#8217;re dreaming. They stick out like a geek on American Idol because:

Anchor text targets money terms. Anchor text like &#8220;buy viagra&#8221;, &#8220;car insurance&#8221;, and &#8220;search engine optimization&#8221; are [...]]]></description>
			<content:encoded><![CDATA[<p>Not all <a href="http://www.mattcutts.com/blog/how-to-report-paid-links/">paid links</a> (don&#8217;t forget to read the update) are easy to detect, but if you think Google can&#8217;t detect blatant schemes like Text Link Ads you&#8217;re dreaming. They stick out like a geek on American Idol because:</p>
<ol>
<li><strong>Anchor text targets money terms.</strong> Anchor text like &#8220;buy viagra&#8221;, &#8220;car insurance&#8221;, and &#8220;search engine optimization&#8221; are more likely to raise Google&#8217;s suspicion than if you used anchor text like &#8220;click here for more info on why Text Link Ads suck.&#8221;</li>
<li><strong>Links are off-topic.</strong> (e.g. &#8220;online casino&#8221; link on a &#8220;paris hotel&#8221; site)</li>
<li><strong>No context.</strong> Anchor text without context is like fish out of water. Sure, there are legit reasons for linking out from your sidebar. But the reason is unclear because there&#8217;s no surrounding text to provide context.</li>
</ol>
<p>If you&#8217;re buying links through link brokers like Text Link Ads, <strong>stop wasting your money</strong>. Your chance of avoiding detection is lower than Boris Yeltsin scoring a one-night-stand with Maria Belluci. Instead, surf on over to technorati, search for topics related to your niche, sort by authority, and start at the top of your list. Email the owner of every blogger you see there, offering them money for blogging about your site. Remember, <strong>every blogger has a price</strong>, so don&#8217;t take no for an answer.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/05/14/stop-blowing-money-on-text-link-ads.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>If You Lack Motivation, Read This</title>
		<link>http://seo4fun.com/blog/2007/04/29/if-you-lack-motivation-read-this.html</link>
		<comments>http://seo4fun.com/blog/2007/04/29/if-you-lack-motivation-read-this.html#comments</comments>
		<pubDate>Sun, 29 Apr 2007 06:00:59 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/04/29/if-you-lack-motivation-read-this.html</guid>
		<description><![CDATA[Success is going from failure to failure without a loss of enthusiasm. - Winston Churchill
Work is the refuge of people who have nothing better to do. - Oscar Wilde
There&#8217;s a difference between knowing the path and walking the path - Morpheus, The Matrix
If we keep doing what we&#8217;re doing, we&#8217;re going to keep getting what [...]]]></description>
			<content:encoded><![CDATA[<p>Success is going from failure to failure without a loss of enthusiasm. - Winston Churchill</p>
<p>Work is the refuge of people who have nothing better to do. - Oscar Wilde</p>
<p>There&#8217;s a difference between knowing the path and walking the path - Morpheus, The Matrix</p>
<p>If we keep doing what we&#8217;re doing, we&#8217;re going to keep getting what we&#8217;re getting. - Stephen R. Covey</p>
<p>If you are failing to plan, you are planning to fail.</p>
<p>Don&#8217;t sweat the small stuff&#8230;and it&#8217;s all small stuff.</p>
<p>The measure of success is not whether you have a tough problem to deal with, but whether it&#8217;s the same problem you had last year.</p>
<p>Everything that irritates us about others can lead us to an understanding of ourselves.</p>
<p>Work is love made visible. And if you cannot work with love but only with distaste, it is better that you should leave your work and sit at the gate of the temple and take alms of those who work with joy.</p>
<p>If you don&#8217;t do it excellently, don&#8217;t do it at all. Because if it&#8217;s not excellent, it won&#8217;t be profitable or fun, and if you&#8217;re not in business for fun or profit, what the hell are you doing there?</p>
<p>Many attempts to communicate are nullified by saying too much.</p>
<p>I am grateful for all my problems. I became stronger and more able to meet those that were still to come. - J.C. Penney</p>
<p>The vision must be followed by the venture. It is not enough to stare up the steps - we must step up the stairs.</p>
<p>If each of us sweeps in front of our own steps, the whole world would be clean.</p>
<p>Flow with whatever is happening and let your mind be free. Stay centered by accepting whatever you are doing. This is the ultimate.</p>
<p>If the world seems cold to you, kindle fires to warm it.</p>
<p>The heart of a fool is in his mouth, but the mouth of the wise man is in his heart.</p>
<p>All that is gold does not glitter; not all those that wander are lost.</p>
<p>More <a href="http://www.heartquotes.net/Business.html">wise sayings</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/04/29/if-you-lack-motivation-read-this.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google&#8217;s Motives Are Selfish - So Are Yours and Mine</title>
		<link>http://seo4fun.com/blog/2007/04/17/googles-motives-are-selfish-so-are-yours-and-mine.html</link>
		<comments>http://seo4fun.com/blog/2007/04/17/googles-motives-are-selfish-so-are-yours-and-mine.html#comments</comments>
		<pubDate>Tue, 17 Apr 2007 23:52:24 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/04/17/googles-motives-are-selfish-so-are-yours-and-mine.html</guid>
		<description><![CDATA[As Graywolf said, Google&#8217;s motive for cleaning up the SERPs is self-serving and revenue-driven. Robert Scoble explains in more detail:
Why does Google care? Well, Google’s relevancy rankings will be hurt if people can buy their way onto their pages instead of earn their way to those search results pages by doing the best content, etc. [...]]]></description>
			<content:encoded><![CDATA[<p>As <a href="http://www.wolf-howl.com/google/how-can-so-many-phds-be-so-wrong/">Graywolf said</a>, Google&#8217;s motive for cleaning up the SERPs is self-serving and revenue-driven. <a href="http://scobleizer.com/2007/04/14/google-to-penalize-bloggers-selling-links/">Robert Scoble</a> explains in more detail:</p>
<blockquote><p>Why does Google care? Well, Google’s relevancy rankings will be hurt if people can buy their way onto their pages instead of earn their way to those search results pages by doing the best content, etc. Lots of people are doing comparisons of Google’s search results to Yahoo, Ask, and Microsoft’s search engines. If Google’s result set isn’t the best Google’s market share will start to go down as people figure out there are better engines out there. That, in turn, will hurt Google’s advertising business.</p>
<p>Not to mention that if advertisers know there’s a cheaper way to get onto Google’s search engine than by buying an ad, they’ll go with that system. So, Google has a LOT of incentive to swat down PayPerPost and pay-per-link style systems.</p></blockquote>
<p><a href="http://www.seobook.com/archives/002163.shtml">Aaron Wall</a> also points out that Google would make more money on Adwords if Google makes search results harder to game:</p>
<blockquote><p>
The more I think about it the more I realize why Google doesn&#8217;t like the various flavors of paid links. It has <strong>nothing</strong> to do with organic search relevancy. </p></blockquote>
<p>Not quite. Google cares <strong>alot</strong> about organic search relevancy. But here&#8217;s the catch. As I commented on Aaron&#8217;s blog,</p>
<blockquote><p>
Google wants informational sites on the organic results front page while forcing commercial sites to battle it out in the right column. Searchers looking for information on &#8220;coffee&#8221; will be happy with his organic results (Wikipedia, nationalgeographic, coffeereview, coffeeuniverse), and searchers looking to buy coffee will click on adwords and buy.</p>
<p>If commercial sites show up in organic results, there&#8217;s no reason for people to click on Adwords.</p>
<p>That means Google not only wants highly relevant organic results, but highly <strong>relevant results that are also non-commercial.</strong></p></blockquote>
<p>(Time to write meatier, more &#8220;informational&#8221; articles and dump those aff links from thin pages, huh?)</p>
<p>So Google is smacking down on paid links to increase its profit margin.</p>
<p>So what?</p>
<p>Do you read SEO blogs because they&#8217;re fun to read? Do you buy Adword ads because you want to help people find better products? Give me a break. I run ads to make money - pure and simple. And just because I have selfish reasons for giving you a great product doesn&#8217;t take away from the value of my product. Sure, Sony makes money off selling PS3s - so what? You want them to make them for free?</p>
<p>We&#8217;re all in it for the money, so &#8220;Google is selfish&#8221; objection doesn&#8217;t wash.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/04/17/googles-motives-are-selfish-so-are-yours-and-mine.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>21 Reasons Why Anti-Nofollow SEOs Can&#8217;t Think Straight</title>
		<link>http://seo4fun.com/blog/2007/04/16/21-reasons-why-anti-nofollow-seos-cant-think-straight.html</link>
		<comments>http://seo4fun.com/blog/2007/04/16/21-reasons-why-anti-nofollow-seos-cant-think-straight.html#comments</comments>
		<pubDate>Mon, 16 Apr 2007 18:30:46 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/04/16/21-reasons-why-anti-nofollow-seos-cant-think-straight.html</guid>
		<description><![CDATA[Paid link buyers and sellers are nothing but black hat spammers (though you&#8217;ll never catch me saying spamming is right or wrong - to quote Vlad from Max Payne 2, &#8220;you have to do what you have to do.&#8220;). Now I&#8217;m hearing alot of regurgitation going on, so I compiled a list of 21 major [...]]]></description>
			<content:encoded><![CDATA[<p>Paid link buyers and sellers are nothing but <a href="http://seoblackhat.com/2007/04/15/a-hearty-welcome-to-all-the-new-search-engine-spammers/">black hat</a> <a href="http://www.calacanis.com/2007/04/15/google-checkmates-payperpost/">spammers</a> (though you&#8217;ll never catch me saying spamming is right or wrong - to quote Vlad from <a href="http://www.rockstargames.com/maxpayne2/">Max Payne 2</a>, &#8220;<strong>you have to do what you have to do.</strong>&#8220;). Now I&#8217;m hearing alot of regurgitation going on, so I compiled a list of 21 major compelling (and not so compelling) anti-Google-paid-link-policy objections I came across on the Net in the last few days.</p>
<h2>And the nominees are&#8230;</h2>
<ol>
<li>Google, you&#8217;re just <strong>trying to <a href="http://www.seobook.com/archives/002163.shtml">make more money on Adwords.</a></strong><br />
<em>Rebuttal:</em> Just because I make money on a product <a href="/blog/2007/04/17/googles-motives-are-selfish-so-are-yours-and-mine.html">doesn&#8217;t take away from the value of my product</a>. You&#8217;re gonna tell me Sony&#8217;s evil because it doesn&#8217;t make PS3s for free?</li>
<li><strong>Innocent sites will get penalized</strong> if Google guesses wrong or if a site owner isn&#8217;t familiar with <a href="http://www.google.com/support/webmasters/bin/answer.py?answer=35769">Google&#8217;s Guidelines</a>.<br />
<em>Rebuttal:</em> None. This is a fair point and Google needs to address it.</li>
<li>It&#8217;s Google&#8217;s own fault for building <strong>a link-dependent algo.</strong><br />
<em>Rebuttal:</em> There&#8217;s no way Google can judge the quality or accuracy of content on a page unless they build a machine that can read and think. News flash - We&#8217;re not there yet. Google HAS to rely on links. There&#8217;s no way around it. You might bash NASA for not being able to build space colonies, but I got no respect for armchair quarterbacks that don&#8217;t even understand the game.</li>
<li><strong>There&#8217;s no other way for some niche sites to link build</strong>.<br />
<em>Rebuttal:</em> <a href="http://www.wolf-howl.com/seo/how-to-make-sea-slugs-cool/">Graywolf argues</a> against people &#8220;who don’t think linkbaiting can be used to their boring clients such as carpet cleaners, when I came up with five ideas in less than 15 minutes.&#8221; Porn is one tough niche to crack because no one links for free, but other niches in comparison aren&#8217;t so tough. Is no one linking to you? Maybe your website isn&#8217;t offering anything new or valuable.</li>
<li>Google is telling webmasters to <strong>build for search engines, not for people.</strong><br />
<em>Rebuttal:</em>  Have we all forgotten that Matt Cutts said <a href="http://video.google.com/videoplay?docid=8553629667451959310&#038;hl=en">build for both search engines and people?</a>. Now I agree rel=nofollow is building for search engines. Absolutely. Most users won&#8217;t even know its there. So we&#8217;re stuck in a grey area. But so what? Building for search engines is what SEOs have always done. Why do you pay $10,000 a month for links that send you 2 hits a day? Why do you waste hours writing unique META description tags? Why do you charge $600 an hour for SEO services? To improve usability? Gimme a break. And if you really have issues with building for search engines, maybe its time you quit your profession, because SEO isn&#8217;t about content building.</li>
<li>Paid links are whiter than other spam tactics like cloaking or hidden text, so <strong>why doesn&#8217;t Google go after more nefarious tactics first?</strong><br />
<em>Rebuttal:</em> Who says Google isn&#8217;t working on a solution for better cloak detection? If you assumed paid links is the only thing Google&#8217;s spam team is working on, you assumed wrong.</li>
<li><strong>Small site owners who make a living off selling links will go broke.</strong> You don&#8217;t wanna see them get evicted or something, do you Matt?<br />
<em>Rebuttal:</em> You&#8217;re making a living off contributing to spam and my heart should bleed because &#8230; why? Find something better to do with your websites.</li>
<li><strong>This is all FUD.</strong> Google is lousy at paid link detection.<br />
<em>Rebuttal:</em> Yeah, some of it is FUD. If Google could detect paid links, they wouldn&#8217;t need site owners to tag paid links with nofollow; they&#8217;d just auto devalue paid links without all this media hype and move on. And for easy-to-detect links (can you say Text Link Ads?) they probably already do. If a big chunk of your paid links are automated or above the radar, this isn&#8217;t FUD. You are fucked. If all your paid links are contextual, relevant, and point to high quality sites, then yeah, its FUD.</li>
<li><strong>Most links involve some sort of compensation</strong>, even if money doesn&#8217;t exchange hands.<br />
<em>Rebuttal:</em> Compensation isn&#8217;t the problem. Even marriage is a kind of a trade. I get to have sex every night with a beautiful wife in exchange for providing a roof over her head, helping her make babies, and giving her money to buy expensive jewelry and clothes. Ok so did she marry me for my money or did she marry me because she loves me? Compensation is a non-issue; almost everything in life is a trade. It&#8217;s the intent that&#8217;s in question.</li>
<li>I aint&#8217; worried. <strong><a href="http://www.sugarrae.com/blog/why-google-shouldnt-penalize-me-for-their-incompetence/">Some paid links are impossible to detect.</a></strong><br />
<em>Rebuttal:</em> Yeah, some individual links are undetectable. But many aren&#8217;t that hard to detect. And if Google detects a pattern of manipulative intent, your entire IBLs will become suspect.</li>
<li>Google, you&#8217;re not being realistic. <strong>You can&#8217;t expect dishonest people to behave honestly.</strong><br />
<em>Rebuttal:</em> None. Google needs to find a completely automated paid link detection method that doesn&#8217;t depend on people&#8217;s good will. <a href="http://www.people.com/people/article/0,,20034923,00.html">Rule breakers will always break rules</a>. From that POV, nofollow doesn&#8217;t work.</li>
<li><strong>Google makes money off link sellers</strong> like Text Link Ads by letting them run Google ads.<br />
<em>Rebuttal:</em> Google&#8217;s out to make money like everyone else. Besides, don&#8217;t blame Google&#8217;s Spam Team for what Google&#8217;s Adwords people do. They&#8217;re two completely different breeds of people.</li>
<li><strong>Pay Per Action doesn&#8217;t offer disclosure</strong> until you mouseover.<br />
<em>Rebuttal:</em> Big deal.</li>
<li><strong>Reporting paid links is snitching.</strong><br />
<em>Rebuttal:</em> Spam Report&#8217;s been available to the public for years, and most people (except Sugarrae) have used it at least once. You&#8217;ve even bitched about <a href="http://www.google.com/search?hl=en&#038;safe=off&#038;client=firefox-a&#038;rls=org.mozilla%3Aen-US%3Aofficial&#038;hs=RoR&#038;q=google+spam+report&#038;btnG=Search" rel="nofollow">Googlers not acting on your spam reports fast enough</a> (34 million results? Wow). So why the sudden uproar? There&#8217;s no bad karma in filing a spam report on a spammer. If you buy links, yeah, that makes you a spammer. But why worry? Googlers go out of its way not to manually penalize sites because manual bans don&#8217;t scale. That means hours of tweaking and testing before you see any spammer get penalized. As for &#8220;paidlinks&#8221; report, Google doesn&#8217;t even have a working algo in place yet. So why are you panicking?</li>
<li><strong>Matt, be clearer about what&#8217;s a paid link and what isn&#8217;t.</strong> How about charities that links to a list of donors. Are those paid links?<br />
<em>Rebuttal:</em> None. Even though some links are obviously paid for, others aren&#8217;t so obvious. Google needs to clearly define what constitutes paid and what doesn&#8217;t.</li>
<li><strong>You can <a href="http://www.10e20.com/2007/04/15/google-wants-you-to-report-paid-links/">damage your competitor</a></strong> by buying links to his site then reporting those links to Google.<br />
<em>Rebuttal:</em> Google is running a beta test on an algorithm that targets thousands of sites, not just one particular site. And history says Google punishes link sellers, not buyers. That <a href="http://www.marketingpilgrim.com/2007/03/googles-lasnik-wishes-nofollow-didnt-exist.html">may change over time</a> (Andy Beal: &#8220;Lasnik explains, why penalize hundreds of sites that sell just a single link, when it’s the recipient that is clearly benefiting?&#8221;) but I don&#8217;t see that day coming anytime soon. This is SEO FUD defense against Google&#8217;s FUD - and it ain&#8217;t pretty.</li>
<li>Using paid link reports to spot spam <strong>introduces a human factor</strong> in Google&#8217;s algo.<br />
<em>Rebuttal:</em> <a href="http://www.webmasterworld.com/google/3311622.htm">According to tedster</a>, a WMW mod, &#8220;Google is already using human input to a degree, and they&#8217;ve even patented a more scalable method for integrating editorial oversight without needing to rely on it for everything.&#8221; I don&#8217;t see Google fully automating everything. There are always going to be Google Adwords reviewers, Google Video submission reviewers, people who read spam reports and site reinclusion requests, engineers who think up new algorithms, PHDs that develop new BETA products&#8230; Sure, I&#8217;m sure Google would like to automate everything, but a human factor isn&#8217;t being &#8220;introduced&#8221;  - its always been a factor.</li>
<li><strong>Google, why are you cramming the Ten Commandments down webmasters&#8217; throats?</strong><br />
<em>Rebuttal:</em> Wanna cloak? Keyword spam? Use hidden text? Build doorway pages? Buy links? Go ahead. No one&#8217;s stopping ya. Do whatever you want with your site - it&#8217;s your site. But when you walk into someone else&#8217;s house you respect their rules or you&#8217;ll be asked to leave. It&#8217;s that simple. I like what Linkmoses said on Matt&#8217;s blog:</p>
<blockquote><p>
Ultimately we only have one choice to make. We either follow Google’s reco’s or we don’t. Nobody is forcing anything on us. Like the speed limit, we can -choose- to drive faster, and usually don’t get caught. Like the speed limit, we have no right to act shocked if we are pulled over.</p></blockquote>
</li>
<li><strong>If a link points to a relevant, quality site then compensation is irrelevant.</strong><br />
<em>Rebuttal:</em> Everyone has a price. Anyone who insists he/she won&#8217;t link to a crap site for any amount of money is blowing smoke. If I offered you one million dollars to link to a page that said something really nasty about your mother, you&#8217;ll not only link to it but send 10K uniques/day to it using Flash banners and Adbrite.</li>
<li><strong>Paid links improve search results.</strong> Successful companies with quality products and the baddest buying power deserve top rankings.<br />
<em>Rebuttal:</em> Let me introduce to you the players in this game. <em>SEOs:</em> These guys make money off fortune 500 companies who pay them $550/hour to buy up links. Without this tool, SEOs lose their edge. As <a href="http://www.seomoz.org/blog/i-disagree-with-danny-the-google-engineers-about-link-buying-practices">Rand Fiskin says</a>, &#8220;you&#8217;d be at <strong>a huge competitive disadvantage</strong> to your slightly less pointy-white-hat competitor.&#8221; <em>Link sellers:</em> these guys live off this monster of a marketplace; they do not want to lose their ability to make thousands of bucks a month on links. <em>Big companies:</em> their ranking <a href="http://www.cre8asiteforums.com/forums/index.php?s=2db884db5b41162ec0c2f84cbb42c1d3&#038;showtopic=48187&#038;pid=224718&#038;st=40&#entry224718">depends heavily on paid links</a> - these guys don&#8217;t want paid links to go away either. <em>Mom and Pop website owners (yeah, you):</em> you guys are basically screwed. The top 10 spots will be dominated by companies with millions to blow on links and these small website owners can&#8217;t compete. The delusion is that every webmaster is made to believe that by buying links, he can someday rank in the top 10, or if not rank at least a few spots higher. The reality is that no matter how much money you spend, if 10 websites outspend you, you&#8217;re never gonna show up on the front page. If the 10 richest companies dominate top SERP positions, <em>99% of you are screwed.</em></li>
<li><strong>Google, you made PageRank a commodity</strong> by displaying it in the Toolbar.<br />
<em>Rebuttal:</em> Only novice link buyers rely on the toolbar to find potential link sources. I assume Jim Boykin does alot of link buying, but I doubt he ever looks at the toolbar when measuring up a potential buy.</li>
<li><strong>Google, you&#8217;re <a href="http://www.wolf-howl.com/google/how-can-so-many-phds-be-so-wrong/">being hypocritical</a>.</strong> You said Yahoo Directory is ok because people pay for the review, not the link. So if someone pays me, I review his/her link, and then add the link to my site, why should I get penalized?<br />
<em>Rebuttal:</em> None. I can&#8217;t wrap my head around this one. I understand Google needs expert pages like Yahoo! Directory or DMOZ to calculate topic-dependent authority scores or calculate TrustRank, but to me it sounds like you&#8217;re skirting the issue. Alot of people who sell links review and reject link requests.</li>
<li><strong>Aren&#8217;t Adwords and adsense paid links?</strong><br />
Rebuttal: First, <em>Google has no problem with paid links</em> for traffic/advertisement. Get that through your thick skull. Second, neither Adsense nor Adwords pass PageRank. <a href="http://google.com/robots.txt" rel="nofollow">Google&#8217;s search pages are disallowed</a>. Notice the line &#8220;Disallow:/search&#8221;?</li>
<li><strong>It&#8217;s not our job to police the internet.</strong>.<br />
<em>Rebuttal:</em> Did Matt Cutts offer you money to report paid links? If not, I don&#8217;t consider that a job. As for policing the internet, you&#8217;re not policing unless you spot a cheater and get him banned. Google isn&#8217;t interested in banning anyone. They&#8217;re interested in BETA testing their new algorithms. So its more like collecting guinea pigs than playing the town sheriff.</li>
</ol>
<p><strike>You want to see me try to counter them, right? I might later, but its Monday and I got alot of other stuff to do. *ducks*</strike></p>
<p>For now, here&#8217;s my off-the-cuff advice - something you already know. If you&#8217;re shopping around for links, <a href="http://www.jimboykin.com/">buy them under the radar</a>. There are some paid links Google will never be able to detect, but a service like Text Link Ads isn&#8217;t one of them. Those links scream &#8220;paid links&#8221;, and the company is too visible. Assuming Text Link Ads links still carry some juice, they&#8217;ll be the first sinking ship among many if Google has its way.</p>
<p><strong>UPDATE:</strong> This list is growing by the minute - 24 objections and counting. I updated this post this morning with rebuttals so people like Nick will have something to sink their teeth into.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/04/16/21-reasons-why-anti-nofollow-seos-cant-think-straight.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Why Does Viagra.com 302 Its Home Page?</title>
		<link>http://seo4fun.com/blog/2007/04/11/why-does-viagracom-302-its-home-page.html</link>
		<comments>http://seo4fun.com/blog/2007/04/11/why-does-viagracom-302-its-home-page.html#comments</comments>
		<pubDate>Wed, 11 Apr 2007 16:50:30 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/04/11/why-does-viagracom-302-its-home-page.html</guid>
		<description><![CDATA[Why did viagra.com drill a 302 into their home page? And why does the redirect dump me on this cryptic, Google-unfriendly URL:
http://www.viagra.com/content/index.jsp?setShowOn=../content/index.jsp
&#038;setShowHighlightOn=../content/index.jsp
No wonder viagra.com is beat by .edu spam for &#8220;buy viagra.&#8221; It seems the company opted out of hiring a competent webmaster (no you don&#8217;t need an SEO to figure out redirects). Or is [...]]]></description>
			<content:encoded><![CDATA[<p>Why did viagra.com drill a 302 into their home page? And why does the redirect dump me on this cryptic, Google-unfriendly URL:</p>
<p>http://www.viagra.com/content/index.jsp?setShowOn=../content/index.jsp<br />
&#038;setShowHighlightOn=../content/index.jsp</p>
<p>No wonder viagra.com is beat by .edu spam for &#8220;buy viagra.&#8221; It seems the company opted out of hiring a competent webmaster (no you don&#8217;t need an SEO to figure out redirects). Or is there something major usability-oriented thingie I&#8217;m missing? Nah, I can&#8217;t think of one legit explanation. Can you?</p>
<p>BTW, <a href="http://www.vanessafoxnude.com/">Vanessa Fox Nude</a> is now 5th for &#8220;vanessa fox&#8221;, and 1st for &#8220;vanessa fox nude.&#8221; Her new blog outranks DaveN and Search Engine Land, two very &#8220;authoritative&#8221; sites. That&#8217;s reality folks. On the other hand, she won&#8217;t have an easy time outranking webmastercentral, but if she reveals more skin (er..how Google really works), she&#8217;ll make it to the top.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/04/11/why-does-viagracom-302-its-home-page.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Why in Elvis&#8217; Name Do I Blog?</title>
		<link>http://seo4fun.com/blog/2007/04/10/why-in-elvis-name-do-i-blog.html</link>
		<comments>http://seo4fun.com/blog/2007/04/10/why-in-elvis-name-do-i-blog.html#comments</comments>
		<pubDate>Tue, 10 Apr 2007 06:04:54 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/04/10/why-in-elvis-name-do-i-blog.html</guid>
		<description><![CDATA[Michael Goldberg tagged me. I was also tagged earlier by SEM Zone and I even wrote a response but I forgot all about it. UPDATE: JLH and John aka Softplus also tagged me. John says I&#8217;m not revealing enough about myself, hmmm&#8230;
So, should I post a serious reply, try to be funny like Wayne Knight [...]]]></description>
			<content:encoded><![CDATA[<p><a href="http://www.googleorganic.com/wordpress/">Michael Goldberg</a> tagged me. <a href="http://www.thesemzone.com/2007/02/tagged-why-i-blog.html">I was also tagged earlier</a> by SEM Zone and I even wrote a response but I forgot all about it. UPDATE: <a href="http://www.jlh-design.com/">JLH</a> and <a href="http://seside.net/">John aka Softplus</a> also tagged me. John says I&#8217;m not revealing enough about myself, hmmm&#8230;</p>
<p>So, should I post a serious reply, try to be funny like Wayne Knight fumbling through <a href="http://www.nbc.com/TGYH/">Thank God You&#8217;re Here</a>, or write a link bait piece like I was Rand Fishkin Jr.?</p>
<p>I suck at talking about my cats, so I&#8217;ll let my honesty bore you to death.</p>
<p>In December 2005, one of my money sites went completely supplemental. I&#8217;m talkin&#8217; a 2000+ page site reduced to 3 pages. This was back before Big Daddy when no one had a clue about the supplemental index. On WMW, everyone was fixated on duplicate content, gs1md leading that discussion with a ton of insights into combatting canonical issues. The SEO &#8220;experts&#8221; outside of WMW didn&#8217;t have much of a clue. For example, <a href="http://www.jimboykin.com/damned-to-google-hell-supplemental-results/">This is the first article</a> I ever read about supplemental results, written by Jim Boykin:</p>
<blockquote><p>Now, for the dirt - how to get out.<br />
1. If you stole content - change it.<br />
2. If there’s no content - add some.<br />
3. If it’s orphaned - link to it.</p></blockquote>
<p>(Love your blog Jim, but first impressions die hard :D) On Sept 6, 2006, Ammon over on cre8 <a href="http://www.cre8asiteforums.com/forums/index.php?showtopic=41208">said this</a> about the supplemental index:</p>
<blockquote><p>
Supplemental usually means “Google knows of this URL, but has not spidered the document recently for it to be in the main index”.</p></blockquote>
<p>As late as Nov 2006, in reaction to my post claiming low PageRank was the primary factor producing supplemental results, Rand Fishkin <a href="http://www.seomoz.org/blog/duplicate-content-revisited">responded</a>:</p>
<blockquote><p>  I think it’s bogus - maybe it’s the primary factor in that a huge number of pages that are no longer linked to (in site structures from large sites) drop into supplemental, but for most of the real pages that webmasters want in the index that get dropped, I don’t think PageRank is playing a big role.
</p></blockquote>
<p>Now, Michael Martinez bitched about my <a href="/notes/supplementals.html">Supplemental Results page</a> being irrelevant, out of date, and first tier, (which, in retrospect, isn&#8217;t completely untrue)</p>
<blockquote><p>The article may very well create a buzz and go on to become one of the SEO community’s standard references on how to deal with Google’s Supplemental Index. And the irony is that it’s wrong, even though the correct answer (as far as what I have seen work in the past few weeks) is buried amidst all the bad/good advice.</p></blockquote>
<p>but what he doesn&#8217;t know is there was virtually nothing on the web about supplemental results in the Spring of 2006 which is when I started writing that page. Even though the page feels near obsolete now, back then it was ahead of the curve.</p>
<p>Anyway, during Big Daddy&#8217;s release, Matt Cutts mentioned that lack of trust in the inlinks/outlinks of a site leads to PageRank devaluation, which leads to low overall PageRank for a domain, which leads to pages dropping out of the main index - which <em>exposes</em> supplemental results.</p>
<p>But guess what? No one was listening, or didn&#8217;t want to listen, because</p>
<p><strong>They resisted letting go of the idea that PageRank is dead.</strong></p>
<p>When you look at the mechanics behind any piece of code, you discover function calls, loops, if/then statements, variables. Like it or not, PageRank is one of those variables. While it remains inside Google&#8217;s code, it maintains its influence, however slight, regardless of what anyone outside of Googleplex wants to believe.</p>
<p>Marketers who excell at writing digg-happy headlines will tell you what sounds cool - but what do they know? For example, marketers say &#8220;trust&#8221; alot (yeah, I know, I do too). Trust in the scope of supplemental results isn&#8217;t about domains; its about links, exchanged links, paid links. It&#8217;s about pattern detection, not authority.  It&#8217;s about link devaluation, not a ranking boost. Trust in the scope of TrustRank has to do with high PageRank sites penalized in search results when they got lousy link profiles. TrustRank doesn&#8217;t effect low PageRank sites.</p>
<p>But stuff like that bores the crap out of most readers. You want to read stuff that gets you more sales. You want to know how to game Digg. You don&#8217;t want to waste time trying to postulate theories about an uncracked algorithm.</p>
<p>If you look at my archive links, you&#8217;d notice I started this blog on March 2006, right around the release of Big Daddy.</p>
<p>So, I guess the one and only reason I started this blog is a selfish one - I used this blog like a sailor uses a compass while lost at sea.</p>
<p>Then again, I was never really that lost to begin with.</p>
<p>I&#8217;ll tag these guys:</p>
<p><a href="http://www.irelandseomarketing.com/">Ireland SEO Marketing</a><br />
<a href="http://www.johnon.com/">John Andrews</a><br />
<a href="http://petertdavis.net/">Peter T Davis</a><br />
<a href="http://www.redcardinal.ie/">Red Cardinal</a><br />
<a href="http://www.scoreboard-media.com/">Scoreboard Media Group</a></p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/04/10/why-in-elvis-name-do-i-blog.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Proof is in the SERPS - Overestimating Domain Authority, Take 2</title>
		<link>http://seo4fun.com/blog/2007/04/06/proof-is-in-the-serps-overestimating-domain-authority-take-2.html</link>
		<comments>http://seo4fun.com/blog/2007/04/06/proof-is-in-the-serps-overestimating-domain-authority-take-2.html#comments</comments>
		<pubDate>Fri, 06 Apr 2007 05:03:00 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/04/06/proof-is-in-the-serps-overestimating-domain-authority-take-2.html</guid>
		<description><![CDATA[In my previous long-ass post titled Free SEO Course - Overestimating Domain Authority, I dared to outrank Peter Da Vanzo from V7n  blog for the keyphrases &#8220;free seo course&#8221; and &#8220;seo course.&#8221; More than anything, I felt like challenging a slick sound bite that - while holding a morsel of truth - distorted the [...]]]></description>
			<content:encoded><![CDATA[<p>In my previous long-ass post titled <a href="/blog/2007/03/20/free-seo-course-offering-expert-training-overestimating-domain-authority.html">Free SEO Course - Overestimating Domain Authority</a>, I dared to outrank Peter Da Vanzo from <a href="http://blog.v7n.com/">V7n  blog</a> for the keyphrases &#8220;free seo course&#8221; and &#8220;seo course.&#8221; More than anything, I felt like challenging a slick sound bite that - while holding a morsel of truth - distorted the picture beyond recognition:</p>
<blockquote><p>It&#8217;s not what you publish, its where.</p></blockquote>
<h2>Authority Matters</h2>
<p><strong>Where you publish a post matters.</strong> No question about it. V7n ranking high with just a handful of links says volumes about how authority score still remains a force to be reckoned with. <a href="http://www.redcardinal.ie/">Red Cardinal</a> outranking my post with very few backlinks also proves content - even anchor text (if there aren&#8217;t enough of them from the right sources) - isn&#8217;t everything.</p>
<h2>Nothing Really Matters</h2>
<p>But &#8220;trust&#8221; and &#8220;reputation&#8221; isn&#8217;t everything either. Peter is telling me <strong>all that really matters is a big rep.</strong> Is SEO that simple? Is driving in NASCAR just about having <a href="http://www.nascar.com/drivers/dps/jgordon00/cup/">Jeff Gordon</a> behind the wheel? Is a <a href="http://billiards.about.com/od/stroketechniques/ss/01_02_06power.htm">9-ball power break</a> just about hitting a rack hard as hell?</p>
<p>If reputation is everything, I dare Peter to change the title of his post to &#8220;<a href="http://www.trumpgolf.com/trumplosangeles/index.asp">Expensive Golf Course</a>&#8221; and see if it can maintain its position for &#8220;seo course.&#8221; All the <a href="http://www.seomoz.org/article/search-ranking-factors">so-called SEO&#8217;s in-the-know</a> will tell you the post will tank.</p>
<p>There are hundreds of factors. We hear that 24/7 - but let&#8217;s say it again.</p>
<p><strong>There are hundreds of factors.</strong> Yeah, only a few of them really make a big difference. But there&#8217;s no factor that is so dominant that you can forsake all the rest.</p>
<p><img src="/images/free-seo-course-2.gif" alt="free SEO course SERP" /></p>
<p><em>SERP for &#8220;free seo course&#8221;. Screen cap for posterity (plus this SERP flip-flops more often than John Kerry)</em></p>
<p><img src="/images/seo-course.gif" alt="seo course SERP" /></p>
<p><em>SERP for &#8220;seo course.&#8221; V7n is now off the front page.</em></p>
<h2>Don&#8217;t Think You Can&#8217;t Outrank Wikipedia</h2>
<p><strong>What does this prove?</strong> First, it proves that even if your website doesn&#8217;t have that visibility you need to break into the &#8220;SEO circule of trust&#8221;, you can still outrank authoritative sites like V7N, SE Roundtable, and even Wikipedia. A gazillion links to a domain makes internal links that much more powerful, but you only need to outrank a page, not the entire domain. Put another way - even if a domain has millions of links pointing at it, if one of its pages has a weak link profile, you can beat it.</p>
<p>Second, it tells me anchor text creates relevance, and enough relevance will trump &#8220;reputation,&#8221; PageRank, or what have you.</p>
<p>Thanks to everyone that linked to me, but for the final push that put me on the map, the credit goes to <a href="http://seo-theory.com/">SEO Theory</a>.</p>
<p>UPDATE: Since I&#8217;m done with this test, I asked Michael to remove links to my post.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/04/06/proof-is-in-the-serps-overestimating-domain-authority-take-2.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Matt Cutts Defaces Dark SEO Team - or Does He?</title>
		<link>http://seo4fun.com/blog/2007/04/02/matt-cutts-defaces-dark-seo-team-or-does-he.html</link>
		<comments>http://seo4fun.com/blog/2007/04/02/matt-cutts-defaces-dark-seo-team-or-does-he.html#comments</comments>
		<pubDate>Mon, 02 Apr 2007 16:34:54 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/04/02/matt-cutts-defaces-dark-seo-team-or-does-he.html</guid>
		<description><![CDATA[Yesterday, you probably noticed Matt Cutts&#8217; blog supposedly defaced by the Dark SEO team  (ya know, those guys who have all those fake TBPR pages up), as reported by Search Engine Land in a piece titled Matt Cutts gets hacked.
The Dark SEO Team has had a bit of a beef with Google&#8217;s Matt Cutts [...]]]></description>
			<content:encoded><![CDATA[<p>Yesterday, you probably noticed <a href="http://www.mattcutts.com/blog/">Matt Cutts&#8217; blog</a> supposedly defaced by the Dark SEO team  (ya know, those guys who have all those fake TBPR pages up), as reported by Search Engine Land in a piece titled <a href="http://searchengineland.com/070331-225536.php">Matt Cutts gets hacked</a>.</p>
<blockquote><p>The Dark SEO Team has had a bit of a beef with Google&#8217;s Matt Cutts from back in 2005 over URL hijacking. Looks like they&#8217;ve pulled a prank on him today. Matt&#8217;s blog is down, hacked</p></blockquote>
<p>Well it turns out the prank was <a href="http://www.mattcutts.com/blog/april-fools-day-2007/">Matt Cutts&#8217; own doing</a>, inspired by his wife. To tell ya the truth, he got me too because I really believed his site got hacked. A WMW member even pointed out t<a href="http://www.webmasterworld.com/google/3298808-3-20.htm">here&#8217;s some French in the source code</a>, and that he didn&#8217;t think Matt spoke French.</p>
<p>But now check out Dark seo team&#8217;s <a href="http://www.darkseoteam.com/" rel="nofollow">home page</a>. Who&#8217;s responsible for that one?</p>
<p>If you view source, you&#8217;ll find stuff like &#8220;phe4r !! THIS IS NOT KEYWORDS STUFFING !! Matt just try to be funny and that&#8217;s work, all the seo world is loud of laugh !&#8221;</p>
<p>The funniest line on the page is &#8220;Google is just a stupid algorithm relying on spammy backlinks. But you guys had no right to let everyone know.&#8221;</p>
<p>The photo you see is <a href="http://infolab.stanford.edu/pub/voy/museum/pictures/display/0-4-Google.htm">The Original GOOGLE Computer Storage</a> wrapped in a wall of rainbow-colored LEGO:</p>
<blockquote><p>
Crawling the web to obtain its link structure required an enormous amount of storage in comparison with typical student projects at that time. We show here the original storage assembly, containing 10 4 Gigabyte disk drives, giving 40 Gbytes total.</p></blockquote>
<p>40 Gigs! (we&#8217;re talking 1996) That&#8217;s about how much space I try to keep open for periodic degrag. Come to think of it, I got 200 GB inside my rig and 200 GB on an external drive (yeah, I should buy a cheap 500GB drive). So I could theoretically crawl the entire web without overflowing my hard drive.</p>
<p>Anyway, I hope you had a fun April Fools Day.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/04/02/matt-cutts-defaces-dark-seo-team-or-does-he.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Free SEO Course Offering  Expert Training - Overestimating Domain Authority</title>
		<link>http://seo4fun.com/blog/2007/03/20/free-seo-course-offering-expert-training-overestimating-domain-authority.html</link>
		<comments>http://seo4fun.com/blog/2007/03/20/free-seo-course-offering-expert-training-overestimating-domain-authority.html#comments</comments>
		<pubDate>Tue, 20 Mar 2007 11:47:49 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/03/20/free-seo-course-offering-expert-training-overestimating-domain-authority.html</guid>
		<description><![CDATA[If you ever paid for an SEO course run by an expert SEO training you to conquer the Intarweb, you&#8217;d be taught what you&#8217;ve known all along - that the two major factors Google uses to rank a page are relevance and value.
(Hey, if you decide to skip reading this long-ass free expert seo course [...]]]></description>
			<content:encoded><![CDATA[<p>If you ever paid for an SEO course run by an expert SEO training you to conquer the Intarweb, you&#8217;d be taught what you&#8217;ve known all along - that the two major factors Google uses to rank a page are <em>relevance</em> and <em>value</em>.</p>
<p>(Hey, if you decide to skip reading this long-ass free expert seo course training post<em>, please do me a favor and at least read the last two short paragraphs.</em>)</p>
<h2>Relevance, How You Build It</h2>
<p>For the uninitiated SEO course newbie, relevance is determined in part by what a page claims its about (e.g. page TITLE, on-page text, keywords in domain name, keywords in url) and what other people say its about (anchor text). If a page claims to be about &#8220;minestrone soup&#8221; but if everybody links to the page with the anchor text &#8220;cold pizza bought two days ago&#8221;, its not going to rank well for either of those terms. You can&#8217;t just claim to be <a href="http://abc.go.com/latenight/jimmykimmel/index?pn=index">Jimmy Kimmel</a> if you&#8217;re calling up Mastercard; you need to prove it (e.g. social security number, driver&#8217;s license number, your favorite rock band while sleeping through classes at <a href="http://en.wikipedia.org/wiki/Ed_W._Clark_High_School" rel="nofollow">Ed W. Clark High School</a>). <strong>Anchor text verifies who you say you are.</strong> Any basic free SEO training course worth its salt will teach you that.</p>
<h2>Value, the other side of the Coin</h2>
<p>Value, on the other hand, is determined primarily by PageRank. Put simply, Jim Boykin recommending one of your blog posts, for example, is a greater indication of value than some nameless guy on 12lzafksaldf0234.wordpress.org mentioning you.</p>
<p>According to a Google patent filed in 2004 titled <a href="http://www.google.com/patents?vid=USPAT7028029&#038;id=UTh4AAAAEBAJ&#038;pg=RA1-PA1&#038;dq=pagerank#PRA1-PA6,M1">Adaptive computation of ranking</a>, calculating relevance is more computationally intensive than calculating value:</p>
<blockquote><p>
One approach to ranking documents involves examining the intrinsic content of each document or the back-link anchor text in parents to each document. <strong>This approach can be computationally intensive and often fails to assign highest ranks to the most important documents.</strong> Another approach to ranking involves examining the extrinsic  relationships between documents, i.e. from the link structure of the directed graph&#8230;</p>
<p>Although link-based ranking techniques are improvements over prior techniques, in the case of an extremely large database, such as the World Wide Web which contains billions of pages, the computation of the ranks for all the pages can take considerable time.</p></blockquote>
<p>(Emphasis mine)</p>
<p>That&#8217;s one strike against relevance.</p>
<h2>Domain Strength Inflates Page Value?</h2>
<p>In addition, some SEO experts claim that <strong>domain strength adds ranking power to a page</strong>. Rand Fishkin <a href="http://www.seomoz.org/blog/two-clarifications-on-how-search-engines-interpret-links">recently claimed</a> that migrating (301 redirecting) his Web 2.0 piece from a relatively new domain to a directory under the wings of SEOmoz.org dramatically improved the page&#8217;s ranking. <a href="http://www.seomoz.org/web2.0/">His page</a> currently sits at 5th position for &#8220;Web 2.0.&#8221;</p>
<p>Peter Da Vanzo from V7N <a href="http://blog.v7n.com/2007/03/19/top-ten-in-serps-in-a-couple-of-days/">published</a> a post today also supporting that claim, citing his recently published <a href="http://blog.v7n.com/2007/03/11/free-seo-course-what-is-search-marketing/">piece</a> titled <em>Free SEO Course: What Is Search Marketing?</em>, which currently ranks 7th on Google for the term &#8220;seo course&#8221;, as evidence.</p>
<p>Peter didn&#8217;t mention the fact that his page isn&#8217;t unoptimized for the term &#8220;SEO course.&#8221; He&#8217;s got the term &#8220;FREE SEO course&#8221; in the blog title. He also has &#8220;SEO course&#8221; embedded in two H tags. There are 7 on-page occurances of the word &#8220;course&#8221; and 5 occurances of the term &#8220;SEO course.&#8221; He&#8217;s got words &#8220;free&#8221;, &#8220;seo&#8221;, &#8220;course&#8221; in the URL. He also has a couple of other domains (Search Engine Land included) linking to his post using the title, which validates the page&#8217;s own assertion as being about &#8220;free SEO course.&#8221; Finally, there&#8217;s an internal link from <em>free SEO Course Part Six:Advanced Tips and Tricks</em> pointing to the first installment of his SEO course series with the anchor text &#8220;free SEO course.&#8221;</p>
<p>I agree that a page published on a strong domain is likely to be more trusted. A new page on a new domain is like a new kid on the block that needs to pass a few tests before being accepted by the community. But does domain strength improve a page&#8217;s value? I&#8217;m not too sure.</p>
<p>PageRank per page on a domain with massive IBLs also tend to average out higher. In the Web 2.0 page&#8217;s case, Rand claims IBLs remained basically unchanged.</p>
<h2>Page Value or Relevance? No-brainer: Be Valuable and Relevant</h2>
<p>Many folks shooting down PageRank on forums and attributing power to relevance instead still seem to maintain that both trust and domain authority are important. However, domain authority, trust, and PageRank are metrics of page value, not relevance. Google uses both to rank pages, <strong>but which gives you more bang for your buck?</strong></p>
<p>According to this <a href="http://www.cs.cornell.edu/home/kleinber/auth.pdf">TrustRank paper (PDF)</a>, as I posted over on <a href="http://www.cre8asiteforums.com/forums/">cre8asiteforums</a>, <strong>pages with higher PageRank tend to be displayed higher in search results</strong>. TrustRank algorithm targets high PageRank spam pages because they are more likely to be seen by users:</p>
<blockquote><p>For example, say we have four pages p, q, r, and s, whose <em>contents match a given set of query terms equally well</em> [they&#8217;re equally relevant]. If the search engine uses PageRank to order the results, the page with highest rank, say p, will be displayed first, followed by the page with next highest rank, say q, and so on. Since it is more likely the user will be interested in pages p and q, as opposed to pages r and s (pages r and s may even appear on later result pages and may not even be seen by the user), it seems more useful to obtain accurate trust scores for pages p and q rather than for r and s.</p></blockquote>
<p>In other words, theTrustRank paper attributes importance to both relevance and PageRank, where <strong>PageRank acts as a tie breaker when pages are equally relevant</strong>. How often Google finds equally relevant documents is another question, though. If that scenario seldom comes up, PageRank becomes a non-issue, because there are no ties to break.</p>
<p>But wait a minute - as Mike Grehan points out in his <a href="http://www.searchguild.com/topic_distillation.pdf">Topic Distillation paper</a> (PDF):</p>
<blockquote><p>
..just because they are relevant doesn&#8217;t mean they are the most useful, or for that matter, the most important. Kleinberg himself calls this the &#8220;abundance problem&#8221; and states that <strong>the number of pages that could reasonably be returned as relevant is far too large for a human user to digest.</strong></p></blockquote>
<h2>What You Get from this Free SEO Course Expert Training Post</h2>
<p>My SEO blog is invisible compared to blogs like v7n, so if Peter is right, I can&#8217;t outrank him just by keyword spamming &#8220;seo course&#8221; all over this post or by gaining links to this post with &#8220;seo course&#8221; in the anchor text, because, according to him, &#8220;it&#8217;s not what you publish, it’s where.&#8221;</p>
<p>I disagree. If you&#8217;re still reading, <strong>please help me prove it</strong>. So far, I see 7+ links pointing to Peter&#8217;s free seo course post (according to live.com). If I can get at least 8 links pointing to this post with &#8220;seo course&#8221; somewhere in your anchor text, that should be enough to either prove or disprove my claim that <strong>what you publish and what people think of your page is as important if not more than where a page is published</strong>.</p>
<p><strong>UPDATE:</strong> Red Cardinal left me kinda <a href="http://www.redcardinal.ie/">speechless with a generous post</a>. Also thanks to <strong>megabluewave</strong> and <strong>irelandseomarketing</strong> for linking up. (Trackbacks are nofollow-free). </p>
<p>3/24/2007: Hmm, looks like Google dropped this page (guess I tripped a filter); it was ranking 7th for &#8220;free seo course&#8221; and 19th for &#8220;seo course.&#8221; But notice now Red Cardinal is ranking 7th for &#8220;free seo course&#8221; and 15th for &#8220;seo course.&#8221; :) Is his TBPR 6 home page helping him out? My TBPR 4 blog root is still ranking, but is ranking low, IMO mainly due to lack of keywords in the TITLE.</p>
<p>UPDATE: 4/9/2007. I outrank v7n for both keywords, at least temporarily. <a href="/blog/2007/04/06/proof-is-in-the-serps-overestimating-domain-authority-take-2.html">Read this</a> for the full update.<br />
UPDATE: I asked some IBLS to be removed, so don&#8217;t be surprised to see my site fall back in the SERPs. </p>
<p>P.S. For a stronger counterargument, see <a href="http://www.seomoz.org/blog/parasite-hosting-now-dominating-spam-results">Rand&#8217;s piece</a> on Parasite Hosting.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/03/20/free-seo-course-offering-expert-training-overestimating-domain-authority.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>How Do 6 SEOs Miss The Obvious?</title>
		<link>http://seo4fun.com/blog/2007/03/07/how-does-6-seos-miss-the-obvious.html</link>
		<comments>http://seo4fun.com/blog/2007/03/07/how-does-6-seos-miss-the-obvious.html#comments</comments>
		<pubDate>Wed, 07 Mar 2007 07:31:06 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/03/07/how-does-6-seos-miss-the-obvious.html</guid>
		<description><![CDATA[I used to spend some time looking at sites people posted up on Google Group Webmaster Help as a form of recreation. I enjoyed looking at real examples instead of listening to assertions on WMW people could not back up. So when I read Search Engine Journals 6 SEOs-of-caliber write their take on techsmith.com, I [...]]]></description>
			<content:encoded><![CDATA[<p>I used to spend some time looking at sites people posted up on Google Group Webmaster Help as a form of recreation. I enjoyed looking at real examples instead of listening to assertions on WMW people could not back up. So when I read Search Engine Journals 6 SEOs-of-caliber write <a href="http://www.searchenginejournal.com/?p=4488">their take on techsmith.com</a>, I was hoping to be amazed. Its kind of like the SEO-version of American Idol, where 6 expert judges look at your site and tell you what&#8217;s hot, what&#8217;s not, and what you can do to make it better. It&#8217;s potentially a huge plus for site owners looking for SEO advice and a huge plus for the SEO community as well which has been defending itself recently against badly-uninformed claims that SEO is bullshit.</p>
<p>By the time I was done reading, I thought to myself &#8220;these guys got it covered.&#8221; In fact, Michael Martinez commented: &#8220;That was one of the most thorough, informative online analyses I have read in a very, very long time.&#8221; That&#8217;s like hearing <a href="http://www.house.gov/pelosi/">Nancy Pelosi</a> say something nice about the Bush Administration.</p>
<p>To be frank, I wasn&#8217;t too impressed with the ALT tag and H element suggestions, because those are not the prime on-page factors (keyword frequency and keywords at the beginning of TITLE is all you need), but otherwise my initial reaction was that the article was spot on.</p>
<p>But that was before I ran a link: search on techsmith and found several major domains like snagit.com and camtasia.com 302 redirecting into techsmith.com. Now 302 redirects are useful for linking out to affiliate sites, but if you&#8217;re trying to transfer link juice from an old domain to a new directory, you should use a 301 redirect instead. Because they&#8217;re using a 302, that link juice isn&#8217;t getting transferred and the target urls aren&#8217;t benefiting from the redirect.</p>
<p>More importantly, when I dug up techsmith.com&#8217;s outlinks, I was a little baffled to find  <a href="http://www.techsmith.com/community/blogcomments.asp?thread=12" rel="nofollow">this</a>:</p>
<p><img src="/images/spam-comments.jpg" alt="spam comments" /></p>
<p>That&#8217;s right. Outbound blog links to porn spam. And yeah, those links are linking without protection. The URL <a href="http://72.14.203.104/search?q=cache%3Ahttp%3A%2F%2Fwww.techsmith.com%2Fcommunity%2Fblogcomments.asp%3Fthread%3D12&#038;ie=utf-8&#038;oe=utf-8&#038;aq=t&#038;rls=org.mozilla:en-US:official&#038;client=firefox-a" rel="nofollow">isn&#8217;t cached yet</a>, but <a href="http://www.techsmith.com/robots.txt" rel="nofollow">the directory isn&#8217;t noindexed via robots.txt</a> either, so the links will do some harm once the page gets indexed.</p>
<p>Am I the only guy who bothers to look at the outlinks of a site to make sure a site isn&#8217;t accidentally linking out to a bunch of crap? To be fair, the site does have a load of outbound links, but it only took me a few seconds to find these links, and there&#8217;s enough links on that post its not like looking for a needle in a haystack. Anyway, I&#8217;m mostly impressed and a little disappointed, which reminds me I wanna watch American Idol tomorrow night, but I&#8217;ll probably watch <a href="http://www.nbc.com/Friday_Night_Lights/">Friday Night Lights</a> instead.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/03/07/how-does-6-seos-miss-the-obvious.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>YouTube Taking Down More Music Videos</title>
		<link>http://seo4fun.com/blog/2007/02/25/youtube-taking-down-more-music-videos.html</link>
		<comments>http://seo4fun.com/blog/2007/02/25/youtube-taking-down-more-music-videos.html#comments</comments>
		<pubDate>Sun, 25 Feb 2007 16:28:52 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/02/25/youtube-taking-down-more-music-videos.html</guid>
		<description><![CDATA[Looking to download videos from YouTube for free?
1. Download a free FLV player from http://www.jeroenwijering.com/?item=Flash_video_Player.
2. Go to Keepvid.com.
Pretty simple huh?
So I woke up today to find out YouTube&#8217;s pulling more and more music videos. Out of 20 videos I bookmarked (yeah I know, not a whole lot), around 16 of them are now gone. It [...]]]></description>
			<content:encoded><![CDATA[<p>Looking to download videos from YouTube for free?</p>
<p>1. Download a free FLV player from http://www.jeroenwijering.com/?item=Flash_video_Player.<br />
2. Go to Keepvid.com.</p>
<p>Pretty simple huh?</p>
<p>So I woke up today to find out YouTube&#8217;s pulling more and more music videos. Out of 20 videos I bookmarked (yeah I know, not a whole lot), around 16 of them are now gone. It was just a matter of time and I understand why they gotta do this but it doesn&#8217;t mean I gotta like it. There are other ways to get my hands on my favorite music vids like P2P but I prefer watching videos on sites like AOL, MSN or YouTube where I don&#8217;t have to wait for songs to download and I didn&#8217;t have to bend over backwards hunting them down. It&#8217;s easy to download songs that are still available on Youtube, though alot of the songs that used to be on there are either going or gone. Yeah, I know, I&#8217;m such a damn freeloader. Anyway, in the next few months, YouTube will go from your ultimate source of copyrighted material to a place where you can only find silly people doing silly things.</p>
<p>UPDATE: According to the Washington Post, YouTube failed to hammer out a deal with Viacom, who owns MTV and Comedy Central. Viacom instead struck up a deal with Joost. By disallowing YouTube from airing videos, websites like MTV.com and comedycentral.com are seeing a rise in traffic.</p>
<p>But some people don&#8217;t think Viacom is being too smart &#8220;because its[YouTube&#8217;s] audience surpasses that of any other video site, with 100 million video streams a day. &#8220;</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/02/25/youtube-taking-down-more-music-videos.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Why Duplicate Content Causes Supplemental Results</title>
		<link>http://seo4fun.com/blog/2007/02/19/why-duplicate-content-causes-supplimental-results.html</link>
		<comments>http://seo4fun.com/blog/2007/02/19/why-duplicate-content-causes-supplimental-results.html#comments</comments>
		<pubDate>Mon, 19 Feb 2007 10:55:28 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/02/19/why-duplicate-content-causes-supplimental-results.html</guid>
		<description><![CDATA[Nearly a year after I had my tentative say about supplemental results, my thinking on what causes supplemental results has shifted away from duplicate content and moved more toward lack of inbounds and internal PageRank distribution, as Matt Cutts recently stated that PageRank is the primary factor in determining supplemental results.
Back around May 18, 2006, [...]]]></description>
			<content:encoded><![CDATA[<p>Nearly a year after I had my tentative say about <a href="/notes/supplementals.html">supplemental results</a>, my thinking on what causes supplemental results has <strong>shifted away from duplicate content</strong> and moved more toward lack of inbounds and internal PageRank distribution, as Matt Cutts recently stated that <a href="/blog/2006/10/12/matt-cutts-pagerank-primary-factor-determining-supplemental-results.html">PageRank is the primary factor in determining supplemental results</a>.</p>
<p>Back around May 18, 2006, Matt Cutts had this to say about combatting supplemental results by optimizing internal link structure:</p>
<blockquote><p>typically the depth of the directory doesn’t make any difference for us; PageRank is a much larger factor. So without knowing your site, <strong>I’d look at trying to make sure that your site is using your PageRank well.</strong> A tree structure with a certain fanout at each level is usually a good way of doing it.</p></blockquote>
<p>Some people questioned my thinking when I advocated optimizing your internal link structure for better PageRank distribution, probably because they just hate hearing the word <em>PageRank</em>.</p>
<h2>The Word PageRank Causes Friction (tangent)</h2>
<p>So I&#8217;ve been replacing the word &#8220;PageRank&#8221; with words like &#8220;authority&#8221;, &#8220;trust&#8221;, &#8220;link juice&#8221;, &#8220;link value&#8221;, &#8220;visibility&#8221; or &#8220;link weight.&#8221; If I say &#8220;your page doesn&#8217;t have enough PageRank,&#8221; that makes some people:</p>
<ol>
<li><strong>cringe, like they&#8217;re at the dentist, when suddenly the dentist whips out a dental drill.</strong> I think its some kinda pavlovian response - the word &#8220;PageRank&#8221; just make some people feel ill.</li>
<li><strong>look at the toolbar</strong>, and tell me &#8220;hey, this PR 0 page is in the main index, while this PR 6 page is marked supplemental. So you&#8217;re obviously wrong.&#8221; First of all, please don&#8217;t say &#8220;PR.&#8221; PR means Public Relations or Peurto Rico. Second of all, when I say TBPR, I&#8217;m talking about the Toolbar PR. When I say PageRank, I&#8217;m talking about internal PageRank, the one you can&#8217;t see. Just because the toolbar says 0 doesn&#8217;t mean there are no links to that page.</li>
<li><strong>argue with me because he/she is conditioned to think <a href="/blog/2006/11/24/pagerank-doesnt-matter-is-now-officially-an-seo-myth.html">PageRank doesn&#8217;t matter</a>.</strong> She might tell me &#8220;PR is just one of 100 factors in Google&#8217;s algorithm and you should ignore it and focus on getting quality links by publishing quality content and marketing it well instead of spending hours nofollowing links to your privacy policy.&#8221; No wonder people claim &#8220;SEO is easy&#8221; when SEOs advocate ignoring internal link structure optimization in favor of &#8220;quality content.&#8221; Not that I disagree with that recommendation, but I think if I&#8217;m <em>optimizing</em> a site, my focus would lie more on the nuts and bolts instead of on copywriting or buying ad space.</li>
</ol>
<p>Seriously, some people just get stupid when they hear the word &#8220;PageRank&#8221;, so to get around that debacle, I say &#8220;your page doesn&#8217;t have enough trust&#8221; or &#8220;your page doesn&#8217;t have enough quality inbound links&#8221; or &#8220;your page lacks visibility.&#8221; But really, <strong>all I&#8217;m saying is your page doesn&#8217;t have enough internal PageRank.</strong></p>
<p>/tangent</p>
<h2>Why Duplicate Content Is Responsible for Supplemental Results</h2>
<p>Anyway, back to duplicate content. I&#8217;m starting to read some people saying <strong>duplicate content doesn&#8217;t matter at all when it comes to supplemental results.</strong> I disagree.</p>
<p>If you have two identical/similar content pages, both with thousands of authority inbound links, those pages are not supplemental-index-bait. Google&#8217;s solution to duplicate content is not the supplemental index. On the contrary, <strong>Googlers insist that they <a href="http://videos.webpronews.com/2006/12/06/vanessa-fox-clarifies-the-role-of-google-sitemaps/">filter out duplicate content.</a></strong></p>
<p>But when you&#8217;ve got links to domain.com and domain.com/index.php, you&#8217;re splitting link juice between two pages instead of one.When you got multiple urls like index.php?sessionId=290302342, index.php?sessionId=20343400, and index.php?sessionId=023123321 generating identical content and you got links to all of those urls, again, you&#8217;re splitting link weight between all those pages. The urls with just a couple of low value inbound links end up &#8220;going supplemental&#8221; while urls with tons of other sites linking gets to sit pretty in the main index.</p>
<p>So yeah, you can have <a href="http://www.google.com/search?hl=en&#038;safe=off&#038;client=firefox-a&#038;rls=org.mozilla%3Aen-US%3Aofficial&#038;hs=MKk&#038;q=%22It%27s+been+a+mystery+and+still+they+try+to+see%22&#038;btnG=Search" rel="nofollow">multiple duplicate content pages in the main index</a> as long as they&#8217;re well-linked-to. But links to multiple urls hosting identical content within your own site will dilute PageRank and cause some link-starved pages to &#8220;go supplemental.&#8221;</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/02/19/why-duplicate-content-causes-supplimental-results.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Results Don&#8217;t Always Say They&#8217;re Personalized</title>
		<link>http://seo4fun.com/blog/2007/02/03/results-dont-always-say-theyre-personalized.html</link>
		<comments>http://seo4fun.com/blog/2007/02/03/results-dont-always-say-theyre-personalized.html#comments</comments>
		<pubDate>Sat, 03 Feb 2007 17:47:02 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/02/03/results-dont-always-say-theyre-personalized.html</guid>
		<description><![CDATA[Aaron Wall and Graywolf reported that Google is turning off personalization notification:
In the past they typically placed a turn off personalized results whenever your results were personalized, but now they do not disclose when they are personalizing the results - Aaron Wall
If Aaron’s correct that they are increasing personalized search and turning off notification, and [...]]]></description>
			<content:encoded><![CDATA[<p><a href="http://www.seobook.com/archives/002023.shtml" rel="nofollow">Aaron Wall</a> and <a href="http://www.wolf-howl.com/google/google-is-not-my-mother-and-should-kill-personalized-search/" rel="nofollow">Graywolf</a> reported that Google is turning off personalization notification:</p>
<blockquote><p>In the past they typically placed a turn off personalized results whenever your results were personalized, but <strong>now they do not disclose when they are personalizing the results</strong> - Aaron Wall</p></blockquote>
<blockquote><p>If Aaron’s correct that they are increasing personalized search and turning off notification, and I have no reason to doubt him&#8230; - Graywolf</p></blockquote>
<p>Has Google gone completely mad? I ran a few searches while logged into my Google account. Here&#8217;s a couple screenies of what I found:</p>
<p><img src="/images/personalized.jpg" alt="Google personalized search" /></p>
<p>A search for &#8220;search engines&#8221; returns personalized search results. You can see the text &#8220;Personalized results 1-100 of about&#8230;&#8221; above the digg subscribed link. The text &#8220;Turn off Personalized Search (Beta) for these results&#8221; is gone.</p>
<p><img src="/images/not-personalized.jpg" alt="Google personalized search turned off" /></p>
<p>A search for &#8220;paris hilton&#8221; returns ordinary results. Notice the usual text &#8220;Results 1-100 of about&#8230;&#8221; above the results.</p>
<p>After having run a few other searches, these are personalized:</p>
<p>&#8216;mysql tree&#8221;<br />
&#8220;ajax&#8221;<br />
&#8220;youtube&#8221;<br />
&#8220;wired magazine&#8221;</p>
<p>These aren&#8217;t personalized:</p>
<p>&#8220;britney spears&#8221;<br />
&#8220;cpr&#8221;<br />
&#8220;how to bake cake&#8221;<br />
&#8220;is baking cake easy&#8221;<br />
&#8220;is seo easier than baking cake&#8221;</p>
<p>It&#8217;s true that personalization used to be an option I could opt out of while remaining logged into my Google account, but now its no longer optional. Like <a href="http://googleblog.blogspot.com/2007/02/personally-speaking.html">Sep says on Googeblog</a>, &#8220;If you don&#8217;t want to see personalized results, just <strong>sign out of your Google Account.</strong>&#8221; I do think that personalization should remain optional; I shouldn&#8217;t have to log out of my Google account to get rid of personalization. But Google still does tell you when results are being personalized. Of course, if you use Google alot, after a while more results will be personalized than not.</p>
<h2>Is Google being Evil?</h2>
<p>First, repeat after me: <a href="http://en.wikipedia.org/wiki/Wall_Street_(film)">Greed, for lack of a better word, is good</a>. Greed is right; greed works. Greed clarifies, cuts through, and captures the essence of the evolutionary spirit.</p>
<p>But seriously, Google isn&#8217;t restricting your choices by personalizing results. Google is just shifting  the order of items on the menu, moving down sites you hate, and ranking sites you like slightly higher. So saying personalization is like being force-fed by Google is like saying Paris Hilton is the best woman I&#8217;ve ever seen.</p>
<p>Still, I don&#8217;t quite get why search history is dependent on personalization. Like <a href="http://searchengineland.com/070202-224617.php">Danny Sullivan says</a>, personalized results is dependent on search history, not the other way around. Technically, search history doesn&#8217;t need personalized results option activated in order for it to work:</p>
<blockquote><p>
&#8220;Creating a Google Account will enable Search History. Search History is a feature that will provide you with a more personalized experience on Google that includes more relevant search results and recommendations.&#8221;</p></blockquote>
<h2>So How Does this Personalization Stuff Work?</h2>
<p>This Personalization-is-the-enemy-of-all-SEOs-either-living-or-dead debate got me kind of curious about my search history and how it relates to personalized results. Here are some of the queries I used since yesterday:</p>
<p>&#8220;wikipedia long term&#8221; <strong>personalized</strong><br />
&#8220;javascript get element TD&#8221; <strong>Not</strong><br />
&#8220;david bowie changes lyrics&#8221; <strong>Not personalized</strong><br />
&#8220;fire hydrant &#8221; <strong>not personalized</strong>.<br />
&#8220;fire hydrant&#8221; <strong>personalized</strong>.  WTH? But this effect disappeared after I submitted the query a few more times.<br />
&#8220;Modified Preorder tree traversal&#8221; <strong>nope</strong><br />
&#8220;Modified Preorder tree&#8221; <strong>personalized</strong><br />
&#8220;Modified Preorder posts&#8221; <strong>personalized</strong><br />
&#8220;Modified Preorder algorithm&#8221; <strong>personalized</strong><br />
&#8220;Modified Preorder algorithm posts&#8221; <strong>nay</strong></p>
<p>Does Google not like queries longer than 3 words when it comes to personalization?</p>
<p>Let&#8217;s try this again:</p>
<p>&#8220;wikipedia long term&#8221; <strong>personalized</strong><br />
&#8220;wikipedia long&#8221; <strong>personalized</strong><br />
&#8220;wikipedia long term memory&#8221; <strong>niet</strong></p>
<p>&#8220;half life 2 entanglement&#8221; <strong>nicht personifiziert</strong><br />
&#8220;half life entanglement&#8221; <strong>personalizado</strong></p>
<p>last none:</p>
<p>&#8220;google base&#8221; персонализировано<br />
&#8220;google base store&#8221; <strong>personnalisé</strong><br />
&#8220;google base store help&#8221; <strong>ordinary results</strong>.</p>
<p>Ok, my head is starting to hurt.</p>
<p>Interested in digging a little deeper? <a href="http://www.seobythesea.com/?cat=22">A few personalization patents</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/02/03/results-dont-always-say-theyre-personalized.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Long Tail De Jour: what&#8217;s the point in adding more pages to my site if google doesn&#8217;t index them or puts them into the supplemental index?</title>
		<link>http://seo4fun.com/blog/2007/01/23/long-tail-de-jour-whats-the-point-in-adding-more-pages-to-my-site-if-google-doesnt-index-them-or-puts-them-into-the-supplemental-index.html</link>
		<comments>http://seo4fun.com/blog/2007/01/23/long-tail-de-jour-whats-the-point-in-adding-more-pages-to-my-site-if-google-doesnt-index-them-or-puts-them-into-the-supplemental-index.html#comments</comments>
		<pubDate>Tue, 23 Jan 2007 11:12:27 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/01/23/long-tail-de-jour-whats-the-point-in-adding-more-pages-to-my-site-if-google-doesnt-index-them-or-puts-them-into-the-supplemental-index.html</guid>
		<description><![CDATA[A couple days ago, someone hit my site with this longtail:
what&#8217;s the point in adding more pages to my site if google doesn&#8217;t index them or puts them into the supplemental index
Crazy, long huh? :) When I start talking to myself, my friends start to worry. When I start talking to Google, I need a [...]]]></description>
			<content:encoded><![CDATA[<p>A couple days ago, someone hit my site with this longtail:</p>
<blockquote><p>what&#8217;s the point in adding more pages to my site if google doesn&#8217;t index them or puts them into the supplemental index</p></blockquote>
<p>Crazy, long huh? :) When I start talking to myself, my friends start to worry. When I start talking to Google, I need a doctor :D</p>
<p><em>The obvious answer:</em> Forget Google. Build your site for visitors.<br />
<em>Reply:</em> What visitors? Where am I gonna get the traffic if I don&#8217;t show up on Google?<br />
<em>Counter reply:</em> There are tons of other ways of generating traffic besides Google. There&#8217;s always MSN and Yahoo. Besides, people can see you instantly on Technorati if you publish a post on a blog, or *gasp* bookmark yourself on del.icio.us.<br />
<em>Rebuttal:</em> Yeah right. MSN sends me like 1 hit a week. That&#8217;ll buy me a pair of shoes maybe if I wait like 5 months.<br />
<em>Me:</em> That&#8217;s your problem. You&#8217;ve got the unenlightened self-interest bug.<br />
<em>Reply:</em> The&#8230;.what??</p>
<p>Definition of <strong>selfishness</strong>, according to Wikipedia:</p>
<blockquote><p>a selfish person deliberately f<strong>ocuses on his own agenda</strong>, rather than that of others.</p></blockquote>
<p>But wait a minute:</p>
<blockquote><p>un-selfishness is <strong>a deliberate act</strong>, rather than selfishness, which tends to occur naturally. <strong>In the animal kingdom, few species exhibit unselfish behavior.</strong></p></blockquote>
<p>Not to mention selfishness isn&#8217;t all bad:</p>
<blockquote><p>Selfishness has <strong>some good qualities</strong> such as productivity or the taking of personal responsibility. One view is that since one needs to act in a mainly self-interested way in order to advance in life doing so should not be regarded as wrong, or labelled as harmful or inappropriate.</p></blockquote>
<p>Now check out &#8220;rational selfishness&#8221; or &#8220;rational self-interest&#8221;:</p>
<blockquote><p>
The philosophy holds that individuals should not act on momentary self-interested whims but on what is in their long-term self-interest, which is defined to require respecting the individual liberty of others by refraining from initiating coercion against them.
</p></blockquote>
<p>And we finally end up with <strong>Enlightened self-interest</strong>:</p>
<blockquote><p>Enlightened self-interest is a philosophy in ethics which states that persons who act <strong>to further the interests of others</strong> (or the interests of the group or groups to which they belong), ultimately serve their own self-interest.</p></blockquote>
<p>So what&#8217;s <a href="http://en.wikipedia.org/wiki/Enlightened_self-interest" rel="nofollow">unenlightened self-interest</a>?</p>
<blockquote><p>Unenlightened self-interest, in which it is argued that when most or <strong>all persons act according to their own myopic selfishness</strong>, that the group suffers loss as a result of conflict, decreased efficiency because of lack of cooperation, and the increased expense each individual pays for the protection of their own interests.</p></blockquote>
<p>Oook, so now you&#8217;re probably thinking, what the hell does this unenlightened self-interest thingie got to do with getting indexed in Google?</p>
<p>Here&#8217;s a shocking revelation: some people care more about money than about putting out <strong>the best product</strong> in their niche. If you&#8217;re Shoemoney, you don&#8217;t need organic ranking to make big bucks. You really don&#8217;t even need a website to make money with PPC. But if you want to rank high in organic search results, what in the God&#8217;s name do you think you&#8217;re doing trying to get to the top with a mediocre site no one in the right mind wants to recommend? :)</p>
<p>What&#8217;s a good business model? <a href="http://www.seobomb.com/how-to-amaze-your-clients-with-a-killer-viral-campaign/">Amazing your clients</a>. You <a href="http://www.johnchow.com/time-for-a-commercial-break/">under promise and over deliver</a>. Want to rank #1 on Google? <strong>Build a site that deserves to be #1.</strong> Focus on what other people want and contribute to society instead of obsessing about how much money you need to pay your mortage or buy another game on your PS3. What&#8217;s that old saying?</p>
<p>You reap what you sow.</p>
<p>P.S. Yeah I realize this post sounds preachy, but its good to get hit over the head with a hammer once in a while to maintain perspective, don&#8217;t you think? :)</p>
<p><strong>UPDATE:</strong> Matt Cutts asks &#8220;<a href="http://www.mattcutts.com/blog/algorithm-to-reduce-googlebomb-impact/#comment-95036">What value is my site offering to users?</a>&#8220;</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/01/23/long-tail-de-jour-whats-the-point-in-adding-more-pages-to-my-site-if-google-doesnt-index-them-or-puts-them-into-the-supplemental-index.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>SEO Myth: There is No Duplicate Content Penalty</title>
		<link>http://seo4fun.com/blog/2007/01/05/seo-myth-there-is-no-duplicate-content-penalty.html</link>
		<comments>http://seo4fun.com/blog/2007/01/05/seo-myth-there-is-no-duplicate-content-penalty.html#comments</comments>
		<pubDate>Fri, 05 Jan 2007 09:59:52 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Duplicate Content]]></category>

		<category><![CDATA[Google]]></category>

		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2007/01/05/seo-myth-there-is-no-duplicate-content-penalty.html</guid>
		<description><![CDATA[This is probably old news to black hats, but I often hear people say &#8220;there&#8217;s no duplicate content penalty.&#8221; Newbies worry they&#8217;ll incur some kind of penalty for having identical copyright text across 100 pages or something, and other people like me jump in to alleviate their fears: &#8220;Google doesn&#8217;t penalize duplicate content; it filters [...]]]></description>
			<content:encoded><![CDATA[<p>This is probably old news to black hats, but I often hear people say &#8220;there&#8217;s no duplicate content penalty.&#8221; Newbies worry they&#8217;ll incur some kind of penalty for having identical copyright text across 100 pages or something, and other people like me jump in to alleviate their fears: &#8220;Google doesn&#8217;t penalize duplicate content; it filters them out.&#8221;</p>
<p>However, a few months ago, back when I still believed supplemental results were largely due to duplicate content, I ran a test to try to figure out exactly what % similarity I had to hit for pages to squeeze into the main index. I created a directory with several pages: one original page, then several other similar pages of varying similarity, from 90% similar down to 20%. Initially, all the pages got into the main index. Then after a few months, the entire directory poofed. If Google was filtering duplicate content, then I&#8217;d assume the original page, at least, to remain indexed. I also expected a page that was only 20% similar to stay in the index. But no. Every page in that directory disappeared from both the main and supplemental index. At that point, I suspected that Googlebot was refusing to index any page in that directory.</p>
<p>Here&#8217;s the logic. Say I have a site with 245,230,991 pages and at least 60% of those pages are very similar. Does Googlebot really want to spend the time and effort to crawl all those pages? Keep in mind, since Big Daddy, Google&#8217;s been very picky with what pages to crawl and index. PageRank became an anti-spammer weapon built to protect Googlebot from crawling thousands of low-value spam pages with nothing but guest book links pointing at them. So if Googlebot thinks that a good number of pages within a directory are too similar, then it would make sense to not only filter those pages out but to prevent future crawling of any pages in that directory.</p>
<p>Caveman says something similar in <a href="http://www.webmasterworld.com/forum30/31430.htm">this post</a> started way back in Sep. 29, 2005 (several months before Big Daddy):</p>
<blockquote><p>The fact that even within a single site, when pages are deemed too similar, G is not throwing out the dups - they&#8217;re throwing out ALL the similar pages&#8230;if they find four pages on the same site about a certain kind of bee, and the four pages are similarly structured, and one is a main page for that bee, and the other three are subpages about the same bee, each reflecting a variation of that bee, the site owner now seems to run the risk that they will find all of the pages too similar, and filter them all, not just the three subpages.</p></blockquote>
<p>Anyway, today I was re-reading a <a href="http://www.webmasterworld.com/google/3192967-3-50.htm">Webmasterworld thread</a> regarding Adam Lasnik&#8217;s Duplicate Content post, and happened on a few interesting comments Adam wrote:</p>
<p>Some guy asked: Why not build into your webmaster toolkit something like a &#8220;Duplicate Content&#8221; threshold meter?</p>
<p>Adam responds:</p>
<blockquote><p>
The fact that duplicate content isn&#8217;t very cut and dry for us either (e.g., it&#8217;s not &#8220;if more than [x]% of words on page A match page B&#8230;&#8221;) makes this a complicated prospect.</p></blockquote>
<p><a href="http://www.stuntdubl.com/2006/06/12/dupe-content/">Todd</a> wrote about this 6 months ago:</p>
<blockquote><p>If it was as easy as saying that any page with more than 42% duplicate content will be filtered from the search results, then all site owners and SEO’s would probably grab 40% duplicate content for every page filler. It IS NOT a percentage.</p></blockquote>
<p>Maybe I should go over to <a href="http://www.seobythesea.com/">Bill Slawski&#8217;s blog</a> and search for <a href="http://www.google.com/search?q=site%3Aseobythesea.com+duplicate+content&#038;ie=utf-8&#038;oe=utf-8&#038;rls=org.mozilla:en-US:official&#038;client=firefox-a">some duplicate content patents</a>.</p>
<p>In regards to duplicate content penalty (emphasis mine), Adam says:</p>
<blockquote><p>As I noted in the original post, penalties in the context of duplicate content <strong>are rare</strong>. Ignoring duplicate content or just picking a canonical version is MUCH more typical&#8230;Again, this very, very rarely triggers a penalty. <strong>I can only recall seeing penalties</strong> when a site is perceived to be particularly &#8220;empty&#8221; + redundant; e.g., a reasonable person looking at it would cry &#8220;krikey! it&#8217;s all basically the same junk on every page!&#8221;</p></blockquote>
<p>So if I take Adam&#8217;s word for it, Google does penalize sites for duplicate content, though its a once on a DVD night kinda thing (I cancelled my Netflix like a year ago).</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2007/01/05/seo-myth-there-is-no-duplicate-content-penalty.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Sex Blogs Tank Over Christmas - Matt Cutts Responds</title>
		<link>http://seo4fun.com/blog/2006/12/28/sex-blogs-tank-in-organic-serps-over-christmas.html</link>
		<comments>http://seo4fun.com/blog/2006/12/28/sex-blogs-tank-in-organic-serps-over-christmas.html#comments</comments>
		<pubDate>Thu, 28 Dec 2006 15:00:04 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/12/28/sex-blogs-tank-in-organic-serps-over-christmas.html</guid>
		<description><![CDATA[Just an early morning post about a Boing Boing post titled Google &#8220;disappears&#8221; sex blogs? Something&#8217;s broken. For the main dish, you can go read their post, but I did find an interesting trail left by Matt Cutts on Comstock Films comment section, dated Dec. 27, 2006:
Hey, I’m an engineer at Google. I just wanted [...]]]></description>
			<content:encoded><![CDATA[<p>Just an early morning post about a Boing Boing post titled <a href="http://www.boingboing.net/2006/12/27/google_disappears_se.html" rel="nofollow">Google &#8220;disappears&#8221; sex blogs? Something&#8217;s broken</a>. For the main dish, you can go read their post, but I did find an interesting trail left by Matt Cutts on <a href="http://www.comstockfilms.com/blog/tony/2006/12/27/will-google-kill-comstock-films" rel="nofollow">Comstock Films</a> comment section, dated Dec. 27, 2006:</p>
<blockquote><p>Hey, I’m an engineer at Google. I just wanted to say that different people at Google saw this and were asking about it, so we’ll check out these reports from the sites such as tiny nibbles and others.</p></blockquote>
<p>On the 23rd, Matt said in a blog post: </p>
<blockquote><p>I know for a fact that there haven’t been any major algorithm updates to our scoring in the last few days</p></blockquote>
<p>I doubt Google did anything major between Dec 23 and Dec 27th, and I must say the hype I&#8217;ve read so far is a little over the top, but I&#8217;m not in the habit of jumping to conclusions either way.</p>
<p>Since no one is actively competing against these sex blogs for terms like &#8220;Violet Blue&#8221;, &#8220;Tiny Nibbles&#8221; or &#8220;Comstock Films&#8221;, a drop in ranking is a big signal - but of what? Algorithm shift? I doubt it. A bug at the Plex? More likely. Also, its interesting to note that these blogs are already back on #1 for their site names, which was their biggest complaint.</p>
<p>I don&#8217;t buy the Adsense / organic search conspiracy theory and I also don&#8217;t buy the SEO Blackhat theory about Google actively suppressing adult sites from page one, though I do see it happening if people who ran Google Ads ever took over Google&#8217;s Search Quality team.</p>
<p>BTW, did you notice SEO Blackhat is <a href="http://www.google.com/search?q=free+porn&#038;ie=utf-8&#038;oe=utf-8&#038;rls=org.mozilla:en-US:official&#038;client=firefox-a" rel="nofollow">on page one for &#8220;free porn&#8221;</a> for a blog post that&#8217;s no more than one month old (published Nov 21, 2006)? <a href="http://search.msn.com/results.aspx?q=link%3Ahttp%3A%2F%2Fseoblackhat.com%2F2006%2F11%2F21%2Ffree-porn%2F&#038;FORM=MSNH" rel="nofollow">350+ clean IBLs</a> and bam, you leave hundreds of thousands of cheap, bad-neighborhood link trades in the dust.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/12/28/sex-blogs-tank-in-organic-serps-over-christmas.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Happy Holidays Guys</title>
		<link>http://seo4fun.com/blog/2006/12/24/happy-holidays-guys.html</link>
		<comments>http://seo4fun.com/blog/2006/12/24/happy-holidays-guys.html#comments</comments>
		<pubDate>Mon, 25 Dec 2006 01:50:05 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/12/24/happy-holidays-guys.html</guid>
		<description><![CDATA[I&#8217;m still kinda irked at myself for not buying myself a PS3 this Xmas, but I guess I&#8217;ll get over it. Anyway, I hope you guys are having a great holiday.
The real reason I&#8217;m posting though, is I was looking through my traffic tracker log tonight, and I saw another hit on this page I [...]]]></description>
			<content:encoded><![CDATA[<p>I&#8217;m still kinda irked at myself for not buying myself a PS3 this Xmas, but I guess I&#8217;ll get over it. Anyway, I hope you guys are having a great holiday.</p>
<p>The real reason I&#8217;m posting though, is I was looking through my traffic tracker log tonight, and I saw another hit on <a href="/notes/supplementals.html">this page I wrote about supplemental results</a> for the term &#8220;supplemental hell.&#8221; It&#8217;s a zero-competition (~525 websites referencing the exact phrase), close-to-nil traffic query term (0 searches a month according to Overture), and makes me absolutely no money, but I&#8217;ve been stuck in 2nd place below Jim Boykin for a while, and I wanted to try ranking one spot higher.</p>
<p><img src="/images/supplemental-hell-1.jpg" alt="supplemental results" /></p>
<p>You might have noticed I recently added &#8220;aka supplemental hell&#8221; to the page title, though I doubt that was enough, since ranking stayed the same when Google refreshed the page&#8217;s cache. And those cheap blogdrive links have been there since, like, forever.</p>
<p>Ok, so at least temporarily, I got what I wanted. But this is a good reminder to not neglect keyword research - because, the truth is, ranking for &#8220;supplemental hell&#8221; doesn&#8217;t do me any good - at least in the short term - for paying my bills. In contrast, I got a different non-SEO related page ranking 1st for a money term right now, and with conversion at ~1:150 (last period: 1:161; previous period:1:197; peroid before that:1:62), sales just keep coming and coming (knock on wood). And the difference in conversions once the page hit 1st place was rather dramatic. I was like, &#8220;where the hell are all these sales coming from? And what&#8217;s up with ultra low ratios?&#8221; (previously, I was converting at ~1:600+, and niche average ratio can fall anywhere between 1:300 and 1:1200. 1:62 is like, insanely good, though I admit I have another page averaging 1:35ish over the last 3 months.).</p>
<p>So I strongly disagree with those who think SEO is not about ranking. It&#8217;s not JUST about ranking - hopefully that we can all agree on.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/12/24/happy-holidays-guys.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>WMW Google Bashing Over Reciprocal Links</title>
		<link>http://seo4fun.com/blog/2006/12/19/wmw-google-bashing-over-reciprocal-links.html</link>
		<comments>http://seo4fun.com/blog/2006/12/19/wmw-google-bashing-over-reciprocal-links.html#comments</comments>
		<pubDate>Tue, 19 Dec 2006 13:57:47 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/12/19/wmw-google-bashing-over-reciprocal-links.html</guid>
		<description><![CDATA[Some people over at WMW are going ballistic over Stephanie&#8217;s recent article on Google Webmaster Central Blog about &#8220;non-earned&#8221; links, titled Building link-based popularity. These guys who, I guess, pay their bills by swapping and buying links are miffed over the thought of losing their Google ranking and are shooting the messanger. A big chunk [...]]]></description>
			<content:encoded><![CDATA[<p><a href="http://www.webmasterworld.com/link_development/3190790.htm">Some people</a> over at WMW are going ballistic over Stephanie&#8217;s recent article on Google Webmaster Central Blog about <a href="/blog/2006/12/16/still-plenty-of-loopholes-in-googles-paid-link-detection-algo.html">&#8220;non-earned&#8221; links</a>, titled <em>Building link-based popularity</em>. These guys who, I guess, pay their bills by swapping and buying links are miffed over the thought of losing their Google ranking and are shooting the messanger. A big chunk of the thread is pure anti-Google noise, but here&#8217;s a few &#8220;note-to-self&#8221; kinda quotes I found interesting.</p>
<p>For example, Adam Lasnik steps in to clear up the confusion in a thread inaptly titled <em>It&#8217;s Official: Google Discounting <strong>Reciprocal Link Exchanges</strong></em> , started by martinibuster, a mod for the WMW link building forum:</p>
<blockquote><p> Whoa!<br />
This is a lot of speculation about reciprocal linking in response to an official blog entry, when <strong>there&#8217;s not even one mention of &#8220;reciprocal&#8221; on the entire page</strong> ;-). Take a step back, look at the bigger picture, take a deep breath!
</p></blockquote>
<p>He continues, debunking a claim that the blogspot post was just a Googler&#8217;s opinion:</p>
<blockquote><p>
What part of &#8220;Official&#8221; in the title didn&#8217;t resonate with you? :). The people who write on our Webmaster blog are either <strong>engineers</strong> or <strong>product managers</strong> &#8212; or those who work directly with them &#8212; in Search Quality and Webmaster Tools.</p></blockquote>
<p>Ok, so it&#8217;s official. Got it.</p>
<p>Then he elaborates on the difference between a reciprocal link <em>intended as a genuine citation</em> versus a reciprocal link <em>intended to increase PageRank</em>:</p>
<blockquote><p>If a Webmaster is engaging in reciprocal linking <strong>in a way that clearly indicates to us that he or she is doing so to garner PageRank</strong>, not out of a genuine interest for that other site&#8230; well, that&#8217;s the sort of linking scheme we don&#8217;t see as very user-friendly. Are we apt to ban that Webmaster&#8217;s site? I highly doubt it. Are we likely to value those links less? Quite possibly.</p></blockquote>
<p>Glengara, a senior WMW member, makes a good distinction: </p>
<blockquote><p><strong>A reciprocal link can just be coincidental, an exchanged one denotes some deliberation</strong>, and it&#8217;s the deliberate targeting of the PR algo through linkage that the blogpost is all about.</p></blockquote>
<p><strong>Note to self:</strong> From now on, I&#8217;m not going to use the term &#8220;reciprocal&#8221; links to refer to traded links. I&#8217;m going to say &#8220;exchanged&#8221; or &#8220;traded&#8221; links instead.</p>
<p>In regards to Google penalizing sites for exchanged links, Adam calms all fears:</p>
<blockquote><p>&#8220;I hope not. That isn&#8217;t reality. Our aim isn&#8217;t to penalize sites, it&#8217;s to deftly determine when <strong>and to what extent</strong> a link is indeed a &#8220;vote&#8221; for a site.&#8221;</p></blockquote>
<p>Marcia, another Senior WMW member, also reminds people of an Adam Lasnik quote posted in that long-ass Supplemental Thread in Google Groups: </p>
<blockquote><p>The key here is, indeed, moderation :). If, say, 90% of your backlinks are reciprocal, that&#8217;s probably not going to improve how our algorithms view your site. Or worse, if 90% of your backlinks are reciprocal and not likely to be of interest to your user.</p>
<p>But exchanging links here and there &#8212; *especially* when<br />
done with clear editorial judgement (e.g., you&#8217;re not just<br />
accepting dozens of link exchanges willy-nilly) &#8212; that&#8217;s<br />
not the sort of thing Google looks down upon. </p></blockquote>
<p>To sum up, <strong>reciprocal linking isn&#8217;t dead.</strong> To what extent Google devalues those types of links depends, based on what we&#8217;ve heard from Googlers so far, on a few factors, including 1) intent, 2) topical relevance, 3) percentage of reciprocal links to the total number of inbound links, and 4) most likely the level of trust Google has with the sites you&#8217;re linking out to / linked from (number of pages a site has in Google&#8217;s main index versus number of actual pages is one indicator).</p>
<p><strong>UPDATE:</strong> I just read <a href="http://www.seomoz.org/blogdetail.php?ID=1597" rel="nofollow">Rand&#8217;s take</a> on the whole issue (Yeah I nofollowed it - it feels like he&#8217;s baiting). To be frank, I&#8217;m a little surprised by his attempt at working the tired &#8220;there&#8217;s nothing wrong with paid links&#8221; angle. Like I posted over at SEOmoz, buying links is gaming the system.  &#8220;I admit paying for links skews search results in my favor, but as long as I don&#8217;t get caught, I&#8217;ll keep doing it.&#8221; - a statement like that I&#8217;d had no problems with.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/12/19/wmw-google-bashing-over-reciprocal-links.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Blog Tag: Five Things You Didn&#8217;t Know about Me</title>
		<link>http://seo4fun.com/blog/2006/12/17/blog-tag-five-things-you-didnt-know-about-me.html</link>
		<comments>http://seo4fun.com/blog/2006/12/17/blog-tag-five-things-you-didnt-know-about-me.html#comments</comments>
		<pubDate>Sun, 17 Dec 2006 13:40:48 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/12/17/blog-tag-five-things-you-didnt-know-about-me.html</guid>
		<description><![CDATA[Thanks alot John for tagging me so I can spend my Sunday morning in front of a computer - before the sun even has a chance to rise :D Here&#8217;s five things you didn&#8217;t know about me and would have never ever guessed:
1. I never got drunk, smoked cigs, or did any drugs till I [...]]]></description>
			<content:encoded><![CDATA[<p>Thanks alot <a href="http://seside.net/">John</a> for tagging me so I can spend my Sunday morning in front of a computer - before the sun even has a chance to rise :D Here&#8217;s five things you didn&#8217;t know about me and would have never ever guessed:</p>
<p>1. <strong>I never got drunk, smoked cigs, or did any drugs till I went to <a href="http://admissions.vassar.edu/tour/tourlaunch.html">Vassar College</a></strong>. I started out from &#8220;what&#8217;s sex on the beach?&#8221; to stone drunk every night by the end of my first semester. Oh, and yeah, coffee IS a drug.</p>
<p>2. <strong>I majored in Comp Sci / Cognitive Science because I didn&#8217;t believe in God or the existence of a human soul.</strong> By creating a machine that can think, I&#8217;d prove that man is just a complex machine. I guess I was in part rebelling against my Catholic background. BTW, back then, I was answering emails on a VAX, writing papers on MacWrite, and the internet didn&#8217;t even exist.</li>
<p>3. <strong>When I was 21, the local media reported I was missing and believed to be dead.</strong> On their front page, they wrote: &#8220;a straight-A Vassar Student (close to flunking out is more like it) has been missing for 3 days and was last seen with a major drug kingpin, who at this time is in police custody. He claims he did not kill [me] and doesn&#8217;t know of his whereabouts.&#8221; </p>
<p>It&#8217;s weird to see yourself on the front cover of a newspaper, to see yourself when you turn on the evening news, to have kids run up to you asking for autographs, or to hear a Yale English teacher call you a celebrity.</p>
<p>But it&#8217;s a long story. Plus this is an SEO blog, remember?</p>
<p>4. <strong>Before working on the Web, I played pool.</strong> I&#8217;m ranked 7 in <a href="http://www.poolplayers.com/">APA</a> National Team Championships, and unless you&#8217;re an A~Open player, I&#8217;ll give you a beating on the pool table :p I&#8217;ve won a free trip to Vegas a few years ago, which includes a roundtrip ticket to Vegas, and a few nights stay at the Riviera (because its probably the crappiest hotel there. A teammate gave up his free room and blew 5K on a room at the Belagio). I went to NYC and gambled away $200 bucks the night before the trip, and went there with only $100 bucks in my pocket. I ran 11 racks of 9 ball on a bar table playing a 2 am tourney and won a couple hundred from that, plus my team earned me a couple more hundred bucks. The food was near-free with all the buffets, so the trip eneded up costing me nothing (Nope, I didn&#8217;t gamble or blow money away in strip joints).</p>
<p>5. <strong>I met the love of my life when I was 21.</strong> Unlike Todd or Aaron, my relationship started off in a zoo (literally).</p>
<p>Now, I&#8217;m gonna tag&#8230;<a href="http://seobloguk.blogspot.com/">Lisa Ditlefsen</a> (let&#8217;s just all admit it - her avatar is beyond hott),  <a href="http://www.jimwestergren.com/">Jim Westergren</a>, <a href="http://www.jimboykin.com/">Jim Boykin</a> (someone musta tagged him already..but whatev), <a href="http://www.seo-scoop.com/">Donna Fontenot</a>, and <a href="http://seo-theory.blogspot.com/">Michael Martinez</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/12/17/blog-tag-five-things-you-didnt-know-about-me.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Still Plenty of Loopholes in Google&#8217;s Paid Link Detection Algo</title>
		<link>http://seo4fun.com/blog/2006/12/16/still-plenty-of-loopholes-in-googles-paid-link-detection-algo.html</link>
		<comments>http://seo4fun.com/blog/2006/12/16/still-plenty-of-loopholes-in-googles-paid-link-detection-algo.html#comments</comments>
		<pubDate>Sat, 16 Dec 2006 14:25:41 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/12/16/still-plenty-of-loopholes-in-googles-paid-link-detection-algo.html</guid>
		<description><![CDATA[Warning: I&#8217;m just rambling to get a free link on Google&#8217;s blog. Why do I want that? Cuz I&#8217;m bored? :)
There&#8217;s not much new in Google&#8217;s recent post, Building Link-Based Popularity, in which Stephanie Ulrike underscores Google&#8217;s increasing ability to find and devalue artificial links. Matt Cutts has mentioned it during the release of Big [...]]]></description>
			<content:encoded><![CDATA[<p>Warning: I&#8217;m just rambling to get a free link on Google&#8217;s blog. Why do I want that? Cuz I&#8217;m bored? :)</p>
<p>There&#8217;s not much new in Google&#8217;s recent post, <a href="http://googlewebmastercentral.blogspot.com/2006/12/building-link-based-popularity.html">Building Link-Based Popularity</a>, in which Stephanie Ulrike underscores Google&#8217;s increasing ability to find and devalue artificial links. Matt Cutts has mentioned it during the release of Big Daddy earlier this year. In my sector, I still see reciprocal links remain effective to a large extent. </p>
<p>Honestly, I think it&#8217;ll take a while before Google reaches a point where only organic inbounds are counted toward a site&#8217;s popularity. Why? Guessing the intent of a link isn&#8217;t always easy. For example, I linked to a blog I read every day, just because it made it easier for me to visit the blog if I have the link on my blog&#8217;s front page. The blogger noticed my link and linked back to my blog. To Google that may look like reciprocal linking, but if Google discounts the links as such, its clearly missing the intention of those links.</p>
<p>There are also instances where an artifical link can made to look like citations. For example, write a fake article and bury a paid link in one of the paragraphs. How would Google figure out that link was paid for? I doubt it can. What about this? Instead of paying for a link, you pay for an &#8220;organic link generating service&#8221;, where a company will use several people to generate diggs for your site or bookmark you via del.icio.us. You&#8217;re still paying for links, except the company you&#8217;re paying doesn&#8217;t directly link to you anymore.</p>
<p>I don&#8217;t know. I still think there are alot of loopholes out there. Anyway, I think I&#8217;ve rambled on long enough.</p>
<p><strong>P.S.</strong> My link popped up on their blog in like under a minute. That&#8217;s faster than I expected. I do see their blog home page is TBPR 0. Though the links aren&#8217;t nofollowed, I doubt I&#8217;m getting anything besides a traffic/anchor text benefit out of my link appearing on their page.</p>
<p>UPDATE: <a href="http://seo-theory.blogspot.com/2006/12/google-takes-on-rand-fishkin-and-link.html">Michael Martinez</a> wrote the most interesting response to Google&#8217;s announcement I read so far, though I&#8217;m not sure how he goes from the ability to link &#8220;spam&#8221; via Digg to Google VS Link Baiters:</p>
<blockquote><p>I don&#8217;t rely on social link spamming, either, though Rand has publicly admitted to seeding DIGG and other social linking sites with stories and links&#8230;.Call this the first round of &#8220;Google versus the Link Baiters&#8221;.</p></blockquote>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/12/16/still-plenty-of-loopholes-in-googles-paid-link-detection-algo.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Why Google Will Not Move Away From PageRank</title>
		<link>http://seo4fun.com/blog/2006/12/13/why-google-will-not-move-away-from-pagerank.html</link>
		<comments>http://seo4fun.com/blog/2006/12/13/why-google-will-not-move-away-from-pagerank.html#comments</comments>
		<pubDate>Wed, 13 Dec 2006 10:18:10 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/12/13/why-google-will-not-move-away-from-pagerank.html</guid>
		<description><![CDATA[Recently, I&#8217;ve got a little flak in Google Group Webmaster Help for coming down hard on people in the &#8220;Google is broke&#8221; camp. Basically, some of them were upset that the supplemental index was based heavily on PageRank because Google&#8217;s PageRank paradigm is broken and unfair:

Google may misread the intent of a link. For example, [...]]]></description>
			<content:encoded><![CDATA[<p>Recently, I&#8217;ve got a little flak in Google Group Webmaster Help for coming down hard on people in the &#8220;Google is broke&#8221; camp. Basically, some of them were upset that the supplemental index was based heavily on PageRank because Google&#8217;s PageRank paradigm is broken and unfair:</p>
<ul>
<li><strong>Google may misread the intent of a link.</strong> For example, due partly to many people linking to domain.com, it&#8217;s amassed a TBPR 8. While those links aren&#8217;t meant to be citations, Google apparently counts them as such.</li>
<li><strong>A site can&#8217;t get organic links unless it already has links</strong> (a.k.a Mike Grehan&#8217;s &#8220;rich get richer&#8221; syndrome). This would be true IF Google is the only source of a site&#8217;s visibility. However, we&#8217;ve got Technorati, RSS, Yahoo, MSN, Reddit, Digg, Myspace, YouTube, paid advertising&#8230;True - in some niches (e.g. porn), Reddit or Digg isn&#8217;t going to work, and people are more hesitant to link to you. But in general, although the &#8220;richer get richer&#8221; is a fact of life, there are many ways around it. Like Adam Lasnik pointed out, once upon a time, YouTube.com was TBPR 0.</li>
<li><strong>The guy with the deepest pocket wins.</strong> The guy who can afford to spend the most money on paid advertising and paid links will in the end come up on top. Thus mom and pop sites will never have a chance in Google Search, or so they argue.</li>
<li><strong>The Paris Hilton syndrome.</strong> PageRank paradigm degrades search into a popularity contest. People link to what&#8217;s popular, even if  it has no value or it&#8217;s completely untrue. </li>
</ul>
<p>Basing a page&#8217;s value on links thus have several potential downsides (like anything else). Instead of PageRank, they argue, indexing should be based the quality of on-page text - a statement that begs the question: &#8220;Are you on crack?&#8221;</p>
<p>Answer me this. How can a computer program read, understand, and judge the quality of an article in comparison to other articles written on the same topic? It can&#8217;t - until Google discovers Artificial Intelligence. Sure - there are ways to look for on-page spammy finger prints (e.g. illogial sentence structures, excessively high keyword density, overuse of bold and italics). But given two well-written articles, how does a machine decide - based solely on on-page text - which article is more valuable?</p>
<p>It can&#8217;t.</p>
<p>Relevance for a keyword can, of course, be guessed at by looking at things like the TITLE tag, keyword frequency, keyword location on the page, and keywords in H1. Relevancy, however, has nada to do with page value or page quality.</p>
<p>How can a program judge the value of a page using on-page text alone when, from its POV, everything looks like a random string of symbols? To gauge a page&#8217;s value, there is simply no other option than to analyze off-page factors.</p>
<p>Ok, so it sucks that without inbound links or without decent internal link structure, Google will chuck a potentially great site into the supplemental index. Been there, done that. But if you think Google should base their indexing on on-page text quality instead of inbound links, I suggest you try spending a few days coding your own search engine. Then you&#8217;ll eventually realize what you want Google to do, at the present state of technology, is like wanting Wordpress to write posts for you, or like wanting your wife to become a rock legend overnight, or like wanting a billion-dollar <em>white-hat</em> website that builds and markets itself (and all you have to do is deposit checks in the bank every month).</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/12/13/why-google-will-not-move-away-from-pagerank.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>A Quick Update: AJAX/PHP/MYSQL and Google Stealin&#8217;</title>
		<link>http://seo4fun.com/blog/2006/12/13/a-quick-update-ajaxphpmysql-and-google-stealin.html</link>
		<comments>http://seo4fun.com/blog/2006/12/13/a-quick-update-ajaxphpmysql-and-google-stealin.html#comments</comments>
		<pubDate>Wed, 13 Dec 2006 08:42:41 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/12/13/a-quick-update-ajaxphpmysql-and-google-stealin.html</guid>
		<description><![CDATA[Just to let you know if you&#8217;re reading this blog, I&#8217;m working on a non-SEO-related script right now and I&#8217;m won&#8217;t be really posting for another week or two (Stop reading here if you&#8217;re not a Javascropt geek).
What kind of script is it? Well, like other scripts I wrote, its mainly for my own use, [...]]]></description>
			<content:encoded><![CDATA[<p>Just to let you know if you&#8217;re reading this blog, I&#8217;m working on a non-SEO-related script right now and I&#8217;m won&#8217;t be really posting for another week or two (Stop reading here if you&#8217;re not a Javascropt geek).</p>
<p>What kind of script is it? Well, like other scripts I wrote, its mainly for my own use, though it would be cool if I could somehow monetize it, which is one reason I&#8217;m gonna be pretty tight-lipped about it (at least till its done), but I can say with absolute certainty in its present state it already beats the pants off all apps Google came out with this year (and I&#8217;m not just saying that because I wrote it).</p>
<p>I&#8217;ll also add, in case you&#8217;re into Javascript, that till this year I wrote everything in PHP/MYSQL, and never even realized how limiting that 100% server-side model was. This should come as no surprise to anyone, but if you&#8217;re doing a huge amount of manual data entry (in cases where you hit a wall in a mass automation approach), you want as few keystrokes, mouse clicks, page scrolls, and page reloads as possible to get whatever you want done. Less input means less time wasted. If I was just designing a blog, I wouldn&#8217;t really care. But when I need to save thousands of work hours, good UI design can&#8217;t be overlooked. Client-side funtionality provided by AJAX and Javascript makes it all possible.</p>
<p>A few other things I&#8217;ve come to appreciate while hacking away for a few days is coding event handlers in Javascript instead of using Onclick in HTML (ya know, the CSS/HTML/Javascript = Design/Content/functionality model). I still prefer OnClick in situation where I generate the HTML via Javascript, but if I do View Source, it feels nice when all I see is HTML.</p>
<p>And yeah, I&#8217;ve been reading about the Yahoo IE7 VS Google IE7 non-sense but it seriously doesn&#8217;t interest me. Am I going to get richer by knowing about that design theft or is it supposed to be the spark that lights the &#8220;Google is Evil&#8221; firestorm that brings way too many &#8220;Google fucked up my site, I tried everything, I give up, I hate Google&#8221; whiners out of the woodwork? Just fire whoever is responsible (or give him/her a good spanking), and put some kind of approval process in place so things like this done by (I assume) a Googler newbie doesn&#8217;t go unchecked. I&#8217;m still waiting for an admission from Google that someone did in fact steal Yahoo&#8217;s design (as if removing the page wasn&#8217;t good enough - I guess it is - but I&#8217;d still like to hear it) and I&#8217;m also curious as to the reason why they decided to steal it, but like I said, talking about it isn&#8217;t going to make me any richer.</p>
<p>P.S. The other day, I noticed I was No. 1 on Google for &#8220;Pagerank doesn&#8217;t matter.&#8221; (Hehe) I know it&#8217;s a low-traffic, non-competitive term and I&#8217;ll lose position as I post more stuff on here, but I did take a screenshot just to remind myself I was there (too lazy at 3:33 in the morning to post it, but I got it). I keep imagining PageRank haters typing in &#8220;pagerank doesn&#8217;t matter&#8221; in Google and reading my post that proves they were wrong for way too long. Of course, I also noticed Andy Hagans posted a similar post way back in 2005, so I&#8217;m no SEO pioneer on that, but still. Too many people either say &#8220;it doesn&#8217;t matter&#8221; or obsess about it. The healthy POV is to look at it like your TITLE tag.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/12/13/a-quick-update-ajaxphpmysql-and-google-stealin.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>&#8220;PageRank Doesn&#8217;t Matter&#8221; is Now Officially an SEO Myth</title>
		<link>http://seo4fun.com/blog/2006/11/24/pagerank-doesnt-matter-is-now-officially-an-seo-myth.html</link>
		<comments>http://seo4fun.com/blog/2006/11/24/pagerank-doesnt-matter-is-now-officially-an-seo-myth.html#comments</comments>
		<pubDate>Fri, 24 Nov 2006 16:51:32 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/11/24/pagerank-doesnt-matter-is-now-officially-an-seo-myth.html</guid>
		<description><![CDATA[Back in August, after having fixed my 99% supplemental site to the point where duplicate content could not possibly be an issue, Google nonchalantly stuck my original pages back in the supplemental index. At that point, I claimed the existence of a PageRank hurdle preventing supplemental pages from getting back into the main index:
On gfe-eh.google.com [...]]]></description>
			<content:encoded><![CDATA[<p>Back in August, after having fixed my 99% supplemental site to the point where duplicate content could not possibly be an issue, Google nonchalantly stuck my original pages back in the supplemental index. At that point, I claimed <a href="/blog/2006/08/17/supplemental-index-fuzzier-than-ever.html">the existence of a PageRank hurdle preventing supplemental pages from getting back into the main index</a>:</p>
<blockquote><p>On gfe-eh.google.com and other DCs, the supplemental pages cache dates no longer go all the way back to Aug 2005. But the new “system” makes it even harder to tell why a page is listed in the supplemental index, because now you’re required to jump over at least two major hurdles to break out of supplemental hell: 1) duplicate content issues (i.e. identical meta tags, multiple urls resolving to the same content, www/non-www, etc) and 2) “Trust” / PageRank. A perfectly structured page with original content could remain stuck in the supplemental index if a domain lacks juice.</p></blockquote>
<p>On October 11, 2006, <a href="/blog/2006/10/12/matt-cutts-pagerank-primary-factor-determining-supplemental-results.html">Matt Cutts confirmed that PageRank is the primary factor used to determine whether or not to chuck pages into supplemental hell</a>. He later reiterated his point at Pubcon, and Adam Lasnik also recently echoed the same idea in Google Group Webmaster Help.</p>
<p>That PageRank still matters in this day and age didn&#8217;t fly over too well with some SEOs, who often de-emphasize PageRank in favor of other factors like &#8220;trust&#8221;, domain age, user data, links from authority sites, link neighborhood, co-citation, link history, link age, webmaster profile, SERP CTR, dmoz listing, and link topical relevance. Some black hats, G-Man for example, claimed supplemental results are mainly due to duplicate content issues. <a href="http://www.seomoz.org/blogdetail.php?ID=1516">Rand Fishkin also claimed</a> that PageRank has a limited role in determining a page&#8217;s supplemental status. He reasoned orphaned pages with no links to them may turn supplemental, but that in most cases, duplicate content is the predominant factor.</p>
<p>This anti-PageRank attitude stems from several commonly-held beliefs: that 1) PageRank has been devalued during the past few years; 2) PageRank by itself will not guarantee high rankings, and 3) PageRank is just one of hundreds of factors Google considers when evaluating ranking for a search term. Some webmasters also obsess too much over PageRank, paying hundreds of dollars to buy links on high TBPR pages in hopes of boosting their own TBPR.</p>
<p>Still, PageRank isn&#8217;t quite dead. Google still cares about the quantity and quality of links pointing to a page.  How does Google keep track of that? You guessed it.</p>
<p>The supplemental index is Google&#8217;s version of the Interweb junkyard. It&#8217;s the holding space for what got cut on Google&#8217;s cutting floor. It&#8217;s where your pages end up if Google thinks they lack value. Again, how does Google measure page value? As far as dupe issues are concerned, it uses on-page text. Otherwise, Google looks at links. The primary metric Google uses to size up links pointing to a page, of course, is PageRank.</p>
<p>That doesn&#8217;t mean low PageRank is the only reason a page is in the supplemental index. For example, Adam Lasnik&#8217;s following comment implies duplicate content is clearly another major factor:</p>
<blockquote><p> Joe Parts, I took a look at the examples you gave (thanks!) and &#8212; aside from the PR Toolbar issue I noted above &#8212; I did notice that at least a few of the URLs you noted come up directly in searches for text on the page.  But there&#8217;s not much text on those pages for us to go by, unfortunately, and so it&#8217;s not surprising that some seem to be perceived as similar content to other pages on the net.</p></blockquote>
<p>So its not just about low PageRank, even if that&#8217;s the main reason pages &#8220;go&#8221; supplemental. In Joe&#8217;s case, several other reasons, a few of which Adam alludes to, include 1) Similar shingles across pages 2) Devalued inbound links 3) Outdated TBPR.</p>
<p>&#8220;PageRank (toolbar PR) doesn&#8217;t matter (much anymore (in ranking))&#8221; is now officially an SEO myth and a misleading statement at best. Links matter and have always mattered. PageRank is just a metric for links. It used to be an inaccurate and spam-prone metric, but by nuking sold links, devaluing link schemes, blocking PageRank to spam sites with nofollow, and updating PageRank on a daily basis, PageRank has become a more reliable metric that better reflects page value.</p>
<p>The fact that Google already uses PageRank to determine crawl frequency, crawl depth, and supplemental indexing shows how confident Googlers feel (rightly or wrongly) about the accuracy of PageRank in place today.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/11/24/pagerank-doesnt-matter-is-now-officially-an-seo-myth.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>NSFW Reddit is No More</title>
		<link>http://seo4fun.com/blog/2006/11/13/nsfw-reddit-is-no-more.html</link>
		<comments>http://seo4fun.com/blog/2006/11/13/nsfw-reddit-is-no-more.html#comments</comments>
		<pubDate>Mon, 13 Nov 2006 23:38:36 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/11/13/nsfw-reddit-is-no-more.html</guid>
		<description><![CDATA[Just a few minutes ago, I noticed reddit dropped nsfw.reddit.com. If you haven&#8217;t already, you can read a loooong thread about it posted 13 days ago here. Why did they kill NSFW Reddit?

Lack of community compared to regular reddit.
It was never intended for porn: &#8220;Our intention for nsfw.reddit was for adolescent humor that, when on [...]]]></description>
			<content:encoded><![CDATA[<p>Just a few minutes ago, I noticed reddit dropped nsfw.reddit.com. If you haven&#8217;t already, you can read a loooong thread about it posted 13 days ago <a href="http://reddit.com/info/olu8/comments/coly3" rel="nofollow">here</a>. Why did they kill NSFW Reddit?</p>
<ul>
<li>Lack of community compared to regular reddit.</li>
<li>It was never intended for porn: &#8220;Our intention for nsfw.reddit was for adolescent humor that, when on reddit itself, was causing users trouble at work. Unfortunately, it quickly degenerated into porn.&#8221;</li>
<li>NSFW content scares away advertisers: &#8220;Unfortunately, having NSFW material available on a site means that it gets blocked by many filtering programs/gateways.&#8221;</li>
<li>&#8220;unrestrained fake voting from sockpuppets&#8221; - Probably because most NSFW surfers&#8217; hands are too tied up to cast a vote, so posters vote for themselves to fatten their bottom line.</li>
</ul>
<p>What a fine example of users steering a Web 2.0 site in a completely wrong direction. Anyway, if you are dying for a nsfw.reddit replacement, you might want to check out <a href="http://www.mosexindex.com/" rel="nofollow">MoSexIndex</a>, which btw, I&#8217;m not affiliated with in any way whatsoever (thus, the nofollow).</p>
<p>Meanwhile, <a href="http://groups.google.com/group/Google_Webmaster_Help-Indexing/browse_thread/thread/a4a866327b1172a1">Adam Lasnik is doing some detective work right now</a> trying to nail down the <a href="/blog/2006/10/17/wackiest-google-site-command-bug-ive-ever-seen.html">weird site: bug</a> I posted a while ago.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/11/13/nsfw-reddit-is-no-more.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Adam Lasnik Posting More Often?</title>
		<link>http://seo4fun.com/blog/2006/11/10/adam-lasnik-posting-more-often.html</link>
		<comments>http://seo4fun.com/blog/2006/11/10/adam-lasnik-posting-more-often.html#comments</comments>
		<pubDate>Fri, 10 Nov 2006 09:50:55 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/11/10/adam-lasnik-posting-more-often.html</guid>
		<description><![CDATA[I&#8217;ve been taking a break from SEO for the last two weeks, instead spending more time working on my own sites, writing up project plans for the next few months, working on a possibly groundbreaking SEO script, and doing a few SEO consulting on the side.
Though tonight, I couldn&#8217;t sleep, so here I am at [...]]]></description>
			<content:encoded><![CDATA[<p>I&#8217;ve been taking a break from SEO for the last two weeks, instead spending more time working on my own sites, writing up project plans for the next few months, working on a possibly groundbreaking SEO script, and doing a few SEO consulting on the side.</p>
<p>Though tonight, I couldn&#8217;t sleep, so here I am at 4:20 in the morning reading through <a href="http://groups.google.com/group/Google_Webmaster_Help">Google Group Webmaster Help</a>, when I noticed Adam Lasnik has been posting a hell of alot more in there recently. He doesn&#8217;t give away any hard kept secrets, mind you, but he does shell out some incredibly helpful feedback, enough not to make webmasters who post in that group to feel like they&#8217;re talking to a wall, like I sometimes feel when I chat with Google Adword reviewers, who I think are completely braindead and can&#8217;t tell right from wrong (/rant).</p>
<p>Anyway, if you&#8217;re interested, you can look for some of his comments by following the recent post links in <a href="http://groups.google.com/groups/profile?enc_user=Cc3iUTUAAAC0ZCEBAysSlShC_gPAdXUZ6uYgiBXIh5DaeqWCtoTJbOW2QCIjchDn6OYwMZe7JK3IgJw9QHTrXt9e__Js8L8H">his profile</a>. Don&#8217;t expect any new posting for a couple of weeks though, since pubcon is coming up.</p>
<p>UPDATE: If you wanna keep track of Vanessa Fox&#8217;s comments on Webmaster Help GG, <a href="http://groups.google.com/groups/profile?enc_user=a4E6IzgAAAC0ZCEBAysSlShC_gPAdXUZNvh6ZzwqbrDNmlK6sEGnNLvDmyeMYWPeTK9k6LE--aSJvIqbWFYuzPTfMg11G_KA" rel="nofollow">bookmark this link</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/11/10/adam-lasnik-posting-more-often.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Hack: Installing Rel=Nofollow in PHPAdsNew</title>
		<link>http://seo4fun.com/blog/2006/11/07/hack-installing-relnofollow-in-phpadsnew.html</link>
		<comments>http://seo4fun.com/blog/2006/11/07/hack-installing-relnofollow-in-phpadsnew.html#comments</comments>
		<pubDate>Tue, 07 Nov 2006 16:36:47 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/11/07/hack-installing-relnofollow-in-phpadsnew.html</guid>
		<description><![CDATA[If you use PHPAdsNew to serve ads on a relatively low FBPR (foolbarPR) website and your site is suffering from supplemental problems, installing rel=nofollow on your ad links is a tiny tweak that will help tighten your internal links and get some pages back into Google&#8217;s main index:
around line 366 in libraries/lib-view-main.inc.php, just before these [...]]]></description>
			<content:encoded><![CDATA[<p>If you use PHPAdsNew to serve ads on a relatively low FBPR (foolbarPR) website and your site is suffering from supplemental problems, installing rel=nofollow on your ad links is a tiny tweak that will help tighten your internal links and get some pages back into Google&#8217;s main index:</p>
<p>around line 366 in libraries/lib-view-main.inc.php, just before these lines:</p>
<blockquote><p>
// Return banner<br />
	return( array(&#8217;html&#8217; => $outputbuffer, </p></blockquote>
<p>Type this:</p>
<blockquote><p>
$outputbuffer = preg_replace(&#8221;/&lt;a href/i&#8221;, &#8220;&lt;a rel=&#8217;nofollow&#8217; href&#8221;, $outputbuffer);</p></blockquote>
<p>I&#8217;ve only tested this on HTML banners (a combination of image and text), so your mileage may vary.</p>
<p>Although some SEOs disregard the <a href="/blog/2006/10/12/matt-cutts-pagerank-primary-factor-determining-supplemental-results.html">role of PageRank in combating supplemental pages</a>, its something you do not want to rule out when looking for solutions.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/11/07/hack-installing-relnofollow-in-phpadsnew.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>SEO Surgery: resistinc.org - Banned Without a Clue</title>
		<link>http://seo4fun.com/blog/2006/10/25/196.html</link>
		<comments>http://seo4fun.com/blog/2006/10/25/196.html#comments</comments>
		<pubDate>Wed, 25 Oct 2006 15:39:32 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/10/25/196.html</guid>
		<description><![CDATA[This morning, I saw a guy on Google Webmaster Help asking why his domain is banned. Sure enough, site: query returns no results. According to domain tools, resistinc.org was created back in 1997 with 2 DMOZ listings and 3 Y! Directory listings. A site: command on MSN, though, turned up this cache page (dated 10/20/2006) [...]]]></description>
			<content:encoded><![CDATA[<p>This morning, I saw a guy on Google Webmaster Help asking <a href="http://groups.google.com/group/Google_Webmaster_Help-Indexing/browse_thread/thread/44e93ba18b9e6308" title="why his domain is banned">why his domain is banned</a>. Sure enough, site: query returns no results. According to domain tools, resistinc.org was created back in 1997 with 2 DMOZ listings and 3 Y! Directory listings. A site: command on MSN, though, turned up <a href="http://cc.msnscache.com/cache.aspx?q=4168773605629&#038;lang=en-US&#038;mkt=en-US&#038;FORM=CVRE" title="this cache page">this cache page</a> (dated 10/20/2006) with a javascript on the bottom of the source code:</p>
<blockquote><p>&lt;script&gt;document.write(&#8217;&lt;ma&#8217; + &#8216;rq&#8217; + &#8216;uee wid&#8217; + &#8216;th=1 height=1&gt;&lt;fo&#8217; + &#8216;nt style=&#8221;col&#8217; + &#8216;or:#000&#8242; + &#8216;000;<strong>fon&#8217; + &#8216;t-size:1px</strong>;&#8221;&gt;&#8217;)&lt;/script&gt;</p></blockquote>
<p>What follows are hidden links to stuff like &#8220;brewster ny honda&#8221;, &#8220;drawn horse&#8221;, &#8220;yahoo game back door,&#8221; &#8220;hip hop cartel&#8221;, and &#8220;bronx bankruptcy lawyer&#8221;. Google sees these links because only the marquee element is in Javascript. Also keep in mind DMOZ.org&#8217;s description of resistinc.org has nothing to do as far as I know about hip hop cartels:</p>
<blockquote><p>
&#8220;For more than 30 years, Resist has funded progressive organizations in the United States that are actively part of a movement for social change.&#8221;</p></blockquote>
<p>Though the link to &#8220;hip hop cartel&#8221; seems to 404 now, <a title="MSN site: search" href="http://search.msn.com/results.aspx?q=site%3Ahttp%3A%2F%2Fwww.resistinc.org%2F+&#038;FORM=MSNH">MSN site: search</a>  reveals some other cliche urls (out of 7,172 urls):</p>
<blockquote><p>www.resistinc.org/free-online-poker.dhtml<br />
www.resistinc.org/auto-insurance.dhtml<br />
www.resistinc.org/stud-poker-tracker.dhtml<br />
www.resistinc.org/buy&#8212;viagra.dhtml<br />
www.resistinc.org/cheap-landlord-insurance.dhtml</p></blockquote>
<p>List goes on.</p>
<p>Were they hacked? Who knows. What I find puzzling is what this guy posted (dated Tues, Oct 24 2006 11:12 am - Just a day ago):</p>
<blockquote><p>Hi:</p>
<p>I&#8217;m a volunteer for:</p>
<p><a href="http://www.resistinc.org/" rel="nofollow">http://www.resistinc.org/</a></p>
<p><span style="font-weight: bold;">We think we&#8217;ve been banned</span>, since we get no results for searches with site:resistinc.org </p>
<p><span style="font-weight: bold;">We have reviewed the webmaster guidelines and cannot think of how we might have vioated any of them.</span> We&#8217;d appreciate any feedback or guideance about how to determine what might be wrong and/or how to proceed.</p></blockquote>
<p>The last advice on that Google Help thread is &#8220;install Sitemaps so you can get some Google feedback.&#8221; If he realized he&#8217;d been hacked, I&#8217;d expect him to post &#8220;we were hacked and now we&#8217;re banned.&#8221; The spammy pages (the ones I checked) 404, so someone must have removed them. I don&#8217;t know, it&#8217;s smells kinda fishy to me (not that I don&#8217;t like seafood).</p>
<p><strong>P.S.</strong> I tried publishing this via Google Docs, and somehow the TITLE of the post turned up blank. Of course &lt; and &gt; didn&#8217;t translate. Google also uses BR instead of P so I get a chunk of code in Wordpress with no paragraph breaks. Sorry to those who read me on Bloglines (since I keep republishing this one after mutli-edits). My advice? Switch to Google Reader.</p>
<p><strong>UPDATE:</strong> As you might have noticed, the original poster later added:</p>
<blockquote><p>While looking into this yesterday afternoon, I discovered some PHP files on the site that appear to have been placed there illegitimately.</p>
<p>I removed those, but it&#8217;s still too early to say that the site has been &#8220;cleaned up.&#8221; I am working on that.</p>
<p>To state what I hope is obvious: neither the owners of this site nor I had knowledge of the fact that the website had been hijacked into serving these advertisements.</p>
<p>In defense of the organization that owns this site, I will say that like many small nonprofits, they cannot afford in-house tech support. In my own defense, I will say that I am not in charge of the day-to-day operation of this website; I responded to a call for help.</p>
<p>We appreciate the information that has been provided, and will of course appreciate any further constructive suggestions about how to proceed.</p></blockquote>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/10/25/196.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Wackiest Google Site Command Bug I&#8217;ve Ever Seen</title>
		<link>http://seo4fun.com/blog/2006/10/17/wackiest-google-site-command-bug-ive-ever-seen.html</link>
		<comments>http://seo4fun.com/blog/2006/10/17/wackiest-google-site-command-bug-ive-ever-seen.html#comments</comments>
		<pubDate>Tue, 17 Oct 2006 09:27:10 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/10/17/wackiest-google-site-command-bug-ive-ever-seen.html</guid>
		<description><![CDATA[Google&#8217;s site: command&#8217;s been acting up lately, and Googlers are working feverishly day and night I assume trying to fix the damn thing on various Datacenters, but check this out. While running a site: command tonight, I came across this bizarre SERP:

Googe site search returns ginormous serp snippets
Here&#8217;s the search url I used (broken in [...]]]></description>
			<content:encoded><![CDATA[<p>Google&#8217;s site: command&#8217;s been acting up lately, and Googlers are working feverishly day and night I assume trying to fix the damn thing on various Datacenters, but check this out. While running a site: command tonight, I came across this bizarre SERP:</p>
<p><img src="/images/site-command.jpg" alt="google site command bugged" /></p>
<p><strong>Googe site search returns ginormous serp snippets</strong></p>
<p>Here&#8217;s the search url I used (broken in two parts, so it doesn&#8217;t break this blog&#8217;s template):</p>
<p>http://www.google.com/search?hl=en&#038;lr=&#038;safe=off&#038;client=firefox-a<br />
&#038;rls=org.mozilla%3Aen-US%3Aofficial&#038;q=site%3Agembaby.com&#038;btnG=Search</p>
<p>(EDIT: The search result is back to normal now &#8212; Too bad.)</p>
<p>Huge description snippets, right? Where is it coming from?</p>
<p>META description, repeated 12 times.</p>
<p>So, um..Google, what&#8217;s up with that?</p>
<p><strong>UPDATE:</strong> I removed snippets I previously quoted off the meta description and the SERP, since I didn&#8217;t want to accidentally create a duplicate content type scenario.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/10/17/wackiest-google-site-command-bug-ive-ever-seen.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Matt Cutts: PageRank Primary Factor Determining Supplemental Results</title>
		<link>http://seo4fun.com/blog/2006/10/12/matt-cutts-pagerank-primary-factor-determining-supplemental-results.html</link>
		<comments>http://seo4fun.com/blog/2006/10/12/matt-cutts-pagerank-primary-factor-determining-supplemental-results.html#comments</comments>
		<pubDate>Thu, 12 Oct 2006 13:25:06 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/10/12/matt-cutts-pagerank-primary-factor-determining-supplemental-results.html</guid>
		<description><![CDATA[I was going to write a long post about this Matt Cutts quote, but I think it speaks for itself.
PageRank is the primary factor determining whether a url is in the main web index vs. the supplemental results.
I predict many seasoned SEOs and newbies alike will dismiss his statement. Firstly, PageRank doesn&#8217;t matter, right? Or [...]]]></description>
			<content:encoded><![CDATA[<p>I was going to write a long post about this <a href="http://www.mattcutts.com/blog/fall-weather-forecast/#comment-87795">Matt Cutts quote</a>, but I think it speaks for itself.</p>
<blockquote><p><strong>PageRank is the primary factor</strong> determining whether a url is in the main web index vs. the supplemental results.</p></blockquote>
<p>I predict many seasoned SEOs and newbies alike will dismiss his statement. Firstly, PageRank doesn&#8217;t matter, right? Or were you talking about TBPR? But as Matt Cutts made it clear, TBPR is exported internal PageRank translated on a 0-10 scale. The only inaccuracy involved with TBPR is PageRank updates continuously while TBPR updates every few months, and internal PageRank is more granular than TBPR.</p>
<p>But seriously, how can PageRank be the primary factor of anything? Well, here&#8217;s a reality check for folks who likes to turn SEO into a fairy tale where you believe what you want to believe in the face of irrefutable facts.</p>
<p>Another quote I found interesting wriiten by Marcia on a recent featured homepage WMW thread about <a href="http://www.webmasterworld.com/google/3107196-10-10.htm">supplemental results and inbound links</a>:</p>
<blockquote><p>I&#8217;m absolutely in agreement with that and it&#8217;s stood the test of time - and IBLs &#038; PR. Nope, the toolbar ain&#8217;t dead yet; the reports of its demise are grossly exaggerated and contra-indicated by the Supp results and indicators.</p>
<p>Not one single speck of duplication, what&#8217;s Supplemental and what isn&#8217;t on the test site(s) is 100% dependent on the amount of link love the pages are getting.</p>
<p>People who are looking for dup issues where none exist are, unfortunately, chasing their tails. </p></blockquote>
<p>g1smd (whom I consider to be well versed on issues revolving around duplicate content and supplemental results) responds:</p>
<blockquote><p>Heh, Marcia, you&#8217;re gonna love this Matt Cutts comment:</p>
<p>>> PageRank is the primary factor determining&#8230;.</p>
<p>Note: The &#8220;primary&#8221; factor.</p>
<p>Jeez. No mention of Duplicate Content, and Redirects and 404 URLs at all.</p>
<p>Ah, but maybe he only means for &#8220;live&#8221; URLs, or maybe redirects and 404s no longer have any PageRank associated with them.</p>
<p>Whatever, it agrees with what you&#8217;re saying: and I guess that fuels another link frenzy to start all over again. </p></blockquote>
<p>Caveman follows up with an interesting comment as well, but I don&#8217;t want to turn this page into a duplicate content page by overquoting, ya know what I mean? :D</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/10/12/matt-cutts-pagerank-primary-factor-determining-supplemental-results.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Docs Gone Bad - Google, Fix the Tags</title>
		<link>http://seo4fun.com/blog/2006/10/12/google-docs-gone-bad.html</link>
		<comments>http://seo4fun.com/blog/2006/10/12/google-docs-gone-bad.html#comments</comments>
		<pubDate>Thu, 12 Oct 2006 08:46:27 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/10/12/google-docs-gone-bad.html</guid>
		<description><![CDATA[Google Docs tags aren&#8217;t working right now, which means all the new articles you write in Google Docs will end up in a big disorganized pile of whatev, and you get to go back later (when tags are working again) and clean up the mess. This may not be news to you if you never [...]]]></description>
			<content:encoded><![CDATA[<p>Google Docs tags aren&#8217;t working right now, which means all the new articles you write in Google Docs will end up in a big disorganized pile of whatev, and you get to go back later (when tags are working again) and clean up the mess. This may not be news to you if you never used Writely, but if you do I guarantee its going to annoy you to no end. So much for smooth transitions. I tell ya, if I was working for Google, this kind of thing would NEVER happen. I mean, how do they screw up something that was working perfectly just a few days ago?</p>
<p>Anyhoo, I hope Googlers are aware of this bug and taking it with the utmost seriousness, because without tags, Google Docs for you will be just Google Notebooks Deluxe, with no organizational functionality save sort by last mod timestamp, which forces you to scan through a long list of notebooks every time you want to write something down.</p>
<p>I posted about this over at <a href="http://groups.google.com/group/GoogleDocsSpreadsheets">Google Docs &#038; Spreadsheets Help Group</a> just to help get this issue see the light of day. It&#8217;s good to know I&#8217;m not the only guy having this problem.</p>
<h3>Google Docs Broken Tags Bug Google Groups Threads</h3>
<ul>
<li><a rel="nofollow" href="http://groups.google.com/group/Something-in-Writely-is-Broken/browse_thread/thread/9be308121250ad46">Tagging</a></li>
<li><a rel="nofollow" href="http://groups.google.com/group/Something-in-Writely-is-Broken/browse_thread/thread/7e0dbd45f83fe549">Untagging doesn&#8217;t work</a></li>
<li><a rel="nofollow" href="http://groups.google.com/group/Something-in-Writely-is-Broken/browse_thread/thread/00014b0267b1ea82">Adding a tag doesn&#8217;t work anymore</a></li>
<li><a rel="nofollow" href="http://groups.google.com/group/Something-in-Writely-is-Broken/browse_thread/thread/9cce5150734d5c49">Missing tags</a></li>
<li><a rel="nofollow" href="http://groups.google.com/group/Suggestions-and-Ideas-Writely/browse_thread/thread/de0516f42effedd1">Tags</a></li>
<li><a rel="nofollow" href="http://groups.google.com/group/Something-in-Writely-is-Broken/browse_thread/thread/f2f44df267bb1ef0">Unable to Tag Documents from Documents List</a></li>
<li><a rel="nofollow" href="http://groups.google.com/group/Something-in-Writely-is-Broken/browse_thread/thread/0f9b13fecad5fb1f">Don&#8217;t Work &#8220;Remove Tags&#8221;</a></li>
<li><a rel="nofollow" href="http://groups.google.com/group/Something-in-Writely-is-Broken/browse_thread/thread/3dbf1e398c961b24">Remove Tag</a></li>
<li><a rel="nofollow" href="http://groups.google.com/group/Something-in-Writely-is-Broken/browse_thread/thread/561925087e75b35a">Not Able to Tag do Documents</a></li>
<li><a rel="nofollow" href="http://groups.google.com/group/Something-in-Writely-is-Broken/browse_thread/thread/7838378a64951289">Documents: Can&#8217;t add tags</a></li>
</ul>
<p>You get the picture.</p>
<p><strong>UPDATE:</strong> It looks like Google finally fixed the bug. Let&#8217;s hope its a permanent fix.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/10/12/google-docs-gone-bad.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Lost and Grey&#8217;s Anatomy on My Google Calendar</title>
		<link>http://seo4fun.com/blog/2006/09/28/lost-and-greys-anatomy-on-my-google-calendar.html</link>
		<comments>http://seo4fun.com/blog/2006/09/28/lost-and-greys-anatomy-on-my-google-calendar.html#comments</comments>
		<pubDate>Thu, 28 Sep 2006 20:14:36 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/09/28/lost-and-greys-anatomy-on-my-google-calendar.html</guid>
		<description><![CDATA[We all read Google blog&#8217;s recent announcement about a new Google Calendar feature that lets you add web content events, like &#8220;weather forecasts, moon phases, and even Google doodles.&#8221; Well, to tell you the truth, I don&#8217;t care about all that, though holidays on my calendars I&#8217;ve already added. What I desperately need are TV [...]]]></description>
			<content:encoded><![CDATA[<p>We all read Google blog&#8217;s recent announcement about <a href="http://googleblog.blogspot.com/2006/09/google-calendar-does-something-about.html">a new Google Calendar feature</a> that lets you add web content events, like &#8220;weather forecasts, moon phases, and even Google doodles.&#8221; Well, to tell you the truth, I don&#8217;t care about all that, though holidays on my calendars I&#8217;ve already added. What I desperately need are TV programming schedules like Lost, Dancing with the Stars, Grey&#8217;s Anatomy, and Ugly Betty. Seriously, if I could have those things automatically show up on my calendar, it would save me lots of time (i think?). And nope, I don&#8217;t wanna be a developer.</p>
<p>UPDATE: Now I got Fall TV Season (2006) on my Calendar. Very distracting, but at least it lets me schedule my work around my favorite shows. BTW, since I have a tendency to rant about bugs, I&#8217;m posting a link to <a href="http://groups.google.com/group/Google-Calendar-Help/">Google Groups: Google Calendar Help</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/09/28/lost-and-greys-anatomy-on-my-google-calendar.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Rel NOFOLLOW = Granular META ROBOTS NOFOLLOW</title>
		<link>http://seo4fun.com/blog/2006/09/27/rel-nofollow-granular-meta-robots-nofollow.html</link>
		<comments>http://seo4fun.com/blog/2006/09/27/rel-nofollow-granular-meta-robots-nofollow.html#comments</comments>
		<pubDate>Wed, 27 Sep 2006 20:35:21 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/09/27/rel-nofollow-granular-meta-robots-nofollow.html</guid>
		<description><![CDATA[In a response to John Battelle&#8217;s recently released interview with Matt Cutts, Danni Sullivan asks Matt:
Matt, in the interview, you suggest that Google now views meta robots with a nofollow value as being the same as the completely different nofollow attribute, in terms of flagging links as not trusted. Is that now the case?
Matt&#8217;s response:
Danny, [...]]]></description>
			<content:encoded><![CDATA[<p>In a response to <a href="http://battellemedia.com/archives/002917.php">John Battelle&#8217;s recently released interview with Matt Cutts</a>, Danni Sullivan asks Matt:</p>
<blockquote><p>Matt, in the interview, you suggest that Google now views meta robots with a nofollow value as being the same as the completely different nofollow attribute, in terms of flagging links as not trusted. Is that now the case?</p></blockquote>
<p>Matt&#8217;s response:</p>
<blockquote><p>Danny, that’s always been the case–sorry if I haven’t explained that well. There are many techniques to sell visitors/traffic without selling PageRank or affecting search engines. As long as a link doesn’t affect search engines, there is no problem with selling that link from Google’s perspective. The nofollow attribute on links is the most granular because it’s on a link level, but something like a sponsor page is a fine opportunity to use the nofollow meta tag instead of marking each link.</p></blockquote>
<p>I asked (though I expect Matt to dodge):</p>
<blockquote><p>I think there’s still skepticism out there as to Adam’s recent statement “Google senses much” in terms of link selling. For example, was the wc3 page detected algorithmically or from word of mouth? Also, what would be the likely consequence if wc3 decided not to use the nofollow tag?</p></blockquote>
<p>P.S.</p>
<p>In the original interview, some guy asked:</p>
<blockquote><p>Matt, I have one line of questions. Google&#8217;s job is to measure the relevancy of web pages with algorithms/human computation or other means Google have or could have.</p>
<p>But is it fair for Google to put restriction on what people could do with web?</p></blockquote>
<p>I hear this alot, and for some reason, it always rubs me the wrong way.</p>
<p>Here&#8217;s a scenario where Google puts no restrictions on websites:</p>
<p><strong>SERP in random order.</strong></p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/09/27/rel-nofollow-granular-meta-robots-nofollow.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>SEO Surgery: Monx007</title>
		<link>http://seo4fun.com/blog/2006/09/22/seo-surgery-monx007.html</link>
		<comments>http://seo4fun.com/blog/2006/09/22/seo-surgery-monx007.html#comments</comments>
		<pubDate>Fri, 22 Sep 2006 13:10:27 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/09/22/seo-surgery-monx007.html</guid>
		<description><![CDATA[I&#8217;ve been spending the last couple weeks checking out posts in Google Webmasters Help. One thing that these posts keep hammering home to me is the fact that there&#8217;s a huge gap between what&#8217;s being posted on SEO forums and reality in cases where either people are not allowed to link drop or they choose [...]]]></description>
			<content:encoded><![CDATA[<p>I&#8217;ve been spending the last couple weeks checking out posts in <a href="http://groups.google.com/group/Google_Webmaster_Help">Google Webmasters Help</a>. One thing that these posts keep hammering home to me is the fact that there&#8217;s a huge gap between <a href="http://www.webmasterworld.com/google/3091380.htm">what&#8217;s being posted on SEO forums</a> and reality in cases where either people are not allowed to link drop or they choose to keep their URLs to themselves. And right now I&#8217;m getting a kick out of talking about real sites instead of theories. Even a simple question like &#8220;why is my site not indexed?&#8221; often lead to unexpected answers.</p>
<p>A random example (just fished this one out a few minutes ago):</p>
<p><em>&#8220;I&#8217;m very confuse now, before August, my site got PR 4. But around August, my site PR drop to 0.&#8221;</em></p>
<p>Without a URL, typical responses might be:</p>
<ul>
<li>Maybe you&#8217;re banned. Check to make sure your site is clean. You might want to do a re-inclusion request.</li>
<li>The little green bar doesn&#8217;t matter; don&#8217;t worry about it.</li>
<li>Check urls that are linking to you. Changes in their PageRank will effect yours.</li>
</ul>
<p>But the moment the guy posts his url, I have something I can sink my teeth into: <a href="http://games.monx007.com/" rel="nofollow">http://games.monx007.com/</a> What do you think?</p>
<p>Inbound links, according to Yahoo Site Explorer:</p>
<p><a href="http://technohack.blogspot.com/" rel="nofollow">http://technohack.blogspot.com/</a> TBPR 0<br />
http://www.klubforbiz.com/info-kerja-penawaran-kerjasama/t-offline-jobsvacany-at-telkom-963.html (forum - no connection) TBPR 0<br />
<a href="http://www.friendster.com/user.php?uid=7459848"  rel="nofollow">http://www.friendster.com/user.php?uid=7459848</a> TBPR 0<br />
<a href="http://www.womensshoescritic.com/links/games%5B99%5D-1.htm"  rel="nofollow">http://www.womensshoescritic.com/links/games%5B99%5D-1.htm</a> TBPR 0</p>
<p>Interesting. Hidden text using CSS fontsize=0em.</p>
<p><a href="http://www.ringcards.com/resources/recreation%5B8%5D-56.htm"  rel="nofollow">http://www.ringcards.com/resources/recreation%5B8%5D-56.htm</a> (directory) TBPR 0<br />
http://www.myrtle-beach-sc-vacation.com/metro/recreation%5B3%5D-12.htm (directory - now dead) TBPR 0<br />
<a href="http://www.cellphoneswow.com/links/recreation%5B8%5D-52.htm"  rel="nofollow">http://www.cellphoneswow.com/links/recreation%5B8%5D-52.htm</a> (another similar looking directory) TBPR 0<br />
<a href="http://archive.freespaces.com/"  rel="nofollow">http://archive.freespaces.com/</a> TBPR 0</p>
<p>What&#8217;s this? Links pointing to all his subdomains click-proofed with external JS? Real slick.</p>
<p><a href="http://www.lake-tahoe-local.com/partners/games%5B143%5D.htm"  rel="nofollow">http://www.lake-tahoe-local.com/partners/games%5B143%5D.htm</a> (games directory) TBPR 0<br />
<a href="http://www.shohouyaku.com/links/playstation%5B43%5D.htm"  rel="nofollow">http://www.shohouyaku.com/links/playstation%5B43%5D.htm</a> (another crap directory), TBPR 2<br />
<a href="http://www.yellowsurveys.com/resources/index176.htm"  rel="nofollow">http://www.yellowsurveys.com/resources/index176.htm</a> (crap directory) TBPR 0<br />
http://www.paid2lotto.com/links/recreation%5B8%5D-10.htmm (dead link) TBPR 0</p>
<p>A few things that stand out:</p>
<ul>
<li>Alot of TBPR 0s (doesn&#8217;t mean those pages were penalized though. I didn&#8217;t check their backlinks or see whether they are supplemental).</li>
<li>No &#8220;natural&#8221; citation links. Low quality links, including hidden cross links, forum sig, and cheap directory listings.</li>
</ul>
<p>Especially the cheap directories look too much alike when you compare their url structures:</p>
<p>resources/recreation%5B8%5D-56.htm<br />
metro/recreation%5B3%5D-12.htm<br />
links/recreation%5B8%5D-52.htm<br />
partners/games%5B143%5D.htm<br />
links/playstation%5B43%5D.htm<br />
links/recreation%5B8%5D-10.htm</p>
<p>If the guy didn&#8217;t supply the URL, we&#8217;d still be talking about a &#8220;sandbox&#8221; or &#8220;trustbox&#8221; or whatever-else-box. BTW, I have no moral qualms about &#8220;outing&#8221; a crap network or a lousy link building campaign. Robbers don&#8217;t out each other, but I&#8217;m no robber. As The Clash used to sing:</p>
<p>YOU BETTER CHEAT CHEAT<br />
NO REASON TO PLAY FAIR<br />
CHEAT CHEAT OR DON&#8217;T GET ANYWHERE<br />
CHEAT CHEAT <strong>IF YOU CAN&#8217;T WIN</strong></p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/09/22/seo-surgery-monx007.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>SEO Quotes of the Week September 6, 2006</title>
		<link>http://seo4fun.com/blog/2006/09/06/seo-quotes-of-the-week-september-6-2006.html</link>
		<comments>http://seo4fun.com/blog/2006/09/06/seo-quotes-of-the-week-september-6-2006.html#comments</comments>
		<pubDate>Wed, 06 Sep 2006 11:50:16 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/09/06/seo-quotes-of-the-week-september-6-2006.html</guid>
		<description><![CDATA[Some memorable quotes I came across so far this week  in SEO forums and the blogsphere. None of them may be revolutionary or news breaking, but they resonated with me and hopefully with you too.
the reason some did not believe it for a long time is that &#8212; as is true with so many [...]]]></description>
			<content:encoded><![CDATA[<p>Some memorable quotes I came across so far this week  in SEO forums and the blogsphere. None of them may be revolutionary or news breaking, but they resonated with me and hopefully with you too.</p>
<blockquote><p>the reason some did not believe it for a long time is that &#8212; as is true with so many of G&#8217;s algo elements &#8212; things are co-dependent. So the addition of x 1,000 pages for site A will not have the same effect as the addition of 1,000 pages for site B, and the difference is not just limted to the pre-existing number of pages on each site. Many other factors involved.</p></blockquote>
<p>- caveman, <a href="http://www.webmasterworld.com/google/3070792-6-10.htm">on the fallacy of ignoring the &#8220;co-dependency&#8221; of factors when determining the effect of one</a></p>
<blockquote><p>
 I think many people who listen to Matt are those who are trying to rank &#8220;legitimately&#8221; in Google. That is, they are mostly concerned about sending false positives, or getting caught accidentally in a Google net intended to catch spammers. This is one good reason to listen to what Matt says &#8212; there is usually enough meat there to at least help you steer clear of big problem areas, and sometimes problem areas that only recently rose in importance.</p>
<p>In this case, one comment that is worth paying some attention to, I think, is</p>
<blockquote><p>&#8220;So this is not something that a typical site owner needs to think about or worry about if they&#8217;re not adding hundreds of thousands or millions of URLs very quickly.&#8221;&#8211;Matt Cutts/<a href="http://www.seroundtable.com/archives/006065.html#comments">SER</a></p></blockquote>
<p>Here he gives some sense of the scale that would trip this particular flag &#8212; and it&#8217;s not in the thousands of urls, or even the tens of thousands. Certainly seasonal changes at completely legitimate ecommerce sites can require a couple thousand new urls at once. And Google would also have a historical record of that kind of seasonal change to help boost their confidence and trust&#8230;.</p></blockquote>
<p>- Tedster, <a href="http://www.webmasterworld.com/google/3070792-6-10.htm">Matt Cutts&#8217; Adding Too Many URLS Raise a Flag</a></p>
<blockquote><p>Probably just index churn. The Supplemental folks already had a fix for this ready, so it&#8217;s a matter of when some executables will be pushed. I&#8217;d expect highlighting to be working again within the next few weeks. Thanks again for mentioning this..</p></blockquote>
<p>- Matt Cutts, <a href="http://www.seroundtable.com/archives/006062.html">regarding text in cached supplemental pages not being highlighted</a></p>
<blockquote><p>Since I took Kimberly to task in a public blog I feel that the apology should come from the same place. </p>
<p>So, Kimberly and anyone else I offended with that particular post, I apologize.  I never intended any harm but I think I may have caused some.  I&#8217;ll be more careful in the future.</p></blockquote>
<p>- G-Man, <a href="http://www.seomoz.org/blogdetail.php?ID=1358">A Tip for your life&#8230;</a></p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/09/06/seo-quotes-of-the-week-september-6-2006.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Gotta Love The Web</title>
		<link>http://seo4fun.com/blog/2006/09/06/gotta-love-the-web.html</link>
		<comments>http://seo4fun.com/blog/2006/09/06/gotta-love-the-web.html#comments</comments>
		<pubDate>Wed, 06 Sep 2006 11:12:07 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/09/06/gotta-love-the-web.html</guid>
		<description><![CDATA[Hey, I just made a brand new tape. It could definitely be my stairway to riches and fame. And for your information, I hit the low notes better than Danny. I kid you not. So if I go head to head against him on American Idol, look out!
You betta lose yourself in my music the [...]]]></description>
			<content:encoded><![CDATA[<p>Hey, I just made a brand new tape. It could definitely be my stairway to riches and fame. And for your information, I hit the low notes better than Danny. I kid you not. So if I go head to head against him on American Idol, <em>look out!</em></p>
<p>You betta lose yourself in my music the moment you own it<br />
you better never let it go<br />
You only get one shot, do not miss your chance to blow<br />
this opportunity comes once in a lifetime</p>
<p><img src="/images/why-supplemental.jpg" alt="Google Supplemental Index" /></p>
<p><strong>UPDATE:</strong> Yeah, going through my old feeds, I just noticed <a href="http://www.jimboykin.com/need-an-official-seal-seo-code-of-ethics/">Jim already blogged about this</a> 2 days ago. God dangit&#8230;</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/09/06/gotta-love-the-web.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Supplemental Cache Refresh Revisited</title>
		<link>http://seo4fun.com/blog/2006/09/05/supplemental-cache-refresh-revisited.html</link>
		<comments>http://seo4fun.com/blog/2006/09/05/supplemental-cache-refresh-revisited.html#comments</comments>
		<pubDate>Tue, 05 Sep 2006 17:30:07 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/09/05/supplemental-cache-refresh-revisited.html</guid>
		<description><![CDATA[After the recent supplemental cache refresh, more than a few webmasters asked: &#8220;why are my perfectly structured, original content pages still supplemental?&#8221; and &#8220;Do I have to wait another 12 months for Google to get it right?&#8221;
We fixed our META description tags. We got rid of thin content pages. We installed 301 redirects and even [...]]]></description>
			<content:encoded><![CDATA[<p>After the recent supplemental cache refresh, more than a few webmasters asked: &#8220;why are my perfectly structured, original content pages still supplemental?&#8221; and &#8220;Do I have to wait another 12 months for Google to get it right?&#8221;</p>
<p>We fixed our META description tags. We got rid of thin content pages. We installed 301 redirects and even validated our pages to death. But the pages with fresh cache were still showing up as supplemental for no apparent reason.</p>
<p>What I assumed was that supplemental pages show up in the SERP <em>after</em> Google looked them over for quality, duplicates, trust, etc. Nope. Things happened the other way around:</p>
<p>1. Google recrawled all existing supplemental pages by looking up urls stored in its supplemental database. One clean sweep. The timestamps were updated and the cache contained the most recent version of each page. Basically, a database update &#8212; with zero movement of pages from the supplemental index into the main index.</p>
<p>2. After Google updated its supplemental database, it began a long process of evaluating each URL to decide if its spam, duplicate content, or A1. This is still going on.</p>
<p>I hope you&#8217;re starting to see more pages reappear in the main index.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/09/05/supplemental-cache-refresh-revisited.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Engineer Toys with 64.233.187.whatever</title>
		<link>http://seo4fun.com/blog/2006/08/27/google-engineer-toys-with-64233187whatever.html</link>
		<comments>http://seo4fun.com/blog/2006/08/27/google-engineer-toys-with-64233187whatever.html#comments</comments>
		<pubDate>Mon, 28 Aug 2006 04:42:24 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/08/27/google-engineer-toys-with-64233187whatever.html</guid>
		<description><![CDATA[At the risk of sounding like another SEW echo chamber post, here&#8217;s a recent Googleguy quote (now a random Googler dude, or is it Matt Cutts pretending to be a random Googler dude? You be the judge) regarding recent reports of backlink updates on several DCs:

It&#8217;s Brett&#8217;s decision to call something an update, but I [...]]]></description>
			<content:encoded><![CDATA[<p>At the risk of sounding like another SEW echo chamber post, here&#8217;s a <a href="http://www.webmasterworld.com/google/3058623-6-10.htm#msg3061576">recent Googleguy quote</a> (now a random Googler dude, or is it Matt Cutts pretending to be a random Googler dude? You be the judge) regarding recent reports of backlink updates on several DCs:</p>
<blockquote><p>
It&#8217;s Brett&#8217;s decision to call something an update, but I agree this one isn&#8217;t anything to write home about. Like Pluto, I think this would be a shrinking update; the SERPs aren&#8217;t really changing. :)</p>
<p>I talked to one of the engineers who would know, and it turns out that it&#8217;s an engineer who has grabbed the 64.233.187.whatever datacenter for himself to tinker around with making info: queries slightly more accurate. I don&#8217;t expect the visible/external info: or backlink data to spread to other data centers (or if it did, not for a long time).</p>
<p>So have fun poking around over there if you really enjoy monitoring IP addresses, but just bear in mind that it&#8217;s an engineer tinkering by himself, not part of a larger trend or an upcoming update. </p></blockquote>
<p>Earlier, <a href="http://www.webmasterworld.com/google/3058623-5-10.htm#msg3061445">he says</a>:</p>
<blockquote><p>
 I&#8217;ll ask around with a couple engineeers who keep a closer eye on external backlinks/PR and see if they have anything they&#8217;d like to mention.</p>
<p>Bear in mind that backlinks/PageRank updates happen at their own rate, and we&#8217;re continuously finding and incorporating new backlinks and computing new PageRank all the time. So an update of visible backlinks doesn&#8217;t really cause an &#8220;update&#8221; or big change in rankings, because we&#8217;ve already known about those links for a while.</p>
<p>Anyway, I&#8217;ll ask around and if anything interesting comes out of it, I&#8217;ll let you know. </p></blockquote>
<p>BTW, if you&#8217;re still suffering from supplemental problems, there&#8217;s also a <a href="http://www.webmasterworld.com/google/3060898.htm">thread developing over at WMW started by Whitey</a> that you may want to take a look at (just because I posted twice there, yeah&#8230;).</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/08/27/google-engineer-toys-with-64233187whatever.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Long Tail: &#8220;how many words to make page unique google supplemental&#8221;</title>
		<link>http://seo4fun.com/blog/2006/08/25/long-tail-how-many-words-to-make-page-unique-google-supplemental.html</link>
		<comments>http://seo4fun.com/blog/2006/08/25/long-tail-how-many-words-to-make-page-unique-google-supplemental.html#comments</comments>
		<pubDate>Fri, 25 Aug 2006 14:07:42 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/08/25/long-tail-how-many-words-to-make-page-unique-google-supplemental.html</guid>
		<description><![CDATA[I figured I&#8217;ll serve the public occasionally by answer a few questions people are arriving here with, including this lengthy one: &#8220;how many words to make page unique google supplemental.&#8221; I know these questions will probably bore the hell out of most readers, since they&#8217;re un-news-worthy and fall under the beating-a-dead-horse unbrella, but hey, a [...]]]></description>
			<content:encoded><![CDATA[<p>I figured I&#8217;ll serve the public occasionally by answer a few questions people are arriving here with, including this lengthy one: &#8220;how many words to make page unique google supplemental.&#8221; I know these questions will probably bore the hell out of most readers, since they&#8217;re un-news-worthy and fall under the beating-a-dead-horse unbrella, but hey, a few &#8220;timeless&#8221; posts wouldn&#8217;t hurt, right? Future posts like this will be labeled &#8220;long tail&#8221;, so its easy for you to avoid reading them if the topic doesn&#8217;t interest you.</p>
<p>Also, keep in mind, I don&#8217;t pretend to be an SEO and I don&#8217;t play one on TV or Google Video either, so I&#8217;d recommend you verify whatever I post or at least take it with a dose of salt. If I knew everything about SEO, I wouldn&#8217;t be blogging about it :)  Anyway, if you happen to disagree with me, it&#8217;d be awesome if you leave me a comment (concrete examples and a decent rationale behind it preferred). I&#8217;ve been coming across a few lapse-in-logic type statements on forums and Google Groups like &#8220;sitemaps killed my site!&#8221; - but I&#8217;m sure you&#8217;re way smarter than that.</p>
<p>So there you have it, my &#8220;long tail&#8221; posts official disclaimer.</p>
<p>Anyhoo, this guy is wondering <strong>how many words should he have on a page to keep it from turning supplemental.</strong> I wish I knew. From looking at my domains, low PageRank pages 110 words or less (not counting anchor text) largely end up supplemental, especially if its template based. <a href="http://groups.google.com/group/Google_Webmaster_Help-Indexing/browse_thread/thread/735486618dae8b0d">Vanessa Fox also recently mentioned</a> that thin pages aren&#8217;t a good thing:</p>
<blockquote><p>You should also take a look at your site and make sure it provides unique content. <strong>Most of your categories don&#8217;t seem to have any content.</strong> You&#8217;ll need your pages to have value in order to get them indexed.</p></blockquote>
<p>(Google is more about philosophy, passion, and marketing than tinkering with code.)</p>
<p>Since I see other pages on the same low PageRank domain with over 150+ words remaining in the main index, my money&#8217;s on 200+ words per page. This is a debatable figure, of course. Some people will tell you unique content pages with 300+ words have gone supplemental, therefore page size doesn&#8217;t matter. I&#8217;ve seen both extremes - a page with 10 words indexed by Google, and a page with 300+ words wind up in the supplemental index. But as I recently posted on WMW, would you believe me if I told you being 300 lb. overweight isn&#8217;t a factor in getting hot dates because I have three fat relatives that got married to supermodels before they hit 30?</p>
<p>No, I didn&#8217;t think so.</p>
<p>If a 10 word page stays in Google for 5 years, it&#8217;s there in spite of the low word count. The fact that it&#8217;s not supplemental isn&#8217;t a proof that low word count doesn&#8217;t matter. It just means there are other factors that make that page valuable.</p>
<p>BTW, if you need a tool that counts words (except for text in HREF), check out my <a href="/php/text-size.php">word count tool</a>. It&#8217;s buggy and rudimentary, but I use it, so it&#8217;s gotta be good.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/08/25/long-tail-how-many-words-to-make-page-unique-google-supplemental.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Reactive Blogging</title>
		<link>http://seo4fun.com/blog/2006/08/23/reactive-blogging.html</link>
		<comments>http://seo4fun.com/blog/2006/08/23/reactive-blogging.html#comments</comments>
		<pubDate>Wed, 23 Aug 2006 13:12:54 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/08/23/reactive-blogging.html</guid>
		<description><![CDATA[A few years ago, I used to play America&#8217;s Army Online, a squad-based shooter developed by the U.S. Army. To me it&#8217;s a thinking man&#8217;s game, where aim and luck has less to do with winning than exploiting the elements of surprise and forcing your enemy&#8217;s next move.
Never underestimate the element of surprise. Repeating yourself [...]]]></description>
			<content:encoded><![CDATA[<p>A few years ago, I used to play America&#8217;s Army Online, a squad-based shooter developed by the U.S. Army. To me it&#8217;s a thinking man&#8217;s game, where aim and luck has less to do with winning than exploiting the elements of surprise and forcing your enemy&#8217;s next move.</p>
<p><strong>Never underestimate the element of surprise.</strong> Repeating yourself makes you transparent, which gives your enemies an upper hand. Say you rush Oil Entrance with SAW and kill 6 terrorists on Assault on Round 1. They will expect you there the next round. They may park a 203 outside to blow you up if you decide to take a peek outside, or throw a flash through the door to blind you, after which they&#8217;ll toy with you for a few seconds till they put a bullet through your head. The lesson? What worked yesterday won&#8217;t necessarily work today.<br />
<strong><br />
While your opponents force your hand, your odds of winning are low.</strong> Say you&#8217;re on Defense and you hear someone turning EXT pump. You run to EXT room and open the door expecting an easy kill, only to get a round of bullets sprayed on your face with a SAW before you even know what hit you. Someone yanked your chain and you fell for it.</p>
<p>I see the same thing happening on the Interweb. It&#8217;s less effective to blog in reaction to what someone else said than be the pebble in the water that starts a ripple effect.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/08/23/reactive-blogging.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Supplemental Index Fuzzier Than Ever</title>
		<link>http://seo4fun.com/blog/2006/08/17/supplemental-index-fuzzier-than-ever.html</link>
		<comments>http://seo4fun.com/blog/2006/08/17/supplemental-index-fuzzier-than-ever.html#comments</comments>
		<pubDate>Thu, 17 Aug 2006 23:51:59 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/08/17/supplemental-index-fuzzier-than-ever.html</guid>
		<description><![CDATA[On gfe-eh.google.com and other DCs, the supplemental pages cache dates no longer go all the way back to Aug 2005. But the new &#8220;system&#8221; makes it even harder to tell why a page is listed in the supplemental index, because now you&#8217;re required to jump over at least two major hurdles to break out of [...]]]></description>
			<content:encoded><![CDATA[<p>On gfe-eh.google.com and other DCs, the supplemental pages cache dates no longer go all the way back to Aug 2005. But the new &#8220;system&#8221; makes it even harder to tell why a page is listed in the supplemental index, because now you&#8217;re required to jump over at least two major hurdles to break out of supplemental hell: 1) duplicate content issues (i.e. identical meta tags, multiple urls resolving to the same content, www/non-www, etc) and 2) &#8220;Trust&#8221; / PageRank. A perfectly structured page with original content could remain stuck in the supplemental index if a domain lacks juice.</p>
<p>Here&#8217;s how I see it going down (<strong>warning:</strong> pure speculation):</p>
<p>function refresh_supplemental_hell_cache($domain_name) {</p>
<p>   $supplemental_urls = new Urls_in_Supplemental_Hell($domain_name);<br />
   foreach($supplemental_urls->urls as $screwed_url) {<br />
      $page = file_get_contents($screwed_url->url);<br />
      $mysqlupdate(&#8221;UPDATE supplemental_hell SET content=&#8221;$page&#8221;, cache_date=NOW() WHERE url = \&#8221;$screwed_url\&#8221;);<br />
   }</p>
<p>}</p>
<p>That&#8217;s all Google needs to run to refresh the supplemental cache. Then Google re-evaluates each page as if it just found it. Depending on PageRank, trust, number of pages, and multitude of other factors, Google inserts some pages back into the main index, while keeping others out.</p>
<p><strong>Conclusion:</strong> When your mom &#8216;n pop site is 99.99% supplemental, it&#8217;s no longer enough just to &#8220;fix your site.&#8221; You also need to get Google to trust you enough to re-list those supplemental pages back in the main index.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/08/17/supplemental-index-fuzzier-than-ever.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Video Still Keeps its Hands Off Pornographic Material</title>
		<link>http://seo4fun.com/blog/2006/08/14/google-video-still-keeps-its-hands-off-pornographic-material.html</link>
		<comments>http://seo4fun.com/blog/2006/08/14/google-video-still-keeps-its-hands-off-pornographic-material.html#comments</comments>
		<pubDate>Mon, 14 Aug 2006 18:43:52 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/08/14/google-video-still-keeps-its-hands-off-pornographic-material.html</guid>
		<description><![CDATA[Philip Lensen reported a few days ago that &#8220;Google Video now* allows you to upload adult videos.&#8221; (with the added disclaimer that he&#8217;s not sure if the feature is new.) As Jimmy Ruska points out, the &#8220;Adult/Mature&#8221; option is not new; in fact it&#8217;s probably been there from the get-go. UI may have changed, but [...]]]></description>
			<content:encoded><![CDATA[<p><a href="http://blog.outer-court.com/archive/2006-08-12-n20.html">Philip Lensen reported</a> a few days ago that &#8220;Google Video now* allows you to upload adult videos.&#8221; (with the added disclaimer that he&#8217;s not sure if the feature is new.) As <a href="http://www.jimmyr.com/blog/Google_Video_does_NOT_let_you_upload_Pr0n_223_2006.php">Jimmy Ruska points out</a>, the &#8220;Adult/Mature&#8221; option is not new; in fact it&#8217;s probably been there from the get-go. UI may have changed, but Google&#8217;s position hasn&#8217;t. Pornographic videos is still off limits, as you can see from this screen capture:</p>
<p><img src="/images/google-video-adult.jpg" alt="Google Video Adult Mature Content" /></p>
<div align=center><strong>No pornography allowed on Google Video just yet.</strong></div>
<p>Verify what you read on the Interweb.</p>
<p>So, anyway, what does Adult/Mature category cover? <a href="http://www.techcrunch.com/2006/08/13/google-porn/">A guy on Techcrunch</a> linked to http://video.google.com/videoplay?docid=-8578885628445845834&#038;q=blowjob (obviously edited in Windows Movie Maker) as a porn site advertising on Google Video, but that clip shows no nudity. It just links to a spammy landing page with Yahoo ads plastered all over it. Merely sexual behavior and nudity alone isn&#8217;t enough to label a video as pornographic (think of nude / sex scenes in mainstream movies, i.e. love making scene with Nicole Kidman in Cold Mountain - not porn). Now http://video.google.com/videoplay?docid=-7105575458677746453 <strike>is</strike> used to be porn on Google Video</a> (NSFW), clearly in violation of <a href="https://upload.video.google.com/Terms?hl=en">Google Video TOS</a>, which still says &#8220;c) the Authorized Content is not, in whole or in part, <strong>pornographic or obscene</strong>&#8220;. I&#8217;m a little disappointed Google re-encoded my video, since the original is in mouth-watering HiDef. Let&#8217;s see how long that stays up, shall we?</p>
<p><strong>Update (8/15):</strong></p>
<p>I received an email from video.google letting me know the video I submitted one day earlier was rejected. I did tag the hell out of it to make it easier to spot, but this is a good confirmation as any. The email reads:</p>
<p>Hi,</p>
<p>Your video &#8220;Sapphic Chicks High Def Promo&#8221; was rejected because it<br />
didn&#8217;t comply with our policies.</p>
<p>Videos submitted to our program are subject to an initial review to<br />
ensure that they comply with our guidelines. When videos do not meet<br />
our standards, we disapprove them. The following explains our content<br />
policies for uploading videos:</p>
<p>* You must have all necessary legal rights to the content.<br />
* The video must not contain <strong>pornographic, nude, or obscene</strong> material.<br />
* The subject matter in the video must not be illegal.<br />
* The video cannot contain invasions of personal privacy.<br />
* The video cannot contain promotions of hate or incitement of violence.<br />
* The video cannot contain graphic violence or other acts resulting in<br />
serious injury or death.</p>
<p>For more information regarding Google Video policies, please visit:<br />
http://video.google.com/support/bin/answer.py?answer=27737&#038;topic=1490<br />
<strong><br />
Follow UP (Aug 19):</strong></p>
<p>How did they spot my video? Did a Googler actually spend time watching it, as the words in the email &#8220;subject to an initial review&#8221; imply? Or did Google decide on the content by just using the video title, description, tags? I posted a second clip to test this. I&#8217;ll see how long this one stays up.</p>
<p><strong>Aug 22:</strong></p>
<p>Google took 3 days to reject my last video I uploaded. Since I completely misrepresented the content of my video, I&#8217;m pretty convinced Google is manually reviewing videos submitted to them.</p>
<p>I do think Google Video is a step backwards for adult webmasters even if pornography was allowed, because the video quality is still on the low end.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/08/14/google-video-still-keeps-its-hands-off-pornographic-material.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Robots.txt Not Cumulative</title>
		<link>http://seo4fun.com/blog/2006/08/14/robotstxt-not-cumulative.html</link>
		<comments>http://seo4fun.com/blog/2006/08/14/robotstxt-not-cumulative.html#comments</comments>
		<pubDate>Mon, 14 Aug 2006 16:26:10 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/08/14/robotstxt-not-cumulative.html</guid>
		<description><![CDATA[Yesterday, gs1md wrote a meaty post on WMW regarding how Googlebot responds to robots.txt containing directives for both User-Agent:* and User-Agent: Googlebot. Both Googleguy and Vanessa Fox dropped by to clarify the situation: When you use both specifications, Googlebot will go with User-Agent: Googlebot.
Related links:

Google Webmaster Help:robots.txt
Google Webmaster Help: How Do I Block Googlebot?

]]></description>
			<content:encoded><![CDATA[<p>Yesterday, gs1md wrote a meaty post on WMW regarding <a href="http://www.webmasterworld.com/google/3044757.htm">how Googlebot responds to robots.txt</a> containing directives for both User-Agent:* and User-Agent: Googlebot. Both Googleguy and Vanessa Fox dropped by to clarify the situation: When you use both specifications, Googlebot will go with User-Agent: Googlebot.</p>
<p>Related links:</p>
<ul>
<li><a href="http://www.google.com/support/webmasters/bin/answer.py?answer=40360&#038;topic=8846">Google Webmaster Help:robots.txt</a></li>
<li><a href="http://www.google.com/support/webmasters/bin/answer.py?answer=40364">Google Webmaster Help: How Do I Block Googlebot?</a></li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/08/14/robotstxt-not-cumulative.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>GoogleGuy Announces Radically Fresher Supplemental Index</title>
		<link>http://seo4fun.com/blog/2006/08/08/googleguy-announces-radically-fresh-supplemental-index.html</link>
		<comments>http://seo4fun.com/blog/2006/08/08/googleguy-announces-radically-fresh-supplemental-index.html#comments</comments>
		<pubDate>Tue, 08 Aug 2006 15:03:52 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/08/08/googleguy-announces-radically-fresh-supplemental-index.html</guid>
		<description><![CDATA[Earlier this morning, GoogleGuy announced a supplemental index facelift:
 Okay, I believe most/all U.S. users should see radically fresher supplemental results now. The earliest page I saw was from Feb 2006, and most of the ones that I looked at averaged in the ~2 month old range.
As data gets copied to more places, the fresher [...]]]></description>
			<content:encoded><![CDATA[<p>Earlier this morning, <a href="http://www.webmasterworld.com/google/3035462-7-10.htm">GoogleGuy announced</a> a supplemental index facelift:</p>
<blockquote><p> Okay, I believe most/all U.S. users should see <strong>radically fresher supplemental results</strong> now. The earliest page I saw was from Feb 2006, and most of the ones that I looked at averaged in the ~2 month old range.</p>
<p>As data gets copied to more places, the fresher supplemental results should eventually be visible everywhere, not just the U.S. </p></blockquote>
<p>Unfortunately, I&#8217;m either hitting the wrong DCs or I&#8217;m not seeing any change for my sites. So I won&#8217;t be doing cartwheels anytime soon&#8230;</p>
<p>Holy *** - checking 72.14.207.99 I&#8217;m seeing clean results and a whole load of new pages indexed correctly! If that spreads out to other DCs&#8230; well, I won&#8217;t count my chickens till they hatch, but I&#8217;m close to doing cartwheels. Thanks for the heads up Googleguy.</p>
<p>UPDATES:</p>
<p>Posted on <a href="http://www.webmasterworld.com/google/3035462-12-10.htm">Aug 11 at WMW</a>:</p>
<blockquote><p>
trinorthlighting, the refresh of supplemental results went out to a data center that serves North American and Asian traffic. If you went looking at specific data centers, you could probably find older supplemental pages&#8211;however, as the newer supplemental results roll out at more data centers in the next few weeks, they will completely replace the older supplemental results. </p></blockquote>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/08/08/googleguy-announces-radically-fresh-supplemental-index.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Vanessa Fox and Susie Bright Mingling at BlogHer?</title>
		<link>http://seo4fun.com/blog/2006/08/03/vanessa-fox-and-susie-bright-mingling-at-blogher.html</link>
		<comments>http://seo4fun.com/blog/2006/08/03/vanessa-fox-and-susie-bright-mingling-at-blogher.html#comments</comments>
		<pubDate>Thu, 03 Aug 2006 19:45:08 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/08/03/vanessa-fox-and-susie-bright-mingling-at-blogher.html</guid>
		<description><![CDATA[Earlier today I read Susie Bright talk about &#8220;the elephant in the Blogher living room&#8221;, so I was somewhat surprised to just find out Vanessa Fox was also there. That they have two completely different take on the convention is obviously a no shocker. But it doesn&#8217;t hurt to run a little comparison, does it? [...]]]></description>
			<content:encoded><![CDATA[<p>Earlier today I <a href="http://susiebright.blogs.com/susie_brights_journal_/2006/07/blogher_sex_sur.html">read Susie Bright talk about &#8220;the elephant in the Blogher living room&#8221;</a>, so I was somewhat surprised to just find out <a href="http://sitemaps.blogspot.com/2006/08/back-from-blogher.html">Vanessa Fox was also there</a>. That they have two completely different take on <a href="http://www.blogher.org/">the convention</a> is obviously a no shocker. But it doesn&#8217;t hurt to run a little comparison, does it? This bit&#8217;s from a post written by Susie I dug up a few minutes ago off Google Blogsearch:</p>
<blockquote><p>
This post is for the women at the Blogher conference that I just attended. We had a session about sex and blogging that I thought was the best, and most controversial session of the gathering.</p>
<p>I decided to throw together  an  Un-Official, Un-Authorized, <a href="http://susiebright.blogs.com/susie_brights_journal_/2006/07/blogher_sex_sur.html">Blogher Sex Survey</a>!
</p></blockquote>
<p>And here&#8217;s the <a href="http://susiebright.blogs.com/susie_brights_journal_/2006/08/blog_the_cradle.html">post I was reading</a> earlier this morning:</p>
<blockquote><p>
I&#8217;m home from Blogher, the women&#8217;s blogging meet-up, that happened this past weekend. 750 women bloggers&#8230; kind of shocking. We all came out of our cubbyholes.</p>
<p>I&#8217;ve never been to an IT gathering of any kind before. I live on the coast next to Silicon Valley,but I&#8217;ve never been to <a href="http://maps.google.com/maps?q=San+Jose,+CA&#038;ie=UTF8&#038;ll=37.335224,-121.893311&#038;spn=5.869005,13.886719&#038;om=1">San Jose</a> except to drive to the airport&#8230;.I had no idea that a group of female computer nuts, blogging women, were so revolutionary&#8230;In our sex blog workshop at Blogher, several women raised their concern of writing about sex publicly, in any context. They were fearful of ridicule, discrimination, and dismissive stereotypes&#8230;The conference was astounding for the authority of its women speakers. You can find &#8220;pretty&#8221; girls anywhere— how often can you find ones who can rewire your whole world?</p></blockquote>
<p>Susie Bright made an appearance along side Melissa Gira, Logan Levkoff (meow!), and Halley Suitt on <a href="http://blogher.org/node/7383">Let&#8217;s Talk about Sex</a> on Saturday July 29th, 2006. In her post, Susie goes on to quote Guy Kawasaki, who effectively tells readers who object to reading &#8220;non-business, non-tech, non-male subjects&#8221; to take a hike.</p>
<p>On the other hand, <a href="http://sitemaps.blogspot.com/2006/08/back-from-blogher.html">Vanessa Fox</a> seemed to carefully craft her post to meet her target audience - mostly geeky male webmasters trying to get our sites back in Google&#8217;s index, I suppose. Yet she sounds inspired. 750 women, many of whom &#8220;provide unique perspectives and content on topics&#8221;, of which Susie Bright is an outstanding example, means improving Google and helping female bloggers understand Google better will lead to their unique voices being exposed to a wider audience.</p>
<blockquote><p>I just got back from BlogHer, a conference primarily for women about the technical and community aspects of blogging. As a woman who blogs, I had a wonderful time. As a woman who blogs about topics of interest to site owners, I gained some new perspectives&#8230;The panelists understood visitor awareness. They told the crowd to look closely at their sites to determine how unique they were from other sites out there. They talked about getting visitors to care. This thoughtfulness leads to great sites, which in turn can lead to great search results&#8230;Given all of the dedication these bloggers put into their sites, they are of course interested in attracting visitors. Some want the joy of sharing; others are interested in making money from their writing. Here are a few tips to help make your site easier to find, whatever your motivation:</p></blockquote>
<p>She goes on (as you probably already read) to emphasize high quality links, crawlability (echoing Matt Cutts), and the ease of verifying site ownership for Google Sitempas via <a href="http://sitemaps.blogspot.com/2006/05/more-about-meta-tag-verification.html">META tags</a>. What kinda touched me was the human passion I sensed behind all the techno talk, which to me is a hundred times more motivating than a few checks in the mail (though those don&#8217;t hurt either).</p>
<p><strong>P.S.</strong> Interesting though that the cocktail parties and such were sponsored by Windows Live Spaces and Yahoo, not Google. BTW, here are <a href="http://blogher.org/profile/profile_blogher06?from=0">some of the women who attended Blogher</a> - definitely some lookers in the crowd.</p>
<p><strong>Update Aug. 4</strong>: <a href="http://www.flickr.com/photos/tags/blogher2006/">BlogHer photos on Flickr</a>, tip from <a href="http://lustylady.blogspot.com/2006/08/okay-next-year-im-going-to-blogher.html">Village Voice&#8217;s Rachel Kramer</a>.</p>
<p><strong>Update Aug. 8</strong>: <a href="http://www.ysearchblog.com/archives/000340.html">Yahoo!&#8217;s Havi Hoffman on BlogHer</a>: &#8220;Personally, I think the noisy aftermath is a testimonial to the extraordinary range of people who participated.&#8221; One more tidbit: &#8220;Yahoo! slipped a purple pen and notebook into the schwag bag. We dressed the pool deck in purple and served plenty of our famous Yahootinis at the closing night cocktail party.&#8221; I want one of &#8216;em Yahootinis.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/08/03/vanessa-fox-and-susie-bright-mingling-at-blogher.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Guy, Matt Cutts, Vanessa Fox, and Adam Lasnik in One Thread?</title>
		<link>http://seo4fun.com/blog/2006/08/02/google-guy-matt-cutts-vanessa-fox-and-adam-lasnik-in-one-thread.html</link>
		<comments>http://seo4fun.com/blog/2006/08/02/google-guy-matt-cutts-vanessa-fox-and-adam-lasnik-in-one-thread.html#comments</comments>
		<pubDate>Wed, 02 Aug 2006 13:49:42 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/08/02/google-guy-matt-cutts-vanessa-fox-and-adam-lasnik-in-one-thread.html</guid>
		<description><![CDATA[In the last two days, both Vanessa Fox and Adam Lasnik made cameo appearances in the second installment of WMW&#8217;s Google Webmaster Communications thread. Now that&#8217;s Googleguy, Matt Cutts, Vanessa Fox, and Adam Lasnik all in one thread. Is this for real? Reading reseller greet them (&#8221;Its inspiring and a great pleasure to see our [...]]]></description>
			<content:encoded><![CDATA[<p>In the last two days, both Vanessa Fox and Adam Lasnik made cameo appearances in <a href="http://www.webmasterworld.com/google/3031050.htm">the second installment of WMW&#8217;s Google Webmaster Communications thread</a>. Now that&#8217;s Googleguy, Matt Cutts, Vanessa Fox, and Adam Lasnik all in one thread. Is this for real? Reading reseller greet them (&#8221;Its inspiring and a great pleasure to see our good Googlers friends and WebmasterWorld fellow members&#8230;&#8221;), I feel like I&#8217;m watching an SEO version of The Fellowship of the Ring. Still, there&#8217;s alot of griping and grumbling going on, and there&#8217;s not much to take away from the thread so far except a hope that a good suggestion or a question by one of us may get some revealing responses in the days ahead.</p>
<p>I suggested Google Sitemaps show timestamps of when Supplemental Bot last visited a page, though I doubt Vanessa will bite. I&#8217;d also like to know if Supplemental Bot crawl depth or priority depends on trust / PR. I don&#8217;t plan to hold my breath for too long, but it would be great to know Supplemental Bot crawl priority is trust independent. Otherwise, a PR 0 site with no trust gone 100% supplemental will never see the light of day - which doesn&#8217;t exactly describe any of my sites. My domains are at least PR 2, dammit.</p>
<p><strong>UPDATE (8/29/2006):</strong> Since I&#8217;m getting Google hits from people looking for the dirt on Vanessa Fox, I&#8217;ll list a few relevant links here so you can track her down. First place to look for Vanessa is <a href="http://groups.google.com/group/Google_Webmaster_Help?lnk=li">Google Groups/ Google Webmaster Help</a>. Just look for the green &#8220;G&#8221; logo, and you&#8217;ll find either Adam Lasnik or Vanessa helping webmasters out of tight spots. You can also find Vanessa blogging away at either <a href="http://googleblog.blogspot.com/">The Official Google Blog</a> or <a href="http://googlewebmastercentral.blogspot.com/">Webmaster Central Blog</a>. She also drops by at Webmasterworld as well as other forums like cre8site (links on my sidebar). Finally, I also wrote a bit about <a href="/blog/2006/08/03/vanessa-fox-and-susie-bright-mingling-at-blogher.html">Vanessa Fox blogging about BlogHer in this post</a>.</p>
<p>UPDATE 99/20/2006): Vanessa Fox and Amanda Camp in <a href="http://www.seattle24x7.com/up/googlecentral.htm">A Conversation with Google Webmaster Central</a>. Check out <a href="http://www.seattle24x7.com/people/people.htm">their pics here</a>. Wow..they&#8217;re kinda cute. Why the hell am I not working at Google?</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/08/02/google-guy-matt-cutts-vanessa-fox-and-adam-lasnik-in-one-thread.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Matt Cutts Releases SEO Video Quickies</title>
		<link>http://seo4fun.com/blog/2006/07/31/matt-cutts-releases-seo-video-quickies.html</link>
		<comments>http://seo4fun.com/blog/2006/07/31/matt-cutts-releases-seo-video-quickies.html#comments</comments>
		<pubDate>Mon, 31 Jul 2006 13:05:24 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/07/31/matt-cutts-releases-seo-video-quickies.html</guid>
		<description><![CDATA[Perhaps as part of an extended response to his recent cameo appearance on a WMW thread started by Reseller, Matt Cutts posted three videos answering questions asked previously on The Simple Life, er Grabbag Friday. Here&#8217;s the cliff notes:

Does updates of sitemaps depend on page views? Matts: No, page views isn&#8217;t really a factor when [...]]]></description>
			<content:encoded><![CDATA[<p>Perhaps as part of an extended response to his <a href="http://www.webmasterworld.com/google/3022278.htm">recent cameo appearance on a WMW thread</a> started by Reseller, <a href="http://www.mattcutts.com/blog/seo-answers-on-google-video/">Matt Cutts posted three videos</a> answering questions asked previously on The Simple Life, er <a href="http://www.mattcutts.com/blog/grabbag-friday/">Grabbag Friday</a>. Here&#8217;s the cliff notes:</p>
<ul>
<li><strong>Does updates of sitemaps depend on page views?</strong> Matts: <em>No</em>, page views isn&#8217;t really a factor when things are updated in sitemaps.</li>
<li><strong>How do I improve my site&#8217;s visibility on Google?</strong> <em>(make sure its crawlable, then market it with a hook).</em> 1) Make sure the site is crawlable. Try viewing your site via a text browser, like Lynx. Use an HTML sitemap to help spiders along; use Google Sitemaps in addition to that. 2) Market your site - &#8220;think about the people who are relevant to your niche, and make sure they know about you. You also wanna be thinking about a hook - something that&#8217;s viral.&#8221; Good content, social bookmarking sites, etc. &#8220;Fundamentally you need something interesting that sets you apart from the pack.&#8221;</li>
<li><strong>What conditions cause Google to use the DMOZ snippet, when there is already a valid meta description tag on the page?</strong> <em>Result is query-dependent.</em> If you have a page about Christina Aguirella but your DMOZ listing says the page is about Britney Spears, a &#8220;britney spears&#8221; search will display the DMOZ description, whereas a &#8220;christina aquirela&#8221; search will display the page&#8217;s meta description snippet. Of course, if you don&#8217;t want the DMOZ description, you can use the NOODP meta tag.</li>
<li><strong>Does Google favor bold or strong?</strong> In general, <em>Google slightly favors bold</em>. Keyword: slightly. Recommendation: Just code your site the way you want. Update: In <a href="http://video.google.com/videoplay?docid=-1756437348670651505">his latest video</a>, Matt Cutts mentioned he followed up with a Google engineer about this and was shown that Google&#8217;s algo treats B exactly the same as STRONG, and EM exactly the same as I. So, there ya go.</li>
<li><strong>Does having many sites under one IP/server matter?</strong> There&#8217;s a range: <em>4-5 sites on a similar theme = no problem. 2000 sites = probably spam.</em></li>
<li><strong>If you&#8217;re launching a site with millions of pages, launch softly.</strong></li>
<li><strong>Don&#8217;t worry about including the same JS off different sites</strong>, Google Analytics or Google Adsense for example. People do this alot and unless you use the same sneaky code on 5000 sites, you shouldn&#8217;t worry.</li>
<li><strong>Google image index update happened last weekend.</strong> You can read Matt Cutts&#8217; comment regarding the update on <a href="http://www.webmasterworld.com/google/3018232.htm">a recent WMW thread titled Google: Has There Been an Image Update?</a> started by ianevans.</li>
<li><strong>Should I build for search engines, or for people?</strong> <em>Both SE optimization AND end-user optimization is important</em> - otherwise &#8220;you won&#8217;t do as well.&#8221; &#8220;The trick in my mind, is to try to see the world such that they&#8217;re the same thing. You want to make it so that your users&#8217; interests and the search engines&#8217; interests are aligned as you can.&#8221;</li>
<li><strong>How do I detect spam?</strong> Try <a href="http://siteexplorer.search.yahoo.com/">Yahoo site explorer</a>, which shows you backlinks. Tools to show you everything on one IP address. Use <a href="https://www.google.com/webmasters/sitemaps/siteoverview?hl=en">Google Sitemaps</a> to find any problems with your own sites.</li>
<li><strong>WC3 validation, does it matter?</strong> &#8220;40% of all html has syntax errors. and there&#8217;s no way search engines can remove 40% of its index, just because somebody didn&#8217;t validate. I wouldn&#8217;t put it at the top of the list.&#8221; Could be a signal in the future especially after the release of accessibility search.</li>
</ul>
<p><strong>Follow Up:</strong> Matt posted <a href="http://www.mattcutts.com/blog/more-seo-answers-on-video/">another batch of videos</a> today. In the last video, he said Supplemental googlebot will be refreshing pages and following redirects more frequently as time goes on. I&#8217;ll believe it when I see it.</p>
<p><strong>Aug 2:</strong> In <a href="http://video.google.com/videoplay?docid=-9028425054136856586">this video released today</a>, where Matt makes me want to go grab a drink from the fridge, he mentions something that will have all adult webmasters&#8217; heads reeling. I know many people who like to put a whole bunch of adult keywords in their meta keywords and description tags. Those people are also usually not so fond of using ICRA tags. According to Matt, Google uses those tags to filter out adult content from Safe Search. Gotta love it.</p>
<p>He also goes into explaining <a href="http://video.google.com/videoplay?docid=8475081922887713591">what happened on June 27</a> and the difference between an index update, an algo update, and a data refresh. Matt mentions PageRank enough times in this video to make PR naysayers cringe. Few keypoints: Google&#8217;s index is in an everflux (no more updates ala 2003); PageRank updates daily; over-optimization can hurt your site. Yes we all knew that already, but what exactly is data refresh again?</p>
<p><strong>Aug 3:</strong> Matt Cutts throws another curve ball at SEOs in <a href="http://video.google.com/videoplay?docid=-1756437348670651505">this new video</a>, where he dismisses a .gov myth and credits the power of .edu sites mainly to PageRank.  According to Matt, .gov/.org/.edu sites aren&#8217;t really treated differently from any other type of domains. &#8220;It&#8217;s just those sites tend to have higher PageRank&#8230;because more people link to them.&#8221;</p>
<p><strong>P.S.</strong></p>
<p>I left a question on his blog regarding WC3 because I believe some errors result in crappy description snippets being displayed on a Google site: search.</p>
<p>BTW, here&#8217;s a few GoogleGuy quotes from the same WMW Google communication thread, where a couple of people bitched about Google not responding to frustrated webmasters :</p>
<blockquote><p>There was a data refresh on June 27th that lots of people ask about, but there was also a data refresh in the last 1-2 days that refreshes the same data. Going forward, I&#8217;d expect that the cycle time would go down even more, possibly down to once a week for that particular algorithm.</p></blockquote>
<blockquote><p>72.14.207.104 has some newer infrastructure that makes site: queries more accurate, and in general that infrastructure also improves results for other queries too. But the infrastructure at 72.14.207.104 is orthogonal/independent of many other changes.</p></blockquote>
<p>BTW, Thanks to <a href="http://blog.v7n.com/2006/07/31/cutt-lets-early-christmas-present/">Peter over on V7N Blog</a> for throwing me a bone. I hate to bitch but writing a full-fledged post at 4 in the morning just to publish something that <a href="http://www.seroundtable.com/archives/004251.html">reads almost exactly like every other Matt Cutts video blog post</a> can get frustrating. No point in speedy delivery when you got no visibility.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/07/31/matt-cutts-releases-seo-video-quickies.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Corrupt Titles and Short META Descriptions</title>
		<link>http://seo4fun.com/blog/2006/07/18/google-corrupt-titles-and-short-meta-descriptions.html</link>
		<comments>http://seo4fun.com/blog/2006/07/18/google-corrupt-titles-and-short-meta-descriptions.html#comments</comments>
		<pubDate>Tue, 18 Jul 2006 17:27:44 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/07/18/google-corrupt-titles-and-short-meta-descriptions.html</guid>
		<description><![CDATA[Marcia on WMW commented yesterday that pages with unlinked style declarations are turning up with corrupt TITLEs. Just Guessing claims he has the same problem but without a style declaration (mine does have a style declaration and their titles are still corrupt). Which brings us back to square one.
The thread got me wondering if having [...]]]></description>
			<content:encoded><![CDATA[<p><a href="http://www.webmasterworld.com/google/3011487.htm">Marcia on WMW</a> commented yesterday that pages with unlinked style declarations are turning up with corrupt TITLEs. Just Guessing claims he has the same problem but without a style declaration (mine does have a style declaration and their titles are still corrupt). Which brings us back to square one.</p>
<p>The thread got me wondering if having short META descriptions are one of the common denominators. Don&#8217;t get me wrong - I&#8217;m not saying short META descriptions are the CAUSE - just wondering if all pages with corrupt titles have META descriptions less than 50 chars long. (Disagree? Post a counterexample.)</p>
<p>Google&#8217;s been changing the way they snippetize pages and they may have temporarily injected their spider with a bug - in which case we can forget and move on. Still, this minor corrupt title bug can turn into a huge mess since Supplemental Bot revisits websites supposedly around every 6 months (though for my sites it feels more like once a year) and most if not all pages with this corrupt title problem ends up in the supplemental index.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/07/18/google-corrupt-titles-and-short-meta-descriptions.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Notebook - Error in Loading User Data</title>
		<link>http://seo4fun.com/blog/2006/07/14/google-notebook-error-in-loading-user-data.html</link>
		<comments>http://seo4fun.com/blog/2006/07/14/google-notebook-error-in-loading-user-data.html#comments</comments>
		<pubDate>Sat, 15 Jul 2006 04:05:00 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/07/14/google-notebook-error-in-loading-user-data.html</guid>
		<description><![CDATA[I admit I love Google Notebook. I&#8217;ve got all sorts of clips saved up in this thing as I surf the web. But since 5 pm tonight I&#8217;ve been getting this error: &#8220;Error in loading user data&#8221; and it refuses to go away. Yesterday, gmail refused to load for a few minutes, so I thought [...]]]></description>
			<content:encoded><![CDATA[<p>I admit I love Google Notebook. I&#8217;ve got all sorts of clips saved up in this thing as I surf the web. But since 5 pm tonight I&#8217;ve been getting this error: &#8220;Error in loading user data&#8221; and it refuses to go away. Yesterday, gmail refused to load for a few minutes, so I thought maybe it was the same momentary lapse of service kinda deal, but 4 hours later, Google Notebook still refuses to fire up. What gives? I swear to God things like this make me wanna be DaveN. Recommendation: <strong>Send Google an email letting them know your problem</strong> - it worked for me.</p>
<div align=center>
<img src="/images/google-notebook.jpg" alt="google notebook error" /><br />
<strong>Googe Notebook Error in Loading User Data</strong>
</div>
<p><strong>Update:</strong> Now Google Notebook is really acting screwy:</p>
<div align=center>
<img src="/images/google-notebook-screwup.jpg" alt="google notebook screw up" /><br />
<strong>Googe Notebook Screenshot. Nav links on the left are unclickable . I minimized resolution to protect some sensitive data..so to speak.</strong>
</div>
<p><strong>7/19/2006:</strong> <strong>Google Notebook is back to working order.</strong> About the same time I noticed the bugs disappear, I received a reply to my Google labs feedback (bitching) informing me they fixed a few kinks in their app. I was afraid when Google apps bug out, those bugs go unnoticed by the Google folks, but in this case my fears turned out to be unfounded. Now I&#8217;m officially back to being a Google Notebook fanboy.</p>
<p><strong>8/25/2006:</strong> I&#8217;m still getting hits from Google (e.g. &#8220;error in loading user data&#8221;, &#8220;google notebook error in loading user data&#8221;), so people are still having problems with Google Notebook, though I haven&#8217;t had problems with it for a while. I think it has to do with keeping the Notebook open in multiple browsers and creating a syncronization problem (just a guess).</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/07/14/google-notebook-error-in-loading-user-data.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google&#8217;s Vanessa Fox Announces NOODP Tag</title>
		<link>http://seo4fun.com/blog/2006/07/13/googles-vanessa-fox-announces-noodp-tag.html</link>
		<comments>http://seo4fun.com/blog/2006/07/13/googles-vanessa-fox-announces-noodp-tag.html#comments</comments>
		<pubDate>Thu, 13 Jul 2006 21:19:49 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/07/13/googles-vanessa-fox-announces-noodp-tag.html</guid>
		<description><![CDATA[A couple of hours ago, Vanessa Fox announced a new META tag that allows webmasters to opt out of DMOZ title/descriptions appearing in the SERPs:
&#60;META NAME=&#8221;GOOGLEBOT&#8221; CONTENT=&#8221;NOODP&#8221;&#62;
or &#60;META NAME=&#8221;ROBOTS&#8221; CONTENT=&#8221;NOODP&#8221;&#62; to cover all search engines.
MSN was of course the first to get the ball rolling on this, back on May 22.
Vanessa also wrote:
The way we [...]]]></description>
			<content:encoded><![CDATA[<p>A couple of hours ago, <a href="http://sitemaps.blogspot.com/2006/07/more-control-over-page-snippets.html">Vanessa Fox announced</a> a new META tag that allows webmasters to opt out of DMOZ title/descriptions appearing in the SERPs:</p>
<p>&lt;META NAME=&#8221;GOOGLEBOT&#8221; CONTENT=&#8221;NOODP&#8221;&gt;</p>
<p>or &lt;META NAME=&#8221;ROBOTS&#8221; CONTENT=&#8221;NOODP&#8221;&gt; to cover all search engines.</p>
<p><a href="http://blogs.msdn.com/livesearch/archive/2006/05/22/603917.aspx">MSN was of course the first</a> to get the ball rolling on this, back on May 22.</p>
<p>Vanessa also wrote:</p>
<blockquote><p>The way we generate the descriptions (snippets) that appear under a page in the search results is completely automated. The process uses both the content on a page as well as references to it that appear on other sites.</p></blockquote>
<p>In case you&#8217;re one of those unlucky people who&#8217;s seeing nav links pop up in the search results, I&#8217;ve written previously about <a href="/blog/2006/06/24/seo-101-feeding-google-juicy-description-snippets.html">how Google snippetizes a page</a>.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/07/13/googles-vanessa-fox-announces-noodp-tag.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Log - Emailed Business.com and Built a New Rig</title>
		<link>http://seo4fun.com/blog/2006/07/12/log-emailed-businesscom-and-built-a-new-rig.html</link>
		<comments>http://seo4fun.com/blog/2006/07/12/log-emailed-businesscom-and-built-a-new-rig.html#comments</comments>
		<pubDate>Thu, 13 Jul 2006 00:07:02 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/07/12/log-emailed-businesscom-and-built-a-new-rig.html</guid>
		<description><![CDATA[I finally finished putting my new PC together and I gotta say after frying my old rig, it felt like walking on a tight rope without a break for 2 days (gimme a break, it&#8217;s my first try). Seriously, I connect the reset wire to my new ASUS ATX mobo in bullet time. First time [...]]]></description>
			<content:encoded><![CDATA[<p>I finally finished putting my new PC together and I gotta say after frying my old rig, it felt like walking on a tight rope without a break for 2 days (gimme a break, it&#8217;s my first try). Seriously, I connect the reset wire to my new ASUS ATX mobo in bullet time. First time I pressed the power button, I had one hand on the plug in case the fan didn&#8217;t spin. And when my DVD drive refused to show up, I was afraid I shorted a fuse - but it turned out I just had to set the damn thing on slave instead of CS. One thing I was glad of was I didn&#8217;t have to reformat my hard drive. Windows XP fired up just fine - I just had to call up Microsoft tech support to retrieve my key. Windows bonks out like that whenever it detects a major alteration in your system - like upgrading everything except the drives, which is what I ended up doing. I wish I had a digital cam to snap a pic since the back of my jet black case with a glowing blue PSU looks pretty damn cool. I have to remember never to throw away my old case though, since I got my Microsoft key sticker on it.</p>
<p>After I was done, I spent a day playing F.E.A.R and Far Cry (one good thing about being my own boss). My new Radeon X1800XT, Athelon AMD 64 2.2Ghz 2 GB corsair RAM set up still isn&#8217;t quite good enough to max out FEAR settings, but Far Cry plays flawlessly with Anti Aliasing on high. I wish I went SLI with a pair of GeForce 7900s, but I didn&#8217;t feel like blowing 800+ on a pair of video cards when I could buy an Xbox 360 for around 300. Now I gotta go get me some new titles: probably Oblivion (though that sounds like it&#8217;ll put me in a timewarp and I won&#8217;t ever get back to my work), Doom 3 for supposedly great looking visuals (I&#8217;ll have to re-download the demo to decide), or the Need to Speed Most Wanted. Whatever with tons of eye candy will do.</p>
<p>Anyway, that&#8217;s my excuse to myself for doing barely anything for the last few days. I did email business.com though early on in the week asking them why they added nofollow tags on their links and whether their directory was a paid directory. They promptly emailed me back, refering me to Threadwatch where Lane posted an elaborate response, which I was satisfied with. People go berzerk when their paid links are at stake, I tell ya.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/07/12/log-emailed-businesscom-and-built-a-new-rig.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Snippetization Experiment Follow-Up</title>
		<link>http://seo4fun.com/blog/2006/07/07/snippetization-experiment-follow-up.html</link>
		<comments>http://seo4fun.com/blog/2006/07/07/snippetization-experiment-follow-up.html#comments</comments>
		<pubDate>Fri, 07 Jul 2006 07:41:10 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/07/07/snippetization-experiment-follow-up.html</guid>
		<description><![CDATA[Just a follow-up to my last post about description snippets, posted on July 3, 2006. Lucky for me, Google cached a few pages on the same day, including this category page:

This is what I predicted. BTW, all I did was remove an H2 element so that H2 preceeds a text block, instead of having an [...]]]></description>
			<content:encoded><![CDATA[<p>Just a follow-up to my <a href="/blog/2006/07/03/experiment-with-description-snippets.html">last post about description snippets</a>, posted on July 3, 2006. Lucky for me, Google cached a few pages on the same day, including this category page:</p>
<p><img src="/images/description-modified.jpg" alt="description snippet modified" /></p>
<p>This is what I predicted. BTW, all I did was remove an H2 element so that H2 preceeds a text block, instead of having an H2 immediately follow another H2.</p>
<p>Now you&#8217;re probably thinking, &#8220;What does this textbook seo voodoo have to do with viral marketting, increasing my reputation, or improving my CTR?&#8221;</p>
<p>In case you haven&#8217;t read it yet, here&#8217;s <a href="http://www.highrankings.com/advisor/lowbodycopy/">a recent remark Jill Whalen made</a> in her High Rankings Adivsor newsletter regarding the importance of placing content high up in the source:</p>
<blockquote><p>This is an old SEO myth. It actually makes no difference where in the source code the copy of the page shows up.  The search engines have always known how to ignore the HTML code that is not important to them, and can easily find the “meat” that is important.</p></blockquote>
<p>Let&#8217;s break this down, shall we?</p>
<blockquote><p>It actually makes no diffeence where in the source code the copy of the page shows up.</p></blockquote>
<p>Take a look at this <a href="http://www.google.com/search?hs=d9f&#038;hl=en&#038;lr=&#038;safe=off&#038;client=firefox-a&#038;rls=org.mozilla%3Aen-US%3Aofficial&#038;q=site%3Abrooke-skye-videos.blogspot.com&#038;btnG=Search">site: search</a> (don&#8217;t click if you&#8217;re pornophobic, but hey, its just google SERP). 20 some pages, with plenty of unique wordy posts, all but one page supplemental. Tedster from WMW would tell you they&#8217;re supplemental because Google picked up identical description snippets for every page: &#8220;<em>Notify Blogger about objectionable content. What does this mean? Blogger. Send As SMS. Get your own blog Flag Blog Next blog. BlogThis !</em>&#8221; Agreed. Google is judging my blog based not on the page copy, but the text snippet right below BODY. If the blog template placed the &#8220;Notify Blogger&#8221; text below content, that blog wouldn&#8217;t be having a problem.</p>
<p>Don&#8217;t get me wrong. I&#8217;m not saying copy position in the source must always be high up. It depends on the HTML elements you use. But deeper you bury your copy text in the source, more likely Google will get lost.</p>
<blockquote><p> The search engines have always known how to ignore the HTML code that is not important to them</p></blockquote>
<p>Google doesn&#8217;t ignore any HTML when parsing a page. NOSCRIPT, IMG ALT, TD, DIV, even HR - they all dictate how Google reacts to a page. HTML elements add meaning; Google uses them as markers to find the meat. When those elements are misarranged, you are making Google&#8217;s life more difficult.  An obvious sign of trouble is when you see Google defaulting to snippetizing the top of the source (e.g. nav link text, A HREF, image ALT, NOSCRIPT) instead of content.</p>
<p>Think of page copy position as the distance between you and a hoop on a basketball court. You move farther away from the hoop - you may still make the shot, but you&#8217;re decreasing your odds. If you know anything about games, winning is about playing your odds (though I&#8217;ve beat an open player in a 9-ball tournament finals making a 89 degree cut shot across the short side of the table instead of playing safe and forcing a ball in hand or a volley of safes, which would &#8216;ve been the smart thing to do).</p>
<blockquote><p>[search engines] can easily find the “meat” that is important.</p></blockquote>
<p>Clearly, Google can&#8217;t always find the important meat. If it could, identical description snippets wouldn&#8217;t throw it off and make it think its running into duplicate pages. But that&#8217;s what happens. </p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/07/07/snippetization-experiment-follow-up.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Experiment with Description Snippets</title>
		<link>http://seo4fun.com/blog/2006/07/03/experiment-with-description-snippets.html</link>
		<comments>http://seo4fun.com/blog/2006/07/03/experiment-with-description-snippets.html#comments</comments>
		<pubDate>Mon, 03 Jul 2006 20:52:56 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/07/03/experiment-with-description-snippets.html</guid>
		<description><![CDATA[I just removed H2 elements from archive page posts on this blog; I want to see what effect that has on how Google snippetize my archive pages. HOPEFULLY I didn&#8217;t just shoot myself in the foot, but what the hell - I&#8217;m in the mood for a little experimentation. BTW, this is a follow up [...]]]></description>
			<content:encoded><![CDATA[<p>I just removed H2 elements from archive page posts on this blog; I want to see what effect that has on how Google snippetize my archive pages. HOPEFULLY I didn&#8217;t just shoot myself in the foot, but what the hell - I&#8217;m in the mood for a little experimentation. BTW, this is a follow up to my previous post about <a href="/blog/2006/06/24/seo-101-feeding-google-juicy-description-snippets.html">how I think Google snippetizes a page</a>. Note: The point of this excercise is not to rank higher. The point is to 1) learn how Google operates and 2) to avoid supplemental listings without using META descriptions (Look ma, no hands!).</p>
<p><strong>BEFORE</strong> (post titles are wrapped in H2):</p>
<p><img src="/images/archive-template.jpg" alt="Wordpress archive.php template" /></p>
<p><strong>AFTER</strong> (h2 removed):</p>
<p><img src="/images/archive-after.jpg" alt="Wordpress archive.php template" /></p>
<p>What effect will this have when I run site: on Google a few weeks down the road?</p>
<p><strong>BEFORE</strong> (this is how my listing actually looks right now):</p>
<p><img src="/images/before.jpg" alt="Google SERP description snippet before" /></p>
<p><strong>AFTER</strong> (image below was taken from my Google Notebook. I&#8217;ll have to wait a while to see the actual outcome):</p>
<p><img src="/images/after.jpg" alt="Google SERP description snippet after" /></p>
<p>&#8220;Archive for May, 2006&#8243; is the first H2 text on that page, but it&#8217;s followed by another H2 which I think cancels it out. By removing the post H2, I&#8217;m hoping the first H2 text gets included in the description snippet. &#8220;Google Co-op Annotation&#8230;&#8221; is the post title for the first post on that archive page. Since it&#8217;s separate from the post excerpt wrapped in P, I&#8217;m not sure if it&#8217;ll get skipped over or not. I&#8217;ll bet a buck and a half it won&#8217;t get skipped over.</p>
<p><strong>P.S.</strong> I realize now my archive pages are harder on the eyes. I&#8217;ll work on that later.</p>
<p>Since my main machine is fried and I don&#8217;t have access to Dreamweaver, I used Faststone, SmartFTP and Photoshop to work images into this post. Surprisingly, its way quicker this way than using Dreamweaver.</p>
<p>Ideally, I want to wrap everything (post title, date, excerpt) in P. But doing that requires an involved hack which I think is way too much hassle. I still wrote it down below, in case I change my mind later and implement it.</p>
<p>$post_excerpt is just a DB field. An easy hack here is use str_replace to strip out the P and /P in get_the_excerpt().</p>
<p>In wp-includes/template-functions-post.php, line 109:</p>
<p>$output = $post->post_excerpt;</p>
<p><strong>Replace that with:</strong></p>
<p>$replacement = array(&#8221;&lt;p&gt;&#8221;,&#8221;&lt;/p&gt;&#8221;);<br />
$output = str_replace($replacement, &#8220;&#8221;, $post->post_excerpt);</p>
<p>Now, just wrap excerpt text in P on your archive/index templates.</p>
<p><strong>WARNING:</strong> This hack is completely untested. I take absolutely no responsibility if using it completely wipes your blog off the face of Google. You&#8217;ve been warned.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/07/03/experiment-with-description-snippets.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Mosaic Rambling: Kawasaki, FURL, Fried CPU</title>
		<link>http://seo4fun.com/blog/2006/07/03/mosaic-rambling-kawasaki-furl-fried-cpu.html</link>
		<comments>http://seo4fun.com/blog/2006/07/03/mosaic-rambling-kawasaki-furl-fried-cpu.html#comments</comments>
		<pubDate>Mon, 03 Jul 2006 14:12:50 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/07/03/mosaic-rambling-kawasaki-furl-fried-cpu.html</guid>
		<description><![CDATA[I just finished watching Guy Kawasaki speak. Nearly 40 minutes, but man.. do I feel inspired. &#8220;Don&#8217;t listen to the bozos!&#8221;
I&#8217;m seeing FURL redirect pages ranking high on Google. A double edged sword: More SE traffic, but original pages seem to be getting knocked off the index.
It&#8217;s a long story, but I fried my CPU [...]]]></description>
			<content:encoded><![CDATA[<p>I just finished watching <a href="http://blog.guykawasaki.com/2006/06/the_art_of_the_.html">Guy Kawasaki</a> speak. Nearly 40 minutes, but man.. do I feel inspired. &#8220;Don&#8217;t listen to the bozos!&#8221;</p>
<p>I&#8217;m seeing FURL redirect pages ranking high on Google. A double edged sword: More SE traffic, but original pages seem to be getting knocked off the index.</p>
<p>It&#8217;s a long story, but I fried my CPU a few nights ago, so I&#8217;m fish out of water on a 1GHz / 256MB RAM Windows ME without access to Forum passwords and whatnot. Few lessons I&#8217;ve learned:</p>
<ul>
<li>Don&#8217;t screw around with your main rig. Processors and mobos are cheap these days (found an 2.2Ghz AMD 64 for $100 bucks).</li>
<li>Don&#8217;t use Windows Backup to backup files. That program compresses everything into one huge file, and you can&#8217;t get to individual files from another machine.</li>
<li>Google Notebook, Google Reader, Google Calendar, gmail - when a good chunk of my data is online, one fried machine won&#8217;t slow me down.</li>
<li>Last but not least - don&#8217;t jam anything into the CPU fan.</li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/07/03/mosaic-rambling-kawasaki-furl-fried-cpu.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>SEO 101: Feeding Google Juicy Description Snippets</title>
		<link>http://seo4fun.com/blog/2006/06/24/seo-101-feeding-google-juicy-description-snippets.html</link>
		<comments>http://seo4fun.com/blog/2006/06/24/seo-101-feeding-google-juicy-description-snippets.html#comments</comments>
		<pubDate>Sat, 24 Jun 2006 13:40:14 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[On Page Optimization]]></category>

		<category><![CDATA[SEO]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/06/24/seo-101-feeding-google-juicy-description-snippets.html</guid>
		<description><![CDATA[How can you get Google to pick up the description snippet you want? And why should you care? If you use META descriptions, good. If not, read on. My views on this isn&#8217;t authoritative but if you disagree, post a good counterexample.

Don&#8217;t use H tags just to make text look bigger. Enuf said.
Avoid wrapping text [...]]]></description>
			<content:encoded><![CDATA[<p>How can you get Google to pick up the description snippet you want? And why should you care? If you use META descriptions, good. If not, read on. My views on this isn&#8217;t authoritative but if you disagree, post <strong>a good counterexample</strong>.</p>
<ul>
<li><strong>Don&#8217;t use H tags just to make text look bigger.</strong> Enuf said.</li>
<li><strong>Avoid wrapping text in HREF if you want it snippetized.</strong> Google skips over H1 and H2 on <a href="http://64.233.187.104/search?q=cache:CqDrxoI5a6cJ:horrorbrain.ign.com/+site:ign.com&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=8&#038;client=firefox-a">this page</a> (snippet:&#8221;Horrorbrain.com: Movies, Shows, Reviews and More. &#8230; Today&#8217;s Headlines.  From Beyond: Frightmare · Fan Fiction: Up To Snuff · Slasher Friday: Katiebird*&#8221;), including a chunk of text wrapped in HREF: &#8220;Horror Brain is very proud to present a wonderful screenplay written by Aaron Boehm.&#8221; Another example: On <a href="http://216.239.51.104/search?q=cache:_uez4MJnrH8J:www.philosophypages.com/locke/g00.htm+site:philosophypages.com&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=20&#038;client=firefox-a">this page</a> (snippet:&#8221;<font size="-1">A guide to Locke&#8217;s Essay. &#8230; Introduction. John Locke&#8217;s An Essay Concerning  Human Understanding is a classic statement of empiricist epistemology.</font>&#8220;), the first H tag wrapped in HREF - &#8220;A Guide to Locke&#8217;s Essay&#8221;is ignored.</li>
<li><strong>Keep H and P together.</strong> Don&#8217;t insert HR between them, or wrap them individually in DIV. If there’s a huge chunk of text on a page, will Google always find it, even if its buried under MENUs, TABLEs and DIVS? No. <a href="http://64.233.187.104/search?q=cache:mNSuEBb4unUJ:www.nhc.noaa.gov/aboutnames_text.html+site:nhc.noaa.gov&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=88&#038;client=firefox-a">Here’s an example</a> (snippet: &#8220;<font size="-1">TPC&#8217;S List of World-wide Tropical Cyclone Names. <strong>&#8230;</strong> Atlantic Names: 2002 2003  2004 2005 2006 2007 Arthur Ana Alex Arlene Alberto Andrea Bertha Bill Bonnie&#8221;) </font>where Google doesn’t. Google starts snippetizing text right under H2. On <a href="http://64.233.179.104/search?q=cache:_bddtG89JIoJ:www.themystica.com/mystica/writings/the_journey.html+site:themystica.com&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=20&#038;client=firefox-a">this page</a> (snippet: &#8220;<font size="-1">by David Striar Years ago I found myself struggling with the question of what I  should do with my life. While I was drawn to the arts, especially to music&#8221;)</font>, the text “The journey by David Striar” gets chopped in half because of an HR between “The Journey” and “by David Briar.&#8221; Google ignores a few leading P text with A HREF embedded and chooses text from a bigger P on <a href="http://64.233.179.104/search?q=cache:XGYhhCezPPYJ:www.themystica.com/mystica/writings/curious.html+site:themystica.com&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=17&#038;client=firefox-a">this page</a> (&#8221;<font size="-1">I rarely search for companions that understand my unearthly ways, but I do know  that you are out there, and that I alone do not possess the only key to the&#8221;)</font>, similar to the previous page. The only difference between the two pages is that on this page, the heading text is wrapped in its own P, and Google skips them because they&#8217;re too short. <em>This tells me P is a separator just like DIV.</em></li>
<li><strong>Use H on every page</strong> to tell Google where content begins (as well as what the page is about). SEO 101 sometimes fancy lookin&#8217; sites seem to ignore.</li>
<li><strong>Lots of TABLES suck.</strong> <a href="http://64.233.179.104/search?q=cache:1JsWUiWYCP0J:www.origin.fema.gov/+site:fema.gov&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=17&#038;client=firefox-a">Here’s an example</a>  (snippet: &#8220;<font size="-1">Skip navigations, DHS Seal, FEMA, Background image - Flag, Background image -  DHS seal, Background image - DHS seal. Disaster tab · Emergency tab&#8221;) </font>of a pretty densely structured page with lots of TABLE elements. Google can’t find a good snippet so it defaults to indexing ALT text.</li>
<li><strong>Use CSS to move content up in the source. </strong>Another obvious point but some people believe Google has no problem finding text on a page. Answer is it depends on your page layout. Check out <a href="http://64.233.179.104/search?q=cache:fGN4IY8ohNwJ:www.gismaps.fema.gov/+site:fema.gov&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=19&#038;client=firefox-a">this fema page</a> (snippet: &#8220;<font size="-1">Sorry but your browser does not support JavaScript. Please download the lastest  version of your browser. This JavaScript controls primary navigation &#8230;&#8221;)</font>where Google snippetizes NOSCRIPT text instead of the content. There’s only one H tag, and its followed by paragraphs of text. Content is inside a TABLE though and is buried deep on the bottom of source.On <a href="http://216.239.51.104/search?q=cache:CJbGRkoSyeEJ:www.floridadisaster.org/about_the_division.htm+site:floridadisaster.org&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=6&#038;client=firefox-a">this page</a>, (&#8221;<font size="-1">About The Division Of Emerergency Management. <strong>&#8230;</strong> State Emergency ResponseTeam.  Prepare and Stay Aware! FL Hazard Lookup, NATURAL, - Hurricane, - Lightning &#8230;&#8221;)</font>IMG ALT and text in SELECT listbox in the first TABLE gets picked up; Google doesn’t even get to the content text. There’s a META description, but its not long enough for Google.</li>
<li><strong>Use lengthy META description tags</strong> (over 50 chars long). If Google finds that your META description is too short, it’ll scrape text off your page. This spells trouble if Google ends up snippetizing your nav links instead of content. Ideally, you want to design your page so that it doesn’t even require a META description tag.</li>
<li><strong>HTML Validate</strong> to get rid of nesting problems and other major errors. Replacing XHTML declaration with HTML Transitional for example seems to wack out Google&#8217;s parser, and ill formed TABLE may also cause problems. For example, Google indexes nav text <a href="http://216.239.51.104/search?q=cache:zV6BOtmlpq4J:www.fordham.edu/Alumni_Relations/+site:fordham.edu&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=17&#038;client=firefox-a">here</a> instead of content probably because the content text is wrapped in TD and is missing a TABLE tag. IMG alt <a href="http://64.233.179.104/search?q=cache:WULdC4z49UEJ:www.artchive.com/gallries.htm+site:artchive.com&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=20&#038;client=firefox-a">picked up here</a>. The first IMG alt is ignored because HTML is broken.</li>
<li><strong>Write full paragraphs.</strong> Google prefers full sentences and paragraphs to link text or short phrases. Google defaults to snippetizing the MENU links <a href="http://216.239.51.104/search?q=cache:wwARiIOSzzsJ:www.nhc.noaa.gov/1998MITCHadv_text.html+site:nhc.noaa.gov&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=56&#038;client=firefox-a">on this page</a>, because it doesn’t find any good chunk of text. On <a href="http://216.239.51.104/search?q=cache:oUncFu8AUYQJ:www.nhc.noaa.gov/aboutsshs_text.html+site:nhc.noaa.gov&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=45&#038;client=firefox-a">a similar page</a> with identical structure, Google correctly skips over the MENU links. <a href="http://64.233.187.104/search?q=cache:5UcrzbPvnUYJ:seo4fun.com/blog/contact-me/+site:seo4fun.com&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=17&#038;client=firefox-a">Here</a>, I used to have a short sentence and Google used to snippetize my blogroll, but after beefing up my text, Google indexed the right text.</li>
<li><strong>Wrap navigation links in a DIV, avoid any use of H tags in that section of code, and get rid of unlinked text.</strong> Not bulletproof, but worth a try.</li>
<li><strong>Don’t add ALT text for images if the image isn’t page specific.</strong> Use ALT=&#8221;" instead. This is especially if your page is ill-formed (i.e. without an obvious content starting point), which increases the chance of meaningless ALT text for logos or whatever getting snippetized.</li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/06/24/seo-101-feeding-google-juicy-description-snippets.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Corrupt Titles in Google SERP</title>
		<link>http://seo4fun.com/blog/2006/06/20/corrupt-titles-in-google-serp.html</link>
		<comments>http://seo4fun.com/blog/2006/06/20/corrupt-titles-in-google-serp.html#comments</comments>
		<pubDate>Wed, 21 Jun 2006 00:04:51 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/06/20/corrupt-titles-in-google-serp.html</guid>
		<description><![CDATA[I&#8217;ve been keeping track of a WMW thread about Google displaying corrupt titles in their new SERP. Basically, Google is tagging on on-page text snippets to the end of the TITLE tag (or element, whatever). Since similar TITLE/description snippets throw a duplicate content flag, this little bug may end up causing major problems for some [...]]]></description>
			<content:encoded><![CDATA[<p>I&#8217;ve been keeping track of a <a href="http://www.webmasterworld.com/forum30/34822.htm">WMW thread</a> about Google displaying corrupt titles in their new SERP. Basically, Google is tagging on on-page text snippets to the end of the TITLE tag (or element, whatever). Since similar TITLE/description snippets throw a duplicate content flag, this little bug may end up causing major problems for some sites. In fact, at least one WMW posters claimed thousands of his pages with corrupt titles are now marked supplemental in Google&#8217;s index.</p>
<p>As of today, <a href="http://www.google.com/search?hl=en&#038;lr=&#038;safe=off&#038;q=site%3Abrooke-skye-videos.blogspot.com&#038;btnG=Search">the problem still hasn&#8217;t gone away</a>. Notice the &#8220;notify blogger about objectionable&#8230;&#8221; text at the end of my titles. Also notice (though this is somewhat off-topic) that due to that stupid blogger javascript nav on the top of source code  (and also probably due to me switching the XHTML declaration with HTML transitional), every disription snippet for that blog is identical, throwing 99% of my blog into the supplemental index.</p>
<p>First question - is this intentional or is it a bug? Well, menu link text tagged on to the end of my titles don&#8217;t look all that pretty, and I really can&#8217;t think of any reason why Google would do this intentionally.</p>
<p>Second question - what&#8217;s causing it and how do I prevent Google from screwing up my listing? Well, on one of the pages, the cache date says May 30, 2006 03:24:19 GMT. But interestingly, another page listed correctly is showing the same cache date: May 30, 2006 03:24:18 GMT.. What&#8217;s the difference between the two pages?</p>
<p>First off, in both cases, Google is completely ignoring my META description - assuming the cached page I see corresponds to the SERP snippet (which isn&#8217;t always the case). Both pages have HTML validation errors. Meta description I used: &lt;meta name=&#8221;description&#8221; content=&#8221;dfdfsd&#8221; <strong>/&gt;</strong> on a HTML transitional. I dunno - both pages look pretty much the same to me.</p>
<p>All I can say at this point is if you see a page listed incorrectly, validate it. My guess is Google rewrote their parser used to build description snippets, and the new parser is parsing some pages incorrectly. Validating your pages should minimize the chance of that happening. &#8216;Course if Google is choking on a perfectly HTML valid page, then we&#8217;ll just have to sit it out.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/06/20/corrupt-titles-in-google-serp.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>How Do You Bait Without Looking Stupid?</title>
		<link>http://seo4fun.com/blog/2006/06/08/how-do-you-bait-without-looking-stupid.html</link>
		<comments>http://seo4fun.com/blog/2006/06/08/how-do-you-bait-without-looking-stupid.html#comments</comments>
		<pubDate>Thu, 08 Jun 2006 22:30:25 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/06/08/how-do-you-bait-without-looking-stupid.html</guid>
		<description><![CDATA[Dan Thies wrote a number last week titled Link Building: Are Bloggers a Bunch of Morons?, a title that got one of his foreign readers a little riled up. The intent of the title is plain as grey, but does it get the job done? Don&#8217;t cry wolf unless there&#8217;s really a wolf, and don&#8217;t [...]]]></description>
			<content:encoded><![CDATA[<p>Dan Thies wrote a number last week titled <a href="http://www.seoresearchlabs.com/blog/2006/06/link-building-are-bloggers-bunch-of.html">Link Building: Are Bloggers a Bunch of Morons?</a>, a title that got one of his foreign readers a little riled up. The intent of the title is plain as grey, but does it get the job done? <strong>Don&#8217;t cry wolf unless there&#8217;s really a wolf, and don&#8217;t fret over a wrinkle on your forehead.</strong> <a href="http://www.google.com/search?q=shoemoney+plentyoffish&#038;hl=en&#038;lr=&#038;client=firefox-a&#038;rls=org.mozilla:en-US:official&#038;pws=0&#038;pwst=1&#038;start=20&#038;sa=N">In Shoemoney&#8217;s case</a> (a post that got at least 557 sites buzzing), the wolf was plentyoffish. At the end of the day, his post got more people questioning his own credibility than Markus Frind (one guy at Threadwatch asked: &#8220;Who is Shoemoney anyway?&#8221;, to which SEObook jokingly replied: &#8220;I don&#8217;t know Shoemoney. He is one of DaveN&#8217;s friends.&#8221;). In Dan&#8217;s case, the wolf is <em>natural links of pure hate</em>. What&#8217;s that?</p>
<blockquote><p>When a blogger cleverly links to Microsoft&#8217;s web site with words like &#8220;evil empire&#8221; in the text of the link, they think they&#8217;re sticking it to the man, but they&#8217;re actually passing link love to the man.</p></blockquote>
<p>Is that really a wolf? Or is it a rat in my girlfriend&#8217;s closet? If the point of Dan&#8217;s post is linking to sites you hate is dumb - I&#8217;m not sure what else to take away from it &#8212; isn&#8217;t that like telling a 9-ball A player to chalk up after every shot?</p>
<p>When you shine a spotlight on yourself, deliver a good rendition of <a href="http://www.google.com/url?sa=t&#038;ct=res&#038;cd=1&#038;url=http%3A%2F%2Fwww.brave.com%2Fbo%2Flyrics%2Fstairhea.htm&#038;ei=THWIROLfGaLUqAKjzrC6DA&#038;sig2=IR2Vue4mkTbaiTjhaBTp3A">Stairway to Heaven</a>, not a &#8220;<a href="http://www.comedy-zone.net/jokes/laugh/insults/insult9.htm">yo momma was so ugly</a>, she looked out the window and got arrested for mooning.&#8221;</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/06/08/how-do-you-bait-without-looking-stupid.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Co-op Annotations for Dummies</title>
		<link>http://seo4fun.com/blog/2006/05/27/google-co-op-annotations-for-dummies.html</link>
		<comments>http://seo4fun.com/blog/2006/05/27/google-co-op-annotations-for-dummies.html#comments</comments>
		<pubDate>Sat, 27 May 2006 17:01:33 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/27/google-co-op-annotations-for-dummies.html</guid>
		<description><![CDATA[Just think of annotations like a del.icio.us or FURL save. Every annotation then consists of URL you want to save, and tags. For example, if I wanted to save cnn.com, the data I&#8217;d be  saving using del.icio.us would look like this:

URL: http://www.cnn.com/
tags: news, liberal-bullshit

All you have to do is translate that into Google Co-op [...]]]></description>
			<content:encoded><![CDATA[<p>Just think of annotations like a del.icio.us or FURL save. Every annotation then consists of URL you want to save, and tags. For example, if I wanted to save cnn.com, the data I&#8217;d be  saving using del.icio.us would look like this:</p>
<ul>
<li>URL: http://www.cnn.com/</li>
<li>tags: news, liberal-bullshit</li>
</ul>
<p>All you have to do is translate that into Google Co-op speak:</p>
<p>&lt;Annotation about=&#8221;http://www.cnn.com/&#8221;&gt;<br />
&lt;Label name=&#8221;news&#8221;&gt;&lt;/Label&gt;<br />
&lt;Label name=&#8221;liberal_bullshit&#8221;&gt;&lt;/Label&gt;<br />
&lt;/Annotation&gt;</p>
<p>You also have the option to specify all urls under a specific directory, or all urls that match a number of keywords, instead of annotating a site page by page. You can also add additional information, like comments and score (only one url per domain will show up per query, so score helps to clarify which URL you prefer). And of course, for annotations to work, they need context. I&#8217;ll stop here though, since I need to grab some breakfast, plus I want to stay true to the title of this post.</p>
<p>Google Co-op Annotation references: <a href="http://www.google.com/coop/docs/guide_topics.html">Google Co-op Topics Developers Guide</a></p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/27/google-co-op-annotations-for-dummies.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Sitewide Duplicate Content Checker Tool Useless?</title>
		<link>http://seo4fun.com/blog/2006/05/27/sitewide-duplicate-content-checker-tool-useless.html</link>
		<comments>http://seo4fun.com/blog/2006/05/27/sitewide-duplicate-content-checker-tool-useless.html#comments</comments>
		<pubDate>Sat, 27 May 2006 15:38:50 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Duplicate Content]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/27/sitewide-duplicate-content-checker-tool-useless.html</guid>
		<description><![CDATA[SEO Junkie just released an clientside application you can use to check for duplicate content on your site. I found a link to his app through Search Engine Watch yesterday while I was playing around with Google Co-op. The application is meant for small sites and he warns that there&#8217;s alot of functionality missing.  [...]]]></description>
			<content:encoded><![CDATA[<p><a href="http://www.seojunkie.com/2006/05/24/site-wide-duplicate-content-analyzer/">SEO Junkie just released an clientside application</a> you can use to check for duplicate content on your site. I found a link to his app through Search Engine Watch yesterday while I was playing around with Google Co-op. The application is meant for small sites and he warns that there&#8217;s alot of functionality missing.  When he said &#8220;small sites&#8221; I hoped he meant less than 10,000 pages, but in the end I walked away scratching my head.</p>
<p>First, even when checking just two urls, its incredibly slow, compared to something like the <a href="http://www.webconfs.com/similar-page-checker.php">Similar Pages Checker</a>. Second, it indiscriminantly crawls every URL it finds: affiliate redirect links, links to images, videos, etc. It took me a few minutes to figure out why I was suddenly getting hit with a barrage of pop-ups. Third, it crapped out at the middle of crawling my site with an error message:  Runtime Error: Method &#8216;~&#8217; of object &#8216;~&#8217; failed. I went over to his blog and saw similar error messages being reported. It could be due to hitting a memory cap or something, but why not just elegantly stop crawling links when you get above a certain threshhold?</p>
<p>My biggest complaint though, is the fact that you can&#8217;t use this to check large sites. Those are exactly the kind of sites I&#8217;d want to check for page similarity. I mean, why would I want to run my 30 page blog through something like this?</p>
<p>Suggestions:</p>
<ul>
<li>Reference robots.txt and/or check robots tag and ignore disallowed/noindex pages.</li>
<li>Do not follow redirect urls that lead outside of a given domain.</li>
<li>Automatically limit number of pages crawled to prevent the program from crashing.</li>
<li>Ignore image, audio, video files (or does it already ignore them?)</li>
<li>Give users an option to compare urls in a subdirectory instead of comparing each page to every other page in a domain.</li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/27/sitewide-duplicate-content-checker-tool-useless.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Sitemap Reporting 404s under Summary</title>
		<link>http://seo4fun.com/blog/2006/05/27/google-sitemap-reporting-404s-under-summary.html</link>
		<comments>http://seo4fun.com/blog/2006/05/27/google-sitemap-reporting-404s-under-summary.html#comments</comments>
		<pubDate>Sat, 27 May 2006 12:06:06 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[Google Sitemaps]]></category>

		<category><![CDATA[Googlebot]]></category>

		<category><![CDATA[SEO]]></category>

		<category><![CDATA[Supplemental Bot]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/27/google-sitemap-reporting-404s-under-summary.html</guid>
		<description><![CDATA[Note: This post is a follow up to my earlier post, Googlebot Refreshing Supplementals.
I noticed this morning that Google Sitemap is reporting 404s on the Summary page. The pages Google Sitemap reports missing include some of the pages I&#8217;ve been getting emails for since May 20th.

Not all the urls showed up under HTTP errors, just [...]]]></description>
			<content:encoded><![CDATA[<p><strong>Note:</strong> This post is a follow up to my earlier post, <a href="/blog/2006/05/25/googlebot-refreshing-supplementals.html">Googlebot Refreshing Supplementals</a>.</p>
<p>I noticed this morning that Google Sitemap is reporting 404s on the <strong>Summary page</strong>. The pages Google Sitemap reports missing include some of the pages I&#8217;ve been getting emails for since May 20th.<br />
<img title="Google Sitempa 404 Summary" alt="Google Sitempa 404 Summary" src="/images/google-sitemap-404.jpg" /></p>
<p>Not all the urls showed up under <strong>HTTP errors</strong>, just around 11 of them.</p>
<p>As I said earlier, the only interesting thing about these 404 pages are they no longer have any incoming links, and they exist only in Google&#8217;s supplemental index. So, I&#8217;m kinda hoping Google is in the process of refreshing their supplemental index by checking urls in their database, to see if they changed, 404/410, or 301 redirect to some other url.</p>
<p>In fact, <a href="http://www.mattcutts.com/blog/indexing-timeline">Matt Cutts has previously posted on his blog</a> (520 freaking comments, longer than any forum thread I ever seen, never mind a blog post) that a supplemental refresh has been going on since before April: &#8220;In early April, we started showing some refreshed supplemental results to users.&#8221; He also said the index/crawl team&#8217;s turning their focus to refreshing supplementals: &#8220;Well, now that Bigdaddy is done, we’ve turned our focus to <strong>refreshing our supplemental results.</strong>&#8221; Another quote (this one I remember reading the day he posted it), which suggests we have a glimmer of hope to have all of our supplementals refreshed by September: &#8220;I believe that folks here intend to refresh all of the supplemental results over the summer months, although I’m not 100% sure.&#8221;</p>
<p>Well&#8230;this is what I like to know: will lack of high quality/relevant incoming links or crappy outgoing links put a site at the back of the bus during the big summar supplemental refresh, or what?</p>
<p><strong>P.S.</strong> Anyway, assuming Google will refresh your supplemental pages by the end of the summer, if you or your client is having problems with supplemental pages, this is a good time as any to make sure you don&#8217;t end up with a bunch of new supplemental pages, and wait another year to have them refreshed again. Beef up those product pages, make sure your title/description metas are unique across your entire site, tighten up your dynamic url handling, etc.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/27/google-sitemap-reporting-404s-under-summary.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Turning Blog RSS Feed into Google Co-Op Subscribed Links</title>
		<link>http://seo4fun.com/blog/2006/05/26/turning-blog-rss-feed-into-google-co-op-subscribed-links.html</link>
		<comments>http://seo4fun.com/blog/2006/05/26/turning-blog-rss-feed-into-google-co-op-subscribed-links.html#comments</comments>
		<pubDate>Fri, 26 May 2006 13:54:45 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/26/turning-blog-rss-feed-into-google-co-op-subscribed-links.html</guid>
		<description><![CDATA[If you&#8217;re looking for a way to submit your blog to Google co-op, I just wrote a script that you can use. It makes your blog posts appear on top of Google SERPs, using categories under which you submitted your post. For example, if a post was submitted under &#8220;SEO&#8221; or &#8220;Google coop&#8221;, it&#8217;ll show [...]]]></description>
			<content:encoded><![CDATA[<p>If you&#8217;re looking for a way to submit your blog to Google co-op, I just wrote a script that you can use. It makes your blog posts appear on top of Google SERPs, using categories under which you submitted your post. For example, if a post was submitted under &#8220;SEO&#8221; or &#8220;Google coop&#8221;, it&#8217;ll show up for those keyword searches. Assuming Google recrawls your submitted link occassionally (it seems to recrawl every few hours for me), the results will refresh automatically as you publish more posts.</p>
<p>Just a warning: this is a BETA script I wrote an hour after I woke up without coffee. I&#8217;m also not convinced submitting your blog posts to Google co-op is really productive, but no harm done until somebody loses an eye.</p>
<p>Anyway, back to my script. All you have to do is submit a link to Google Co-op using this format:</p>
<p><b>http://seo4fun.com/php/rss-parser.php?url=YOURRSSFEEDURL</b></p>
<p>The URL you submit must be RSS 2.0. The source is written in PHP 5, in case you&#8217;re wondering. There&#8217;s a ton of features missing (e.g. I&#8217;d want it to scan for technorati tags embedded in posts, though that requires full feeds) and it ignores some guidelines (80 chars per text1 line, no http:// in url, etc). Let me know if you&#8217;re interested in looking at the source code. <a href="http://www.google.com/coop/manage/subscribedlinks">Submit your subscribed links here</a></p>
<p><b>Updates:</b></p>
<ul>
<li>The script might timeout if your feed is huge, but I doubt it.</li>
<li>posts with tabs/carriage returns (\t\n) in description fields weren&#8217;t getting parsed [fixed].</li>
<li>Some feeds like seomoz.org that lacks categories won&#8217;t work.</li>
<li>If your blog doesn&#8217;t run on Wordpress, chances are my script will choke on your feed.</li>
<li>If your post urls use a &#8216;?&#8217;, it won&#8217;t parse [Fixed].</li>
<li>Barfed on titles with commas [fixed].</li>
</ul>
<p> <b>Blogs I&#8217;ve tested:</b></p>
<ul>
<li>First blog to go under the knife: <a href="/php/rss-parser.php?url=http://seo4fun.com/blog/feed/" rel="nofollow">My SEO notebook</a>.</li>
<li>SEO By the Sea (bulky category names but it seems to work): <a href="/php/rss-parser.php?url=http://www.seobythesea.com/?feed=rss2" rel="nofollow">Google Co-op XML</a>. Searches: (design, search engines and directories, usability, search engine optimization, culture, internet advertising)</li>
<li>SEOMoz (doesn&#8217;t work, found no categories).</li>
<li><a href="/php/rss-parser.php?url=http://www.mattcutts.com/blog/feed/" rel="nofollow">Matt Cutt&#8217;s Blog</a> works, but his categories are wacked (i.e. &#8220;weblog/blog&#8221;). searches: (&#8221;weblog blog&#8221;,&#8221;google seo&#8221;, &#8220;personal&#8221;, &#8220;productivity&#8221;, &#8220;web net&#8221;,&#8221;how to&#8221;&#8230;)</li>
<li><a href="/php/rss-parser.php?url=http://www.seobuzzbox.com/feed/" rel="nofollow">SEOBuzzbox</a> uses &lt;dc:subject&gt; instead of &lt;category&gt; [Fixed] Searches: (marketing news, dmoz, blackhat, matt cutts, v7ndotcom elursrebmem, internet marketing, google news&#8230;)</li>
<li><a href="/php/rss-parser.php?url=http://www.jimboykin.com/feed/" rel="nofollow">Jim Boykin&#8217;s Blog</a> seems to load without a hitch. Searches: (places, jim&#8217;s crazy ideas, yahoo, &#8220;CEO, not SEO stuff&#8221;, jim boykin, google, internet marketing tips, seo research, link building&#8230;)</li>
<li><a href="/php/rss-parser.php?url=http://damnjezebel.com/diary/?feed=rss2" rel="nofollow">Damn Jezebel&#8217;s Dairy</a> searches: (conversations, randomness, bitching, what the fuck?)</li>
</ul>
<p><a href="http://www.google.com/coop/profile?user=017275478576632624117" rel="nofollow">Subscribe to my profile</a> instead of submitting those feeds individually, and run a few Google searches to see what happens.</p>
<p> <b> Notes</b></p>
<ul>
<li>I see every damn RSS has its nuances, so it&#8217;ll take a while before this script can handle 50% of the blogs out there.</li>
<li>What&#8217;s the point of subscribing to multiple blog feeds with Google Co-op? (I luv my Google Reader). SER type subscribed link display is cool, but individual blogs? Plus you need to remember categories to pull up results. Not practical.</li>
<li>Only one match per account. Say I&#8217;m subscribed to 10 links, and 5 of them match a search for &#8220;yahoo.&#8221; Only one will show in the Onebox. There&#8217;s also no ResultSpec timestamp element, so there&#8217;s no way to make the most recent item to show. Also, subscriptions complete for the same keywords (e.g. for Google, digg/sew/and ser compete for the same onebox space), something that would steer me away from running a generic search like &#8220;seo.&#8221;</li>
<li>Some blogs use long tail categories or throw several words under one category (i.e. weblog/blog). I prefer using data objects to handle synonyms.</li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/26/turning-blog-rss-feed-into-google-co-op-subscribed-links.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Googlebot Refreshing Supplementals?</title>
		<link>http://seo4fun.com/blog/2006/05/25/googlebot-refreshing-supplementals.html</link>
		<comments>http://seo4fun.com/blog/2006/05/25/googlebot-refreshing-supplementals.html#comments</comments>
		<pubDate>Thu, 25 May 2006 10:27:18 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[Googlebot]]></category>

		<category><![CDATA[Supplemental Bot]]></category>

		<category><![CDATA[Supplementals]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/25/googlebot-refreshing-supplementals.html</guid>
		<description><![CDATA[Weird 404 error email I sent myself yesterday (url hidden to prevent linking to an adult domain):
HTTP_REFERER: [blank] HTTP_HOST: www.domain.com PHP_SELF: /fgdfgfert4534.html REQUEST_URI /NONEXISTENTURL.html REMOTE_ADDR: 66.249.65.69 TIMESTAMP: 5/24/2006 9:15 PM
Quick explanation: I rigged my dynamic pages, so a request to retrieve &#8220;maroon-widget.html&#8221; 404s and triggers an email if I don&#8217;t have &#8220;maroon widget&#8221; in my [...]]]></description>
			<content:encoded><![CDATA[<p>Weird 404 error email I sent myself yesterday (url hidden to prevent linking to an adult domain):</p>
<blockquote><p>HTTP_REFERER: [blank] HTTP_HOST: www.domain.com PHP_SELF: /fgdfgfert4534.html REQUEST_URI /NONEXISTENTURL.html REMOTE_ADDR: 66.249.65.69 TIMESTAMP: 5/24/2006 9:15 PM</p></blockquote>
<p><strong>Quick explanation:</strong> I rigged my dynamic pages, so a request to retrieve &#8220;maroon-widget.html&#8221; 404s and triggers an email if I don&#8217;t have &#8220;maroon widget&#8221; in my database.</p>
<p>REQUEST_URI is linked from nowhere; it exists solely in the supplemental index. I&#8217;ve seen Yahoo do this kinda thing, but this week I&#8217;m starting to see Google do the same thing. I guess Google&#8217;s basically crawling my site using its own database instead of following links. Is this a common behavior/part of a normal crawl, or is Google trying to clean up supplementals?</p>
<p>Looking up 66.249.65.69 in Google returns 208,000 jibberish results (mostly those that display your IP on their page). So I guess its just a regular bot, not supplemental bot?</p>
<p>Update: After cleaning out my inbox, I found similar emails going back to May 20.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/25/googlebot-refreshing-supplementals.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Coop Blog RSS to XML Generator</title>
		<link>http://seo4fun.com/blog/2006/05/24/google-coop-blog-rss-to-xml-generator.html</link>
		<comments>http://seo4fun.com/blog/2006/05/24/google-coop-blog-rss-to-xml-generator.html#comments</comments>
		<pubDate>Wed, 24 May 2006 22:20:46 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/24/google-coop-blog-rss-to-xml-generator.html</guid>
		<description><![CDATA[UPDATE: Read this follow-up post to learn how to feed your Blog RSS to Google Co-op.
Scanning through Google Co-op Group, I came across a Google Co-op XML generator by 1000apps. Unfortunately, it uses words from the title instead of blog categories to generate the XML. So I&#8217;m going to have to rewrite it. This is [...]]]></description>
			<content:encoded><![CDATA[<p><strong>UPDATE:</strong> Read <a href="/blog/2006/05/26/turning-blog-rss-feed-into-google-co-op-subscribed-links.html">this follow-up post</a> to learn how to feed your Blog RSS to Google Co-op.</p>
<p>Scanning through <a href="http://groups.google.com/group/google-co-op">Google Co-op Group</a>, I came across a <a href="http://www.seo-search.net/rsstocoop/">Google Co-op XML generator by 1000apps</a>. Unfortunately, it uses words from the title instead of blog categories to generate the XML. So I&#8217;m going to have to rewrite it. This is the outline of my code (noting for myself, so I don&#8217;t have to think about it later):</p>
<ol>
<li>Parse Blog RSS using regexp; create a class for individual pages, store title, url, and categories per page. Store those classes in an array.</li>
<li>Foreach item in the array, loop through categories. For each category, generate a Google co-op XML entry.</li>
<li>write to output file, using the blog title as part of the url to avoid dup urls.</li>
<li>done</li>
</ol>
<p>Google XML entry structure:</p>
<p>&lt;ResultSpec id=&#8221;CATCOUNT - TITLE&#8221;&gt;<br />
&lt;Query>CATEGORYNAME&lt;/Query&gt;<br />
&lt;Response&gt;<br />
&lt;Output name=&#8221;title&#8221;>POSTORBLOGTITLE&lt;/Output&gt;<br />
&lt;Output name=&#8221;more_url&#8221;>URL&lt;/Output&gt;<br />
&lt;Output name=&#8221;text1&#8243;>DESCRIPTIONSNIPPET&lt;/Output&gt;<br />
&lt;/Response&gt;<br />
&lt;/ResultSpec&gt;</p>
<p>Oh yeah, bracket the page with &lt;Results&gt;, &lt;/Results&gt;<br />
Note to self: Make sure I blog robots from the dir I dump the XML files.</p>
<p>It is kind of cool to show up at the top of Google search, even though only I can see it. If I do get around to rewriting that script, I&#8217;ll link to it from this blog.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/24/google-coop-blog-rss-to-xml-generator.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>My Next Obsession - Getting a Site Back in the Index</title>
		<link>http://seo4fun.com/blog/2006/05/23/my-next-obsession-getting-a-site-back-in-the-index.html</link>
		<comments>http://seo4fun.com/blog/2006/05/23/my-next-obsession-getting-a-site-back-in-the-index.html#comments</comments>
		<pubDate>Tue, 23 May 2006 13:24:22 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Adult Webmastering]]></category>

		<category><![CDATA[Google]]></category>

		<category><![CDATA[SEO]]></category>

		<category><![CDATA[Supplementals]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/23/my-next-obsession-getting-a-site-back-in-the-index.html</guid>
		<description><![CDATA[For the last two months, I&#8217;ve been maintaining a holding pattern with one of my adult sites, to no avail. The general concensus was that something was up at Google and I shouldn&#8217;t do anything drastic. But now I&#8217;m going to start working on my site again. Till I see some progress, how to get [...]]]></description>
			<content:encoded><![CDATA[<p>For the last two months, I&#8217;ve been maintaining a holding pattern with one of my adult sites, to no avail. The general concensus was that something was up at Google and I shouldn&#8217;t do anything drastic. But now I&#8217;m going to start working on my site again. Till I see some progress, how to get a screwed up site reindexed in Google will be the primary focus of my blog.</p>
<p>List of Todos:</p>
<ul>
<li>Write more articles.</li>
<li>Take advantage of technorati and other sidekicks (though the site isn&#8217;t a blog).</li>
<li>Install a blog if necessary.</li>
<li>Optimize for MSN.</li>
<li>Streamline internal linking structure.</li>
<li>Beef up product pages.</li>
<li>Keep track of cache dates.</li>
<li>Periodically, refresh content (the site&#8217;s dynamically generated, but uses static elements).</li>
<li>Linkbait by writing aggressivly and getting it noticed by people I can piss off.</li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/23/my-next-obsession-getting-a-site-back-in-the-index.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>What I&#8217;ve been Up To</title>
		<link>http://seo4fun.com/blog/2006/05/10/what-ive-been-up-to.html</link>
		<comments>http://seo4fun.com/blog/2006/05/10/what-ive-been-up-to.html#comments</comments>
		<pubDate>Wed, 10 May 2006 12:52:44 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/10/what-ive-been-up-to.html</guid>
		<description><![CDATA[I surfed over to Dan Thies&#8217; blog yesterday and noticed he&#8217;s been slacking off on his blog, like I have with this one. Matt&#8217;s also been absent for a couple of days, though now he seems to be back blogging in full force. I&#8217;ve started a couple of new blogs this week, one blog for [...]]]></description>
			<content:encoded><![CDATA[<p>I surfed over to Dan Thies&#8217; blog yesterday and noticed he&#8217;s been slacking off on his blog, like I have with this one. Matt&#8217;s also been absent for a couple of days, though now he seems to be back blogging in full force. I&#8217;ve started a couple of new blogs this week, one blog for generating sales with adult paysites, and another blog ala ProBlogger, except my target audience is adult webmasters. Since Mike G wrote his anti textbook SEO article, I&#8217;ve gotten curious how bloggers generate traffic, and I&#8217;m now spending some time playing around with side kicks, Technorati tags and chicklets. So I write blogging notes on Adult Blogger while I try out my ideas on my other blog. I intentionally kept Google, Yahoo, and MSN out of my new blog just to see how much traffic I can generate without SEs. I won&#8217;t be linking to those blogs from here. I&#8217;m also thinking of building a time management blog just to help myself manage my days better.</p>
<p>Anyway, I&#8217;ve been pretty busy. I might finish up with some of the drafts on this blog too if I get around to it, but with Google dropping pages every day (this domain went from 58 pages indexed back down to 3), it&#8217;s hard to say anything. Most of my sites rank well if I can get pages indexed. But since Matt said BD&#8217;s crawling priorities are different from the past, its all up in the air. I&#8217;ve actually gotten an email back from Google&#8217;s Search Quality Team, telling me &#8220;Your site has not been manually penalized&#8221; which is cool, but Google needs to start indexing some of my pages.</p>
<p>The biggest bomb of the week is Shoemoney&#8217;s post about plentyoffish (noooo I&#8217;m not gonna link to that). I&#8217;m actually surprised SEOMoz haven&#8217;t blogged about it yet. It&#8217;s taken a nasty turn, and nothing against Shoe, but personally I believe Mark makes plenty of money off his site. All you ever need to make money with a site is traffic and clicks&#8230;and with adsense and 890ish Alexa rating, you can&#8217;t lose. With my sites, I have to not only generate clickthrus, but that click has to generate a sale for me to make money. So he really got it easy.</p>
<p>Anyway, that&#8217;s all I got.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/10/what-ive-been-up-to.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Feedster Feeds off del.icio.us</title>
		<link>http://seo4fun.com/blog/2006/05/03/feedster-feeds-off-delicious.html</link>
		<comments>http://seo4fun.com/blog/2006/05/03/feedster-feeds-off-delicious.html#comments</comments>
		<pubDate>Wed, 03 May 2006 20:31:11 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/03/feedster-feeds-off-delicious.html</guid>
		<description><![CDATA[You&#8217;re probably wondering what&#8217;s so great about feedster traffic, right? :)
Quick side note: I&#8217;ve been struggling trying to decide whether to break up this blog into several blogs, so say people interested in reading about Google don&#8217;t get hit with RSS feeds about MYSQL. But since I don&#8217;t run this blog for money, I&#8217;m going [...]]]></description>
			<content:encoded><![CDATA[<p>You&#8217;re probably wondering what&#8217;s so great about feedster traffic, right? :)</p>
<p><em>Quick side note:</em> I&#8217;ve been struggling trying to decide whether to break up this blog into several blogs, so say people interested in reading about Google don&#8217;t get hit with RSS feeds about MYSQL. But since I don&#8217;t run this blog for money, I&#8217;m going to just blog away with whatever I&#8217;m interested in related to SEO. What I will do is use as descriptive titles as I can, so you can save yourself some time deciding what not to read by scanning my post titles.</p>
<p>Also, since this blog is mainly for my own benefit, expect me to go back to an earlier post and add new information, instead of writing short posts. That&#8217;s not good SEO (more pages, the better), and not good blogging either I know, but right now I have about a hundred unposted posts on hold because I&#8217;m worried about publishing newbie/boring/old ideas that&#8217;ll be a waste of time for you to read.</p>
<p>As part of an experiment to feed videos to Yahoo, I added del.icio.us tags to the video pages and socially bookmarked them using <strong>a batch of the most popular tags related to the content of each page</strong>. I didn&#8217;t get a whole lot of clicks from del.icio.us (20 hits after 12 hours to one page), and after a few hours/days I&#8217;ll expect that to trickle off, but for an unindexed page that&#8217;s not too bad. I also noticed my page shows up on Feedster, though the listings lacked descriptions <strong>(This post does not have a description)</strong> probably because I omitted writing descriptions to my del.icio.us entries.</p>
<p>My notes on del.icio.us, feedster, technorati and&#8230;Furl (so far):</p>
<ul>
<li><strong>Use as many popular tags related to a post as possible</strong> to maximize initial burst of traffic.</li>
<li><strong>Add bookmarking tags/RSS links</strong> on a page to generate repeat visits. (These weren&#8217;t blog pages, so I added them manually).</li>
<li><strong>Write descriptions for del.icio.us entries</strong> so they show up in feedster listings.</li>
<li><strong>Publish bookmarkable/linkbait pages</strong>.</li>
<li><strong>Bookmark with Furl using popular technorati tags as categories.</strong> Technorati Links feed off Furl, using Furl categories. Traffic is low unless you know your high traffic tags, so ROI might not be all that. /tag/ pages have robots=nofollow on them too, so those pages get indexed but you get no link juice, which is no big deal since the link wont stay on there forever anyway.</li>
<li><strong>Repeat a keyword a few times in a post</strong> for longer Technorati description in their search results, which equals bigger adspace.</li>
<li><strong>Tag your post before you ping.</strong> (Yep, I&#8217;m doing them by hand, lol. I&#8217;ll have to try out that Wordpress plugin in a bit.)</li>
<li><strong>Use relevant, but different set of tags per post</strong></li>
<p>For 40+ posts/day tags, hammer them every post; less posts/day tags i can spread them around a bit.</ul>
<h3>Technorati Questions</h3>
<p>How do I know which technorati tag gets decent traffic (besides trial and error?) Does it correlate with Wordtracker/Overture?</p>
<p><strong>Google Hits:</strong> &#8220;how to get indexed by feedster&#8221; Page One, 5/26/2006. Talk about a long tail, no traffic listing.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/03/feedster-feeds-off-delicious.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Yahoo Video RSS</title>
		<link>http://seo4fun.com/blog/2006/05/03/yahoo-video-rss.html</link>
		<comments>http://seo4fun.com/blog/2006/05/03/yahoo-video-rss.html#comments</comments>
		<pubDate>Wed, 03 May 2006 07:54:58 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/03/yahoo-video-rss.html</guid>
		<description><![CDATA[WARNING: Unpolished note to self. I&#8217;ll modify it as I know more about video RSS.
I&#8217;m having some problems getting new videos indexed in Yahoo Video. I&#8217;m reusing XML feed that worked in the past, but Yahoo is just not biting.  Here&#8217;s a list of what I know works and what didn&#8217;t work in the [...]]]></description>
			<content:encoded><![CDATA[<p><strong>WARNING:</strong> Unpolished note to self. I&#8217;ll modify it as I know more about video RSS.</p>
<p>I&#8217;m having some problems getting new videos indexed in Yahoo Video. I&#8217;m reusing XML feed that worked in the past, but Yahoo is just not biting.  Here&#8217;s a list of what I know works and what didn&#8217;t work in the past so I can come up with some ideas.  Getting indexed in Yahoo Video is so much easier if I hosted videos on the same domain, but that&#8217;s out of the question at the moment.</p>
<p>BTW, I also noticed that my usual RSS feed validator doesn&#8217;t detect Media RSS errors.  Use the link on the bottom for validation.</p>
<p>What seems to work before:</p>
<ul>
<li>Embedding videos in HTML, and 302 redirecting.</li>
</ul>
<p>What didn&#8217;t work:</p>
<ul>
<li>Linking to video clips with url /xyz.wmv, then redirecting in .htaccess.</li>
</ul>
<p>With my main site, I&#8217;m linking straight to video clips, so that might be the problem.</p>
<p>One validation error I fixed was this:</p>
<p>&lt;guid isPermaLink=&#8221;false&#8221;&gt;video #x from example.com&lt;/guid&gt;</p>
<p>if isPermaLink=&#8221;true&#8221;, then a valid URL should be used instead of text.</p>
<p>Then again, my older RSS feeds don&#8217;t use guid but they worked.</p>
<p>How would I go about testing some stuff with Yahoo video?</p>
<table cellspacing=5 cellpadding=3 border=1>
<tr>
<td>Type of Link</td>
<td>Same Domain</td>
<td>mod_rewrite</td>
<td>Indexed?</td>
</tr>
<tr>
<td>href</td>
<td>Yes</td>
<td>No</td>
<td>-</td>
</tr>
<tr>
<td>href</td>
<td>No</td>
<td>No</td>
<td>-</td>
</tr>
<tr>
<td>href</td>
<td>No</td>
<td>Yes</td>
<td>Not Yet</td>
</tr>
<tr>
<td>embed</td>
<td>Yes</td>
<td>No</td>
<td>-</td>
</tr>
<tr>
<td>embed</td>
<td>No</td>
<td>Yes</td>
<td>Yes</td>
</tr>
</table>
<p>Another possibility is if RSS feed is sitting on domainx.com and the html is on domainy.com.</p>
<h3>Systematic testing</h3>
<p>Upload 10+ video clips. Write a few different versions of Yahoo RSS feeds. Submit them all to Yahoo. Submitting one feed and waiting for results eats up too much time.</p>
<h2>How do I improve my placement in Yahoo Video Search?</h2>
<p>Right now, competition seems to be thin so placement isn&#8217;t a huge concern, but I&#8217;d still like to know the answer to this one.</p>
<p>Other Yahoo video related questions:</p>
<ul>
<li>If I give Yahoo a specific thumbnail, will Yahoo use that instead of scanning the video? (I can test this by giving them a thumbnail showing the middle of the video instead of the beginning)</li>
<li>Would it save me bandwidth if I specified video dimension, bitrate, filesize?  If I fed Yahoo an erroneous spec, will Yahoo believe me?  If not, I can assume no matter what I specify, Yahoo scans the videos anyway? (Yahoo will scan the video if the media:content url is given.  If only playerURL is mentioned, Yahoo won&#8217;t touch the video).</li>
<li>Can I submit my RSS to places like Technorati?</li>
<li>How do I &#8220;ping&#8221; Yahoo to re-read my RSS file?  Do I just resubmit the RSS url?  How quickly will Yahoo index a new video in an updated RSS?</li>
</ul>
<p>Also, it would be cool if I wrote a script for generating Video Search RSS.  Writing RSS by hand is a bit of a chore.</p>
<h3>Interesting quote regarding PlayerURL Field</h3>
<p>Hopefully, that makes the reasoning a bit more clear. One of the<br />
problems we want to solve is in the cases that a RSS feed can only<br />
publish the playerUrl (for whatever reason).</p>
<p>Our playerUrl is just a normal web page<br />
that shows the media content in the general sense of syndicated web<br />
content.</p>
<p>How do we efficiently<br />
allow the feed to include all interesting meta data about the media<br />
object (such as real height/width)? From the search perspective, we<br />
want this information since we can&#8217;t actually &#8220;touch&#8221; the media<br />
object to determine it ourselves.</p>
<p>Ideally, from the search perspective, we&#8217;d much prefer that the<br />
content provider include both a direct link (url) as well as a player<br />
link (playerUrl). That way, we could harvest all the meta information<br />
from the real content, while still delivering the experience to the<br />
end user that the content provider prefers and/or requires.</p>
<p><strong>So this means supplying a direct link will allow Yahoo to scan the video and generate meta data automatically, while just providing player URL will allow you to supply any url to Yahoo.</strong></p>
<p><a href="http://groups.yahoo.com/group/rss-media/message/26">Yahoo Group: PlayerURL</a></p>
<h2>Yahoo Video Hearsay</h2>
<h3>Some untested comments I picked up while surfing.</h3>
<p>&#8220;the example in the spec leads me to believe that urn:mpaa ratings are<br />
expected to be in lowercase.&#8221;</p>
<p>&#8220;The space in the media:thumbnail url needs to be replaced with %20&#8243;</p>
<p>&#8220;Since you are doing bold, italics, and line breaks inside the<br />
media:text, you want to indicate this by specifying &#8216;type=&#8221;html&#8221;&#8216; on<br />
that eleemnt.&#8221;</p>
<p>&#8220;This is pointing to the channel description, which is described in the<br />
RSS 2.0 spec as being a &#8220;Phrase or sentence describing the channel&#8221;.&#8221;</p>
<h2>Yahoo Video Search Resources and Links</h2>
<p><a href="http://search.yahoo.com/mrss/submit">Yahoo Video submit form</a><br />
<a href="http://search.yahoo.com/mrss">Yahoo Video Media RSS Specification</a><br />
<a href="http://groups.yahoo.com/group/rss-media/">Yahoo Group: RSS Media</a><br />
<a href="http://feedvalidator.org/testcases/ext/media/">Feed Validator Test Cases</a><br />
<a href="http://feedvalidator.org/">Feed Validator</a></p>
<p>For my main site, I want to decide on a page layout that will help me get my videos indexed while not throwing video pages into the supplementals. How can I do this?</p>
<p>One site I see link going straight to videos on the same domain.</p>
<p><strong>Log:</strong> Last submitted a feed to Yahoo on 5/4/2006</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/03/yahoo-video-rss.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Update Wp-Config.php When Re-Installing Wordpress</title>
		<link>http://seo4fun.com/blog/2006/05/01/update-wp-configphp-when-re-installing-wordpress.html</link>
		<comments>http://seo4fun.com/blog/2006/05/01/update-wp-configphp-when-re-installing-wordpress.html#comments</comments>
		<pubDate>Mon, 01 May 2006 17:17:27 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/05/01/update-wp-configphp-when-re-installing-wordpress.html</guid>
		<description><![CDATA[After browsing Shoemoney&#8217;s Toronto pics (link&#8217;s dead so I removed it) at 7 am, I decided to start a new blog. Being half asleep, I used the same set of files I used to create this blog. After installation, I tried to load this blog and got a blank page. Oops. So I refreshed the [...]]]></description>
			<content:encoded><![CDATA[<p>After browsing Shoemoney&#8217;s Toronto pics (link&#8217;s dead so I removed it) at 7 am, I decided to start a new blog. Being half asleep, I used the same set of files I used to create this blog. After installation, I tried to load this blog and got a blank page. Oops. So I refreshed the /blog/ directory using Dreamweaver. That gave me a funky error: <strong>&#8220;Database error: [Unknown column &#8216;user_level&#8217; in &#8216;where clause&#8217;]&#8221;</strong>. After Googling that error message, I learned user_level in wordpress_user table is obsolete in 2.0. I must have had Wordpress 1.0 sitting in my local /blog/ directory! So I backed up .htaccess, deleted /blog/, and uploaded a fresh copy of Wordpress 2.0.</p>
<p>Phew.</p>
<p>REMINDER: When installing a new Wordpress blog, use a fresh copy of Wordpress, or go into wp-config.php and make sure the DB info doesn&#8217;t point to one of your other blogs.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/05/01/update-wp-configphp-when-re-installing-wordpress.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Keyword surrounding Links</title>
		<link>http://seo4fun.com/blog/2006/04/28/78.html</link>
		<comments>http://seo4fun.com/blog/2006/04/28/78.html#comments</comments>
		<pubDate>Fri, 28 Apr 2006 15:36:04 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/04/28/78.html</guid>
		<description><![CDATA[Jim Westergren wrote an interesting post in his blog recently. concluding:
both Yahoo and Google uses the text surrounding links in their algo
So, I looked at this test page that has an outgoing link. I took a snippet next to that link and ran a Google search, but the target page refused to come up. What [...]]]></description>
			<content:encoded><![CDATA[<p>Jim Westergren wrote an <a href="http://www.jimwestergren.com/evidence-of-related-text-giving-rank-boost/">interesting post</a> in his blog recently. concluding:</p>
<blockquote><p>both Yahoo and Google uses the text surrounding links in their algo</p></blockquote>
<p>So, I looked at this <a href="http://72.14.203.104/search?q=cache:uJRK8UeV5ccJ:seo4fun.com/test-1/brandnewsacx1-sasahz.html+site:seo4fun.com+Maecenas+auctor+&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=1&#038;ie=UTF-8">test page</a> that has an outgoing link. I took a snippet next to that link and ran a <a href="http://www.google.com/search?hl=en&#038;lr=&#038;ie=ISO-8859-1&#038;safe=off&#038;rls=GGLD%2CGGLD%3A2004-45%2CGGLD%3Aen&#038;q=site%3Aseo4fun.com+Maecenas+auctor+&#038;btnG=Search">Google search</a>, but the target page refused to come up. What am I missing?</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/04/28/78.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Power of META tags in Yahoo</title>
		<link>http://seo4fun.com/blog/2006/04/27/power-of-meta-tags-in-yahoo.html</link>
		<comments>http://seo4fun.com/blog/2006/04/27/power-of-meta-tags-in-yahoo.html#comments</comments>
		<pubDate>Thu, 27 Apr 2006 16:00:10 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[On Page Optimization]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/04/27/power-of-meta-tags-in-yahoo.html</guid>
		<description><![CDATA[After two months, Yahoo is finally starting to show a few pages from this site in their index. This SERP shows pages with keywords in META keyword/description rank in Yahoo, even when keywords are absent from the page copy.  In fact, a page with keyword in just META description tag outranked this site&#8217;s home [...]]]></description>
			<content:encoded><![CDATA[<p>After two months, Yahoo is finally starting to show a few pages from this site in their index. <a href="http://search.yahoo.com/search?p=site%3Aseo4fun.com+Brandnewsacx1+sasahz&#038;prssweb=Search&#038;ei=UTF-8&#038;fr=FP-tab-vid-t&#038;x=wrt">This SERP</a> shows pages with keywords in META keyword/description rank in Yahoo, even when keywords are absent from the page copy.  In fact, a page with keyword in just META description tag outranked this site&#8217;s home page, though since only a few pages are competiting for the same keyword, the SERP ranking order may not actually mean much. Both Google and MSN ignore keywords in META tags, at least when keywords don&#8217;t appear on a page. They may be useless - period - but I&#8217;ll have to confirm that with a test when I have more time to burn.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/04/27/power-of-meta-tags-in-yahoo.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Duplicate Content Revisited</title>
		<link>http://seo4fun.com/blog/2006/04/16/duplicate-content-according-to-tedster.html</link>
		<comments>http://seo4fun.com/blog/2006/04/16/duplicate-content-according-to-tedster.html#comments</comments>
		<pubDate>Sun, 16 Apr 2006 18:16:12 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[Supplementals]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/04/16/duplicate-content-according-to-tedster.html</guid>
		<description><![CDATA[Tedster wrote a meaty post in WMW concerning duplicate content:
Google tries to select the dupes and then put all but one of them into the &#8220;supplemental index&#8221;. If a domain has just a few instances of duplication like this in the Google index, things tend to go on as normal. But when many, many urls [...]]]></description>
			<content:encoded><![CDATA[<p><a href="http://www.webmasterworld.com/forum30/33899.htm">Tedster wrote a meaty post in WMW concerning duplicate content</a>:</p>
<blockquote><p>Google tries to select the dupes and then put all but one of them into the &#8220;supplemental index&#8221;. If a domain has just a few instances of duplication like this in the Google index, things tend to go on as normal. But when many, many urls start showing up, all with identical content, then something seems to get tripped at Google and a site can start to see trouble.</p></blockquote>
<p>One of my domains in supplemental hell (where not even my home page can be found on the first page of a site: query) uses a few directories to link off to sponsors - this adds up to a few thousand supplemental URLs out of 7,000 indexed. I also have session ID type seed in urls that results in thousands of supplementals. The result? I&#8217;m seeing even unqiue pages end up being listed by Google as supplemental.</p>
<p>Still, didn&#8217;t Matt Cutts say there&#8217;s no such thing as &#8220;duplicate content penalty?&#8221; What tedster says doesn&#8217;t sound like a mere filter - it sounds like a site wide penalty. Then again, I never fully trust what anyone says. I mean&#8230;Google didn&#8217;t even notice the sandbox phenomenon until someone pointed it out to them :)</p>
<p>Why doesn&#8217;t Google just want to forget supplementals? (speculation) Because if any page points to them, Moz-bot will have to recrawl them later, and it&#8217;s better to have records of these urls to avoid wasting time recrawling and indexing them.</p>
<p>Makes you want to start a few domains over from scratch, doesn&#8217;t it? My only hope at this point is to increase the number of incoming links, beef up PR sitewide and write more content. Still, seeing site:domain.com return the same SERP for the last 2-3 weeks isn&#8217;t encouraging.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/04/16/duplicate-content-according-to-tedster.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Supplemental Test Update April 13th, 2006</title>
		<link>http://seo4fun.com/blog/2006/04/13/supplemental-test-update-april-13th-2006.html</link>
		<comments>http://seo4fun.com/blog/2006/04/13/supplemental-test-update-april-13th-2006.html#comments</comments>
		<pubDate>Fri, 14 Apr 2006 03:51:10 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Supplementals]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/04/13/supplemental-test-update-april-13th-2006.html</guid>
		<description><![CDATA[I&#8217;ve been generally keeping up with WMW posts and other SEO blogs and even kept webmaster radio running since noon today, but nothing is really grabbing my interest. I do have a pile of blog post drafts sitting around unfinished, though some of it is so specific to my domains that I&#8217;m not sure who [...]]]></description>
			<content:encoded><![CDATA[<p>I&#8217;ve been generally keeping up with WMW posts and other SEO blogs and even kept webmaster radio running since noon today, but nothing is really grabbing my interest. I do have a pile of blog post drafts sitting around unfinished, though some of it is so specific to my domains that I&#8217;m not sure who will benefit from reading them. I opened up this blog to keep my thoughts organized, but I don&#8217;t want to go so far as to air dirty laundry in public (like the mess BD made after March 28). I&#8217;ve also spent some time today checking out Google&#8217;s calendar and Google Reader after Pluck started freezing IE. Because of that, I didn&#8217;t get a whole lot done today. A female buddy from Holland kept ICQing me throughout the day, which didn&#8217;t help either.</p>
<p>Anyway, here&#8217;s a few casual observations on how these <a href="http://www.google.com/search?hl=en&#038;lr=&#038;safe=off&#038;q=site%3Aseo4fun.com%2Ftest-2%2F&#038;btnG=Search">supplemental test pages</a> are doing on Google. The results are temporary so what&#8217;s not supplemental today may not even be in the index a few days from now. Also, the fact that these test pages aren&#8217;t in English may have an effect in keeping these pages from going supplemental. I may have to find a random text generator in English. It could also be that recently crawled pages are handled differently than before and pages are either dropped or indexed, instead of falling into supplemental index (just a guess).</p>
<ul>
<li>Pages with unclosed head/body tags are indexed fine.</li>
<li>Pages 60%, 75%, 80% similar to the original is not supplemental.</li>
<li>A page with unique title/description but the same body text as the original page is not supplemental.</li>
<li>Original page is indexed.</li>
<li>A tiny page with 20 words is indexed and not supplemental.</li>
<li>A previously indexed page 90% similar to the original is gone from the index. The root page http://seo4fun.com/ is also gone; only /index.html remains. Is this an indication that Google is dropping pages from the index instead of storing them in the supplemental index? Or are duplicate content pages just being filtered out from SERPs?</li>
<li>Notice that optimization for these pages are identical (just one occurance of keyword) but <a href="http://www.google.com/search?hl=en&#038;lr=&#038;safe=off&#038;q=site%3Aseo4fun.com%2Ftest-2%2F+se4funsupp+testcases&#038;btnG=Search">this SERP i</a>s making me think #1 is a direct result of keyword placed at the start of a sentence.</li>
<li><a href="http://www.google.com/search?hl=en&#038;lr=&#038;safe=off&#038;q=site%3Aseo4fun.com%2Ftest-2%2F+se4funsupp+testcases&#038;btnG=Search">This SERP</a> shows the 80% similar page get dumped in omitted results, because the paragraph containing the keyword is identical for those pages.</li>
<li>One weird thing I noticed is the normal SERP returned 9 of 13 (4 pages omitted) and 11 when I clicked the omitted results link. The 4 omitted pages are those starting a paragraph &#8220;Se4funsupp testcases nullam velit libero&#8221; phrase. In this case, Google somehow chose the 60% similar page and dumped the rest in omitted. The omitted pages are: 80% similar / The original page / Unique title and desc identical body text / 70% similar. Now for different queries, I can see pages getting omitted, but its interesting that Google picked the 60% similar page to keep in the SERPS. It could be due to something else but only one incoming link goes to each page so everything else should be equal. Also, when I click the omitted results link these 4 pages go back in but why only 11? Which are the 2 missing pages? A page with broken head/body tag (2-15) and (2-11) 100 Words page disappears. Not sure why they poof.</li>
</ul>
<p>By the way, a brief update to my previous MSN post. <a href="http://www.google.com/search?hl=en&#038;lr=&#038;safe=off&#038;q=site%3Aseo4fun.com+Brandnewsacx1+sasahz&#038;btnG=Search">This SERP</a> shows the home page and SEO combined page in 1st/2nd in Google. The home page is probably ranking 1st merely due to PR 3 (other PR 3s are /blog/ and /carcasher page). I&#8217;ll mention the keyword once on this page (&#8221;Brandnewsacx1 sasahz&#8221;) and see how high it ranks. Notice many of the keyword stuffed pages ranking high. Though a page with keyword in the title still ranked higher than any of the keyword stuffed pages, I still got this SERP - keyword repeated 11x (4th) 10x(5th) 9x(6th). Coincidence? Probably, but it&#8217;s eery.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/04/13/supplemental-test-update-april-13th-2006.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>MSN Optimization</title>
		<link>http://seo4fun.com/blog/2006/04/08/msn-optimization.html</link>
		<comments>http://seo4fun.com/blog/2006/04/08/msn-optimization.html#comments</comments>
		<pubDate>Sat, 08 Apr 2006 08:37:00 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/04/08/msn-optimization.html</guid>
		<description><![CDATA[Probably old news to most SEOs, since I dont bother with MSN all that much, the idea that on-page optimization is king for MSN is news to me. 	This serp seems to confirm it, where the page that repeats keywords 11 times ranks the highest among all other types of optimizations.
Looking at my test pages, [...]]]></description>
			<content:encoded><![CDATA[<p>Probably old news to most SEOs, since I dont bother with MSN all that much, the idea that <a href="http://www.seochat.com/c/b/MSN-Optimization-Help/">on-page optimization is king for MSN</a> is news to me. 	<a href="http://search.msn.com/results.aspx?q=site%3Aseo4fun.com+Brandnewsacx1+sasahz&#038;FORM=QBRE">This serp</a> seems to confirm it, where the page that repeats keywords 11 times ranks the highest among all other types of optimizations.</p>
<p>Looking at my test pages, results are like night and day on MSN and Google. I guess I should really build 3 different sites running on the same database for 3 search engines, which is actually ridiculously easy do, except building incoming links &#8212; that would be the hard part, to split links among three versions of the same site. But it would make sense at least in the short term to build a site specifically targetting MSN.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/04/08/msn-optimization.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Googlebot Slowing to a Crawl</title>
		<link>http://seo4fun.com/blog/2006/04/07/googlebot-slowing-to-a-crawl.html</link>
		<comments>http://seo4fun.com/blog/2006/04/07/googlebot-slowing-to-a-crawl.html#comments</comments>
		<pubDate>Fri, 07 Apr 2006 19:55:15 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/04/07/googlebot-slowing-to-a-crawl.html</guid>
		<description><![CDATA[I thought I was the only one stuck in supplemental hell with no Googlebot to rectify the problem/pick up new pages but according to a thread in WMW it seems I&#8217;m not the only one experiencing this problem. Some guys report last visit by Googlebot at the end of March, which is pretty much the [...]]]></description>
			<content:encoded><![CDATA[<p>I thought I was the only one stuck in supplemental hell with no Googlebot to rectify the problem/pick up new pages but according to a <a href="http://www.webmasterworld.com/forum30/33806.htm">thread in WMW</a> it seems I&#8217;m not the only one experiencing this problem. Some guys report last visit by Googlebot at the end of March, which is pretty much the same as me. In fact, this month I&#8217;ve had more visits by Inktomi Slurp than Googlebot.</p>
<p>One thing I&#8217;d like to know is how do I tell Mozilla Googebot and Googlebot apart, but right now I&#8217;m too lazy to look it up :)</p>
<p>I&#8217;m also starting to read about a <a href="http://www.webmasterworld.com/forum30/33808.htm">Google rollback to Aug 2005</a>, with pages indexed from Jan 2006 up to March 28 lost in transition to BD - which sounds like the problem with my site.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/04/07/googlebot-slowing-to-a-crawl.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Update Dynamic Pages like Walking on Glass</title>
		<link>http://seo4fun.com/blog/2006/04/04/update-dynamic-pages-like-walking-on-glass.html</link>
		<comments>http://seo4fun.com/blog/2006/04/04/update-dynamic-pages-like-walking-on-glass.html#comments</comments>
		<pubDate>Tue, 04 Apr 2006 07:49:41 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/04/04/update-dynamic-pages-like-walking-on-glass.html</guid>
		<description><![CDATA[Never change php code on a live page or underlying PHP classes called before writing a test class/test page first. You can crash hundreds or thousands of pages at once with one false move and if spiders happen to be deep crawling your site, boom, you&#8217;ve created a mess you might not be able to [...]]]></description>
			<content:encoded><![CDATA[<p>Never change php code on a live page or underlying PHP classes called before writing a test class/test page first. You can crash hundreds or thousands of pages at once with one false move and if spiders happen to be deep crawling your site, boom, you&#8217;ve created a mess you might not be able to recover from, ever.</p>
<p>Here&#8217;s an example. Say you want to change livepage.html. Don&#8217;t touch it. Create  livepage1.html and make changes to that page. Test it out to make sure it does what you want. Then save the page as livepage.html. Pretty basic stuff, but something you should never forget to do.</p>
<p>Other PHP Good Practices</p>
<p>- Make pages as PHP independent as possible. Try to use cached elements instead of doing live PHP calls.<br />
- Write clean PHP classes. Create sensible class extensions.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/04/04/update-dynamic-pages-like-walking-on-glass.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Dynamic / Fixed Tables</title>
		<link>http://seo4fun.com/blog/2006/04/01/dynamic-fixed-tables.html</link>
		<comments>http://seo4fun.com/blog/2006/04/01/dynamic-fixed-tables.html#comments</comments>
		<pubDate>Sat, 01 Apr 2006 14:03:32 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/04/01/dynamic-fixed-tables.html</guid>
		<description><![CDATA[Read an interesting post by AlexK on WMW on the pros and cons of fixed/dynamic MYSQL tables. I was looking for some info to decide whether to keep my multiple tables or join them. Joining them will simplify inserts/updates but what I&#8217;m also wondering is its impact on speed. According to AlexK, fixed tables will [...]]]></description>
			<content:encoded><![CDATA[<p>Read an interesting post by AlexK on WMW on the pros and cons of <a href="http://www.webmasterworld.com/forum88/9443.htm">fixed/dynamic MYSQL tables</a>. I was looking for some info to decide whether to keep my multiple tables or join them. Joining them will simplify inserts/updates but what I&#8217;m also wondering is its impact on speed. According to AlexK, fixed tables will auto convert to dynamic when I add Text/Blob/varchar (which I usually do), so exporting just those to a separate table and running left joins is his recommendation.</p>
<p><a href="http://dev.mysql.com/doc/mysql/en/static-format.html">MYSQL reference regarding Fixed MYSQL tables</a></p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/04/01/dynamic-fixed-tables.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Why I don&#8217;t Like Index.html</title>
		<link>http://seo4fun.com/blog/2006/04/01/why-i-dont-like-indexhtml.html</link>
		<comments>http://seo4fun.com/blog/2006/04/01/why-i-dont-like-indexhtml.html#comments</comments>
		<pubDate>Sat, 01 Apr 2006 11:22:22 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Supplementals]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/04/01/why-i-dont-like-indexhtml.html</guid>
		<description><![CDATA[The obvious answer is supplementals in Google. I&#8217;ve used index.html on about 3 of my domains and since I use Dreamweaver, and sooner or later I make the mistake of linking to a page using /index.html and boom&#8230; Google will index it. Even this domain has /index.html for root url. It&#8217;s a good thing that&#8217;s [...]]]></description>
			<content:encoded><![CDATA[<p>The obvious answer is supplementals in Google. I&#8217;ve used index.html on about 3 of my domains and since I use Dreamweaver, and sooner or later I make the mistake of linking to a page using /index.html and boom&#8230; Google will index it. Even this domain has /index.html for root url. It&#8217;s a good thing that&#8217;s the only index page I have on this domain (besides the crap wordpress generates, but that&#8217;s another story).</p>
<p>BTW, I noticed both / and /index.html on this domain has the same cache date and still both are managing to stay out of supplementals. This could mean a tweak in how pages are crawled? I&#8217;m not sure what it is, but one of them should be listed as supplemental. I&#8217;m sure later on it will be, but this tells me depending on adjustments made to Google&#8217;s duplicate content filter, which will come eventually, some pages listed as unique pages will get thrown into supplemental listings somewhere down the line.</p>
<p>So just because pages are not flagged as supplementals now won&#8217;t mean it won&#8217;t be later. Scary, huh?</p>
<p>Also, completely unrelated, but i got a hit for &#8220;why did i get deindexed by google&#8221; so I ran a search. I see this site listed under digitalpoint on page one but I notice:</p>
<p>Did you mean: why did i get reindexed by google?</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/04/01/why-i-dont-like-indexhtml.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Doesn&#8217;t Parse Keywords in URLS</title>
		<link>http://seo4fun.com/blog/2006/03/13/google-doesnt-parse-keywords-in-urls.html</link>
		<comments>http://seo4fun.com/blog/2006/03/13/google-doesnt-parse-keywords-in-urls.html#comments</comments>
		<pubDate>Tue, 14 Mar 2006 02:35:27 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/03/13/google-doesnt-parse-keywords-in-urls.html</guid>
		<description><![CDATA[A few days ago, John Scott blogged that Google doesn&#8217;t parse keywords in urls. I was previously told by many reputable webmasters that Google can parse common keywords in urls. Here&#8217;s what would seem to be an indisputable proof: searching for &#8220;search engine optimization&#8221; returns searchenginewatch.com, with the words search engine in the domain name [...]]]></description>
			<content:encoded><![CDATA[<p>A few days ago, <a href="http://www.internet-marketing-blog.com/2006/03/10/parsing-keywords-in-urls/">John Scott blogged that Google doesn&#8217;t parse keywords in urls</a>. I was previously told by many reputable webmasters that Google can parse common keywords in urls. Here&#8217;s what would seem to be an indisputable proof: searching for &#8220;<a href="http://www.google.com/search?hl=en&#038;lr=&#038;q=search+engine+optimization&#038;btnG=Search">search engine optimization</a>&#8221; returns <strong>search</strong><strong>engine</strong>watch.com, with the words <em>search engine</em> in the domain name highlighted. So that proves Google does parse keywords in urls, right?</p>
<p>Wrong.  John Scott&#8217;s tests demystify this notion completely. <strong>He points out that highlighting does not mean parsing.  </strong>Even for common keywords like &#8220;sexual frustration&#8221;, Google fails to parse them in URLs.</p>
<p>Another proof is <a href="http://www.google.com/search?hl=en&#038;lr=&#038;ie=ISO-8859-1&#038;safe=off&#038;rls=GGLD%2CGGLD%3A2004-45%2CGGLD%3Aen&#038;q=site%3Aseo4fun.com+fun">this search</a> for the word &#8220;fun&#8221; on this site. If Google recognized the word fun in the domain name, it would return all pages in the domain; instead, it returns only the home page, which is the only page that contains the word fun on the page.</p>
<p>This teaches me a couple of things: 1) Never blindly buy into what an authority says 2) When in doubt, test it out =)<br />
<strong> </strong></p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/03/13/google-doesnt-parse-keywords-in-urls.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Does Google&#8217;s Spam Reports Work?</title>
		<link>http://seo4fun.com/blog/2006/03/13/does-googles-spam-reports-work.html</link>
		<comments>http://seo4fun.com/blog/2006/03/13/does-googles-spam-reports-work.html#comments</comments>
		<pubDate>Mon, 13 Mar 2006 12:53:42 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/03/13/does-googles-spam-reports-work.html</guid>
		<description><![CDATA[A few days ago, Matt posted a request on his blog for people to file spam reports, especially for keyword stuffing and Asian spam sites. After reading that, I had a chat with another webmaster friend who thought spam reports were useless. Her point was that spammers injected hundreds of new spam domains every day, [...]]]></description>
			<content:encoded><![CDATA[<p>A few days ago, <a href="http://www.mattcutts.com/blog/send-more-spam-reports/#comments">Matt posted a request on his blog for people to file spam reports</a>, especially for keyword stuffing and Asian spam sites. After reading that, I had a chat with another webmaster friend who thought spam reports were useless. Her point was that spammers injected hundreds of new spam domains every day, and reporting each one, even if Google acted on those reports, was a waste of time. Her business also doesn&#8217;t rely much on Google traffic, so that does skew her perspective a bit.</p>
<p>On the other hand, I file spam reports often. I think spam reports are useful and they&#8217;ve worked in the past to knock out hundreds of domains. The dissatisfied? link can be used to report spam in batch mode. I also hardly doubt Google just takes spam reports, verifies them, then black list individual sites. I&#8217;m sure they&#8217;re taking note of spam tactics used to improve their spam filters.</p>
<p>Anyway, a few days after Matt posted the request on his blog, there are over 60 replies on it. Most are your average gripes and views like &#8220;spamming is bad&#8221;, &#8220;adsense created spam&#8221;, etc. Still, here&#8217;s a list of some comments that stuck out in my mind:</p>
<blockquote><p>Google does absoluetly nothing against hidden noscript content (or content in hidden divs). I reported about 5 examples to Google and Matt about hidden div spam ages ago and nothing was done, so now I do it on all my sites and am making an absolute killing. I’ll keep doing it until I see the original spammer I reported get axed (that’s <a rel="nofollow" href="http://www.ambergreeninternetmarketing.com/">www.ambergreeninternetmarketing.com</a> and its clients).</p></blockquote>
<p>I admit after reading that I wanted to see a concrete example to test out on this site.</p>
<p>Rawalex adds:</p>
<blockquote><p>Harith, the response time between “spam in the index” and “spam removed from index” is very critical to how profitable spamming can be.</p></blockquote>
<p>PhilC writes:</p>
<blockquote><p>If I ran a search engine, and I wanted to deal with spam, I definitely wouldn’t remove sites that were reported as spamming. I’d want to write algos that would deal with them, and others that contain the same sort of spam, and I’d want them in the index so that I could see if the algos worked.</p></blockquote>
<p>This line made me chuckle:</p>
<blockquote><p>You’re sending Matt unsolicited job offers through spam reports? That’s sort  of ironic isn’t it?</p></blockquote>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/03/13/does-googles-spam-reports-work.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Robots.txt Before Linking Up</title>
		<link>http://seo4fun.com/blog/2006/03/10/robotstxt-before-linking-up.html</link>
		<comments>http://seo4fun.com/blog/2006/03/10/robotstxt-before-linking-up.html#comments</comments>
		<pubDate>Fri, 10 Mar 2006 11:30:49 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Supplementals]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/03/10/robotstxt-before-linking-up.html</guid>
		<description><![CDATA[First thing you should do after you buy a domain is install .htaccess that deals with canonical issues like non-www and /index.html.
Second thing you should do is install a robots.txt that prevents Google from crawling anything except the domain root.
If a hacker decides to submit your urls with a Google url removal tool, this may [...]]]></description>
			<content:encoded><![CDATA[<p>First thing you should do after you buy a domain is install .htaccess that deals with canonical issues like non-www and /index.html.</p>
<p>Second thing you should do is install a robots.txt that prevents Google from crawling anything except the domain root.</p>
<p>If a hacker decides to submit your urls with a Google url removal tool, this may suck, but in general this will prevent alot of headaches later down the road.</p>
<p>If your site has low PR, Google won&#8217;t come around to deep crawl the site every day. If you screw up and send 100s of pages to supplemental hell, you&#8217;ll have even a harder time getting them recrawled and getting them back into the main index.</p>
<p>So imagine you get one shot for a clean crawl. Block spiders with robots.txt; build the site, check for dup problems, and only when your site&#8217;s ready, let Googlebot in.</p>
<p>&#8212;</p>
<p>Disallowing Googlebot from indexing a page in Robots.txt only seems to work if I put up a robots.txt file before Google ever spiders the page.  If I use robots.txt after the pages are already crawled, the disallowed pages just get thrown into the supplementals and stay there till the end of time.  If an entire site is blocked by robots.txt, it may be hidden from the SERPS but I suspect Google will still have records of the site in its database.</p>
<p>Evidence?</p>
<p>I have two directories on one domain I use to clicktrack out to sponsors.  First domain I noticed these urls were cluttering up my site:index results in Google, so I put a robots.txt to block them from getting indexed.  I also hoped that by putting up robots.txt, Google would eventually drop those urls from its index.</p>
<p>I also included a new directory in the same robots.txt, and switched many of my outgoing sponsor links to use this directory instead.  After 6 months, none of the links under this directory shows up when I run site:xyz.com/directory/.</p>
<p><strong>Well, w</strong><strong>hat about WMW blocking their entire site using robots.txt?</strong>  That seemed to work.  <em>But did those pages really get deindexed</em>, or did they just stopped showing?  My guess is they were just hidden from the SERPs.</p>
<p>Anyway, experience tells me r<strong>obots.txt is not a tool for getting Google to drop pages from their index.</strong>  Your page may be dropped from the main index, but most likely, they&#8217;ll migrate into the supplementals and stay there indefinitely.</p>
<p><strong>If you want to keep Google out of a directory, put up robots.txt before that directory goes live. </strong></p>
<h2>Matt Cutts on Robots.txt</h2>
<p><a href="http://www.mattcutts.com/blog/googlebot-keep-out/">Matt Cutt&#8217;s wrote an interesting post on 3/17/2006 about robots.txt</a> and why sometimes pages show up in Google SERP as URL only even when robots.txt forbade Google from indexing it:</p>
<blockquote><p>Obscure note #1: using the ‘googlebot=nocrawl’ technique would not be the preferred method in my mind. Why? Because it might still show ‘googlebot=nocrawl’ urls as uncrawled urls. <strong>You might wonder why Google will sometimes return an uncrawled url reference, even if Googlebot was forbidden from crawling that url by a robots.txt file.</strong> There’s a pretty good reason for that: back when I started at Google in 2000, several useful websites (eBay, the New York Times, the California DMV) had robots.txt files that forbade any page fetches whatsoever. Now I ask you, what are we supposed to return as a search result when someone does the query [california dmv]? We’d look pretty sad if we didn’t return www.dmv.ca.gov as the first result. But remember: we weren’t allowed to fetch pages from www.dmv.ca.gov at that point. The solution was to show the uncrawled link when we had a high level of confidence that it was the correct link. Sometimes we could even pull a description from the Open Directory Project, so that we could give a lot of info to users even without fetching the page. I’ve fielded questions about Nissan, Metallica, and the Library of Congress where someone believed that Google had crawled a page when in fact it hadn’t; a robots.txt forbade us from crawling, but Google was able to show enough information that someone assumed the page had been crawled. Happily, most major websites (including all the ones I’ve mentioned so far) let Google into more of their pages these days.</p>
<p>That’s why we might show uncrawled urls in response to a query, even if we can’t fetch a url because of robots.txt. So ‘googlebot=nocrawl’ pages might show up as uncrawled. The two preferred ways to have the pages not even show up in Google would be A) to use the “noindex” meta tag that I mentioned above, or B) to use the url removal tool that Google provides. I’ve seen too many people make a mistake with option B and shoot themselves in the foot, so I would recommend just going with the noindex meta tag if you don’t want a page indexed.
</p></blockquote>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/03/10/robotstxt-before-linking-up.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Wrong SERP Snippet for Cache</title>
		<link>http://seo4fun.com/blog/2006/03/08/wrong-serp-snippet-for-cache.html</link>
		<comments>http://seo4fun.com/blog/2006/03/08/wrong-serp-snippet-for-cache.html#comments</comments>
		<pubDate>Wed, 08 Mar 2006 12:10:13 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Supplementals]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/03/08/wrong-serp-snippet-for-cache.html</guid>
		<description><![CDATA[I&#8217;ve always assumed title/description snippets displayed in the SERP reflects what&#8217;s in Google&#8217;s Cache.  But now, I&#8217;m starting to see at least one page where title/description doesn&#8217;t match what&#8217;s stored in the cache.
Here&#8217;s an example:
The cache of one page I&#8217;m looking at (I won&#8217;t post the url since its adult related) is dated 3/5/2006, [...]]]></description>
			<content:encoded><![CDATA[<p>I&#8217;ve always assumed title/description snippets displayed in the SERP reflects what&#8217;s in Google&#8217;s Cache.  But now, <strong>I&#8217;m starting to see at least one page where title/description doesn&#8217;t match what&#8217;s stored in the cache.</strong></p>
<p>Here&#8217;s an example:</p>
<p>The cache of one page I&#8217;m looking at (I won&#8217;t post the url since its adult related) is dated 3/5/2006, and the cached page is up to date, but the title/description that actually displays in the SERP is from an ancient version of the page.</p>
<p>I also noticed that for this particular page <strong>across a dozen+ DCs, there are at least 2 versions of the cache.</strong>  For a particular page I&#8217;m looking at, the dates on the cache are Aug 9 and Mar. 5.  There&#8217;s no way the March. 5 cache is stored in a supplemental database (the way I rewrote the page, there&#8217;s no way it&#8217;ll end up in there), so <strong>even though the page is listed as supplemental, the cache must be from a live repository.</strong></p>
<blockquote><p>My guess is <strong>the document&#8217;s hits aren&#8217;t re-counted yet and its snippet is still outdated.</strong> Though the page is crawled and cached recently, according to the document snippet, the page is supplemental.</p></blockquote>
<p>I&#8217;ll check a few more pages to confirm this.  But if this is true, <strong>I&#8217;m still inclined to believe Google has no more than one copy of a page in their supplemental database.</strong></p>
<p>This is slightly off topic, but here&#8217;s another note to myself when looking at a domain listing in Google pages are either:</p>
<ul>
<li>main index</li>
<li>supplemental</li>
<li>main index but similar</li>
<li><strong>supplemental and similar < - </strong>these are hardest to spot.</strong></li>
</ul>
<p><em>So far, I found 2 urls with snippets not corresponding to the cache.</em></p>
<p><strong>Cached pages for a few other supplementals are identical (timestamps and stored HTML are the same) across all DCs, </strong>so this one page must have been a glitch.</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/03/08/wrong-serp-snippet-for-cache.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Cache</title>
		<link>http://seo4fun.com/blog/2006/03/06/google-cache.html</link>
		<comments>http://seo4fun.com/blog/2006/03/06/google-cache.html#comments</comments>
		<pubDate>Tue, 07 Mar 2006 03:03:57 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Supplementals]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/03/06/google-cache.html</guid>
		<description><![CDATA[I just checked my site at 216.239.59.147 and noticed a huge drop in pages indexed.  Either I&#8217;m still doing something wrong or Google is hiccupping again.
I need to check my pages on this DC and see how many of my pages including subdomains are indexed correctly.
Since Google keeps falling back to cache from August [...]]]></description>
			<content:encoded><![CDATA[<p>I just checked my site at 216.239.59.147 and noticed a huge drop in pages indexed.  Either I&#8217;m still doing something wrong or Google is hiccupping again.</p>
<p>I need to <strong>check my pages on this DC and see how many of my pages including subdomains are indexed correctly.</strong><br />
Since Google keeps falling back to cache from August these days, I&#8217;m wondering if</p>
<p>A. Google keeps several versions of document in its cache per DC</p>
<p>B. Google keeps one version of document per DC.</p>
<p>Also, considering BD seems to shift in and out of a DC, I&#8217;m not sure what the hell I&#8217;m seeing.</p>
<p>Maybe my site is lacking PR to get crawled often, but considering the major supplemental &#8220;bug&#8221; hitting alot of well-established sites out there, lack of PR may only be a small part of the problem.</p>
<p>Question: Are the pages listed as supplementals supposed to be supplemental?  (definitely, some pages are supposed to be supplemental; usually every domain naturally has a few supps). Were the pages close call dupes that got tipped over into supplementals due to a bug in the dup filter aglo?</p>
<p>I&#8217;ll have to <strong>take a look at a few of my competitor sites to see how their pages are holding up in Google.</strong></p>
<p>I just did a &#038;filter=0 to my site and <strong>the page count jumped from 700 to 10,900!?</strong>  Now what in the hell is going on?</p>
<p>Another thing, how did Google crawl/cache sites again?  It has a few Mozilla/regular crawlers (around 300?) and they cache the information where?  In a single repository or do they dump their info on separate DCs (which would make no sense).  They obviously <strong>must all share the same docID and url hash</strong>&#8230;but what about the HTML cache?  Is there some status field which prevents a page from being displayed depending on the DC?<br />
Why would a site show on one DC and not display at all on another?  I&#8217;m missing a basic SEO 101 info here&#8230;</p>
<p>Here&#8217;s an interesting speculation by <a href="http://www.webmasterworld.com/forum30/33386-7-10.htm">lammert on supplementals at wmw</a> (msg #61):</p>
<p><font size="2" face="verdana" color="#000000" class="mo">The search engine index is primarily designed to store pages, pages and even more pages. The rate at which new pages occur is higher than the rate at which old pages disappear. This was at least the situation a few years ago. As a programmer I wouldn&#8217;t be surprised if Google designed the database to be add-only, and solved the delete problem just as DBase did, by marking unwanted records with a flag.</font></p>
<p>According to <a href="http://www.webmasterworld.com/forum30/33386-5-10.htm">g1smd at wmw</a>, Google has a database for supplementals apart from a live cache, and that supplemental cache once it gets in there is permanent, even when hosting is taken down and a domain no longer exists.</p>
<p>Think of this in terms of MYSQL tables: google_supplemental DocID / url hash / cache date/ cache content.</p>
<p>google_live DocID / urlhash / cache date/ cache content.</p>
<p>Whatever record gets added in google_supplemental is never deleted.</p>
<p>Interestingly, steveb added: &#8220;there are two sets of supplementals on different datacenter groups, so depending  on what datacenter you hit you could see a different batch.&#8221;</p>
<p>I can&#8217;t confirm this, but it&#8217;s worth noting.</p>
<p>Another quote, this one from Dayo_uk: &#8221;</p>
<p>These pages have been  crawled and cached in the supplimental index but not been crawled or cached in  the normal index&#8230;.The question that people should be asking themselves is why Google are now not listing there pages in the normal crawl as theses have disappeared rather than the pages going supplimental (<strong>as a supplimental copy was probably already there</strong>).</p>
<p>So&#8230;<strong>pages didn&#8217;t go supplemental, just that pages in the main index is no longer being displayed(?)</strong></p>
<p>I think Dayo_UK is on to something here.  The pages that were crawled correctly (e.g. the homepage, which never had a supplemental problem) is showing perfectly in BD DCs.</p>
<p>Now I ran another subdomains page count, a few minutes ago I saw 584 indexed on 64.233.179.104.  Now I see 684 on the same DC.  Even with filter=0, the resulting number is the same.  I guess it could be a timing thing&#8230;but where is that 10,900 number coming from?  Is that just an approximation?  I ran page counts on every directory and they don&#8217;t add up anywhere close to that huge number.</p>
<p>I know Google sometimes hides pages, but usually if I do a site: targetting a directory, it will show the pages that are being suppressed.  I just can&#8217;t figure it out.</p>
<p>Even with a 684 page count, Google is only displaying 276, which is pretty much identical to the number of pages Yahoo has indexed.  Am I missing something?  And why are the rest of the stuff not being displayed? Are they supplemental? filter=0 won&#8217;t display them.</p>
<p>Is this some kind of subdomain penalty??</p>
<p>site:janesguide.com -inurl:www returns 607 pages today on the same DC, around 86 urls as unique. Lets see why they could be supplemental.</p>
<ul>
<li>Some urls with 5 query strings are supplementals (they should be).</li>
<li>Some urls are not supplental but are hidden as &#8220;similar pages&#8221; due to identical META description: <em>&#8220;Since 1997, JanesGuide and Jane Duvall have been your guide &#8230;&#8221;</em></li>
<li>All subdomain root urls are cached correctly and 607 out of 607 shows up in the SERP.</li>
<li>Interesting to note that the pages are all light, around 7k</li>
<li>The nav menus are below content, and there are 240 unique words on the page.</li>
</ul>
<p>At around 350 urls, the rest are going supplemental.</p>
<p>I refresh a few minutes later and now the page count is 718, and a 10k page count with filter=0. 306 documents shown.</p>
<p>I only use a few &#038;query type urls, and I have most of them blocked by NOINDEX.  Could be Google indexed alot of them and are completely suppressing them.</p>
<p>I&#8217;ll have to <strong>email myself when generating 404 pages just to see if bots are crawling some non-existent pages.</strong></p>
<p>What really bugs me is out of say 724 pages, only around 250 are actually displayed in the SERPS.  Some of those hidden pages are supplementals.  But what about the rest of the pages?</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/03/06/google-cache.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Sitemap FAQ</title>
		<link>http://seo4fun.com/blog/2006/03/06/google-sitemap-faq.html</link>
		<comments>http://seo4fun.com/blog/2006/03/06/google-sitemap-faq.html#comments</comments>
		<pubDate>Mon, 06 Mar 2006 13:05:33 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[Google Sitemaps]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/03/06/google-sitemap-faq.html</guid>
		<description><![CDATA[This is my list of things about Google Sitemaps I&#8217;m pretty certain about.

&#8220;A Sitemap file can contain no more than 50,000 URLs and be no larger than 10MB when uncompressed.&#8221; Read more about Google Sitemaps.
Sitemaps are used to let spiders know about pages on the site that is hard for spiders to get to.  [...]]]></description>
			<content:encoded><![CDATA[<p>This is my list of things about Google Sitemaps I&#8217;m pretty certain about.</p>
<ul>
<li>&#8220;A Sitemap file can contain no more than <strong>50,000 URLs</strong> and be <strong>no larger than 10MB</strong> when uncompressed.&#8221; Read more about <a href="https://www.google.com/webmasters/sitemaps/docs/en/overview.html">Google Sitemaps</a>.</li>
<li>Sitemaps are used to let spiders know about pages on the site that is hard for spiders to get to.  <strong>Don&#8217;t just submit top-level directories.</strong></li>
<li><strong>Google also make use of the sitemap metadata.</strong>  <strong>Priority</strong> and refresh cycle are important in telling G spider how often to crawl the page.  Refresh cycle is useful in preventing Google from spidering the page too often.</li>
<li>Add your <strong>images and videos to the sitemap.</strong></li>
<li>Add <strong>RSS feeds</strong> to the sitemap.</li>
<li><a href="http://www.smart-it-consulting.com/internet/google/submit-validate-sitemap/"><strong>Validate your sitemap</strong></a>.</li>
<li>If you really need to know when the sitemap based crawling begins, then <strong>create a non-indexable page (having a NOINDEX robots META tag) which is not linked from anywhere, and put its URI in your sitemap.</strong></li>
<li>Use <strong>relative paths for ErrorDocument</strong> to enable verification.  Relative paths do not redirect. Read more about <a href="http://64.233.179.104/search?q=cache:GSmngU2mxrAJ:www.smart-it-consulting.com/article.htm?node=133&#038;page=101+google+sitemap+errordocument&#038;hl=en&#038;gl=us&#038;ct=clnk&#038;cd=1">sitemap verificaion</a>.</li>
<li><strong>Google Sitemaps will not work with subdomains.</strong></li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/03/06/google-sitemap-faq.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Google Sitemap Scripts</title>
		<link>http://seo4fun.com/blog/2006/03/06/google-sitemap-scripts.html</link>
		<comments>http://seo4fun.com/blog/2006/03/06/google-sitemap-scripts.html#comments</comments>
		<pubDate>Mon, 06 Mar 2006 12:34:58 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Google]]></category>

		<category><![CDATA[Google Sitemaps]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/03/06/google-sitemap-scripts.html</guid>
		<description><![CDATA[I&#8217;m in the habit of writing own scripts when I can, and since I hear alot of positive things about Google Sitemaps, and I didn&#8217;t find a script I liked out there, I decided to write a script to crawl one of my domains.  Right now, it&#8217;s site-specific, since I&#8217;m excluding certain paths (e.g. [...]]]></description>
			<content:encoded><![CDATA[<p>I&#8217;m in the habit of writing own scripts when I can, and since I hear alot of positive things about Google Sitemaps, and I didn&#8217;t find a script I liked out there, I decided to write a script to crawl one of my domains.  Right now, it&#8217;s site-specific, since I&#8217;m excluding certain paths (e.g. AVI, RSS, #, and certain directories), but I should be able to modify it so I can use it for all of my domains. Once I add an UI, anyone should be able to use it to generate a sitemap.</p>
<p>One thing I overlooked is that the generated sitemap XML is static, so <strong>lastmod will always be outdated</strong>. An article suggested making the XML file dynamic. For example, hardcode  &lt;? echo $lastmod ?&gt; and re-evaluate $lastmod in the XML file header.<br />
P.S. So far, Google seems to like the dynamic sitemap I created. There&#8217;s one URL that&#8217;s broken; Not sure where its coming from&#8230;since I deleted that link from the site; it could be from Google cache. Now I have no clue how to track down this broken link, since I never saw the error using Xenu.</p>
<p>By the way, To <strong>enable PHP parsing of XML files:</strong></p>
<ol>
<li>use: AddType application/x-httpd-php .php .htm .html .xml Note: You can do the same for .rss files and generate them dynamically.</li>
<li>echo the XML header, which has xml with  &lt;? ?&gt; which will be parsed as PHP.</li>
</ol>
<p>One annoyance is XML with PHP parsing enabled won&#8217;t display correctly in NN or FF.</p>
<ol><strong> </strong></ol>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/03/06/google-sitemap-scripts.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>HTACCESS Code (General Purpose)</title>
		<link>http://seo4fun.com/blog/2006/03/06/htaccess-code-general-purpose.html</link>
		<comments>http://seo4fun.com/blog/2006/03/06/htaccess-code-general-purpose.html#comments</comments>
		<pubDate>Mon, 06 Mar 2006 11:58:44 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Mod Rewrite]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/03/06/htaccess-code-general-purpose/</guid>
		<description><![CDATA[This is basically what I use to manage dynamic URLs.

RewriteCond %{QUERY_STRING} ^(.+)$ [NC] RewriteRule ^(.*)$ - [F,L] Protects against bogus query strings attached to the end of a URL. This code is customized to issue a forbidden instead of a 404.
RewriteCond %{REQUEST_URI} ([a-z0-9]+)\.html$ [NC]
RewriteRule ^([a-z0-9]+)\.html /script.file?p_path=$1 [L] This allows dynamic xyz.html under a directory and [...]]]></description>
			<content:encoded><![CDATA[<p>This is basically what I use to manage dynamic URLs.</p>
<ul>
<li><strong>RewriteCond %{QUERY_STRING} ^(.+)$ [NC] RewriteRule ^(.*)$ - [F,L]</strong> Protects against bogus query strings attached to the end of a URL. This code is customized to issue a forbidden instead of a 404.</li>
<li>RewriteCond %{REQUEST_URI} ([a-z0-9]+)\.html$ [NC]<br />
RewriteRule ^([a-z0-9]+)\.html /script.file?p_path=$1 [L] This allows dynamic xyz.html under a directory and block any links to xyz.html%20, xyz.html^20, etc.</li>
<li>I can also cover multiple &#8220;root files&#8221; (urls that are visible to visitors and should not take query strings):</li>
</ul>
<blockquote><p>RewriteCond %{QUERY_STRING} ^(.+)$ [NC]<br />
RewriteCond %{REQUEST_URI} ^/dynamicpage1\.html<br />
RewriteCond %{REQUEST_URI} ^/dynamicpage2\.html<br />
RewriteRule ^(.*)$ - [F,L]</p></blockquote>
<ul>
<li><strong>RewriteRule ^([^/]+)$ http://www.domain.com/path/$1/ [R=301,L]</strong> 301 redirect for path without /. I use this in cases where urls end in / (as opposed to say .html). Just a rewrite may work but use this if you’re worried about duplicates.</li>
<li><strong>RewriteRule ^([^/]+)/$ /encryptedfile04ha8fksdasd.html?query=$1 [L]</strong> This rewrites URL to the actual dynamic URL. The html file name is encrypted to make it difficult to link directly to the actual dynamic page.</li>
</ul>
<p>Additionally, in rare cases where I can&#8217;t catch bad query strings externally or I need to scan query strings for bad query values, I use this PHP 5+ code, where redirectfunction() is a custom function that will generate a 404 page:</p>
<blockquote><p>foreach($_REQUEST as $key => $value) {<br />
$$key = $value;<br />
if(check for values) redirectfunction();<br />
}</p></blockquote>
<p>Last thing: In addition to encrypting dynamic pages, you might want to block them via robots.txt to prevent it from getting them indexed. Of course, doing so can expose that url to your competitors, so personally, I would just make sure no internal links point to the actual dynamic page.</p>
<p>More .htaccess / mod_rewrite links:</p>
<ul>
<li><a href="http://httpd.apache.org/docs/2.0/mod/mod_rewrite.html">Apache Module mod_rewrite</a></li>
<li><a href="http://web-sniffer.net/">HTTP Request and Response Header</a></li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/03/06/htaccess-code-general-purpose.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>META Tag Analyzer Check List</title>
		<link>http://seo4fun.com/blog/2006/03/06/meta-tag-analyzer-check-list.html</link>
		<comments>http://seo4fun.com/blog/2006/03/06/meta-tag-analyzer-check-list.html#comments</comments>
		<pubDate>Mon, 06 Mar 2006 11:41:02 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[SEO]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/03/06/meta-tag-analyzer-check-list/</guid>
		<description><![CDATA[General things I look for when I run a page through a meta tag analyzer.

Title/descriptions/keyword are unique ENOUGH for each page. To be on the safe side, build pages with completely different titles/META descriptions with as few repeating words as possible.
Source code starts off with unique H1 and P 
100 or less links on the [...]]]></description>
			<content:encoded><![CDATA[<p>General things I look for when I run a page through a meta tag analyzer.</p>
<ul>
<li>Title/descriptions/keyword are unique ENOUGH for each page. To be on the safe side, <strong>build pages with completely different titles/META descriptions with as few repeating words as possible.</strong></li>
<li><strong>Source code starts off with unique H1 and P </strong></li>
<li><strong>100 or less links on the page</strong> - There&#8217;s nothing wrong with having more than 100 links on the page, but if you want transfer as much PR as possible per link, its best to limit the number of links per page.  Of course, if you have keyword rich outgoing links, this might balance out PR lost with excessive linking.</li>
<li><strong>At least 20kb+ file size.</strong> (I’ve seen 700kb pages indexed, but that’s not a whole lot of spider food, and a change in the filter may bump them into the supplemental).</li>
</ul>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/03/06/meta-tag-analyzer-check-list.html/feed</wfw:commentRss>
		</item>
		<item>
		<title>Handling Erroneous Dynamic URLS</title>
		<link>http://seo4fun.com/blog/2006/03/06/handling-erroneous-dynamic-urls.html</link>
		<comments>http://seo4fun.com/blog/2006/03/06/handling-erroneous-dynamic-urls.html#comments</comments>
		<pubDate>Mon, 06 Mar 2006 11:23:47 +0000</pubDate>
		<dc:creator>Halfdeck</dc:creator>
		
		<category><![CDATA[Mod Rewrite]]></category>

		<guid isPermaLink="false">http://seo4fun.com/blog/2006/03/06/handling-erroneous-dynamic-urls/</guid>
		<description><![CDATA[When creating dynamic pages, make sure you handle weird urls so that malicious linking doesn’t lead to the indexing of non-existent urls. Use HTTP header checker to see server responses. Catch all invalid requests with a custom 404 page. What URLS to check? Try to let .htaccess handle as much of the 404s as possible [...]]]></description>
			<content:encoded><![CDATA[<p>When creating dynamic pages, make sure you <strong>handle weird urls</strong> so that malicious linking doesn’t lead to the indexing of non-existent urls. <strong>Use HTTP header checker</strong> to see server responses. Catch all invalid requests with a custom 404 page. What URLS to check? Try to let .htaccess handle as much of the 404s as possible instead. PHP should be reserved for checking query values. You don’t want a chain of 301s where invalid URLs are returning 301s.  <strong>If the bad urls already exist, I would replace the 404s with 301s below.</strong></p>
<ul>
<li>/index.html => 404</li>
<li>/index.htm => 404</li>
<li>/?junksdso => 404</li>
<li>// => 404</li>
<li>/index =>404</li>
<li>/%20 =>404 (One time a guy linked me with this and the url got indexed)</li>
<li>/scriptfile.html => 404 Direct access to the file that generates dynamic content should generate a 404. Better yet, rename the file to something difficult to crack. <strong>Don’t forget to delete unencrypted versions of files on your server.</strong></li>
<li><strong>404 any invalid query values.</strong> For example, if a page produce unique pages for values 1-5, a value of 6 should 404.</li>
<li>If you’re <strong>not using unencrypted files: </strong>
<ul>
<li><strong>/validurl.html?junk=query =>404</strong></li>
<li><strong>/validurl.html?junk= should also 404.</strong></li>
</ul>
</li>
<li><strong>/path 301=> /path/</strong> Missing / for urls that end in / must 301 to url/. Yahoo will drop the /, so I can’t have that URL 404.</li>
<li>If a page is not dynamic, make sure page.html?bogusquery=string => 404</li>
<li>Capitalization should not matter. <strong>Use [NC] in Cond</strong> statements to avoid this.</li>
<li><strong>urls with +, _ or - should be handled correctly.</strong></li>
</ul>
<p><strong>Question:</strong> If I link to a url like this: index.html? (with nothing following the question mark) will Google index it?</p>
]]></content:encoded>
			<wfw:commentRss>http://seo4fun.com/blog/2006/03/06/handling-erroneous-dynamic-urls.html/feed</wfw:commentRss>
		</item>
	</channel>
</rss>
