MySpace Tom’s Words of Wisdom on Google+ and Facebook’s Future

July 8, 2011 by Eric Lander · 9 Comments

New to my blog? You may want to learn more about me, subscribe to my RSS feed or friend me on Facebook. If you're looking for an online marketing consultant or just want to communicate offline, please use the contact information available here on the site.

I never thought I’d come back to the blog after a two year, one month and nineteen day hiatus to blog about Tom, the awkward looking friend we all popped our MySpace cherry with. But I have, and with that said, I’d like to buy Tom Anderson a beer for providing some interesting perspective on Facebook, Google+ and the topic of social media innovation. [Read more...]

spacer

Filed Under: Social Media ·

Google: The Internet’s Organized Crime Family

May 20, 2009 by Eric Lander · 41 Comments

Late last week I found myself watching a documentary on organized crime in New York City. Focused on the Gambino crime family, this documentary illustrated how the mafia (which included the Gambino and other families that made up the Cosa Nostra) had their hands in on nearly every facet of city life. From restaurants and delivery trucks to corrupting the law enforcement and legal systems – the organization was in control. [Read more...]

spacer

Filed Under: Search Optimization & Marketing ·

WebmasterRadio’s SEO 101 Program

April 3, 2009 by Eric Lander · 4 Comments

Update: After just three short weeks of co-hosting the program alongside Ross Dunn, I’ve resigned as a host and wish the program the best of luck. My decision not to be part of the program came after what I felt was a very questionable production of SEO 101 which aired immediately following David Szetela’s interview of Guy Kawasaki on Monday April 20. After listening to the show after it aired, I realized that the show was not the best fit for me.

My plans for 2009 continue to become more exciting. I just got word back from Ross Dunn, and the folks at WebmasterRadio.FM have made it official… I’ll be Ross’ new co-host on the SEO 101 program that airs each Monday at 5:00PM ET.

What is SEO 101?
SEO 101 is weekly program that speaks to Search Engine Optimization, or “SEO” from the very beginning. The goal of the SEO 101 program will be to interact and teach listeners about the many aspects of SEO by sharing helpful information without overwhelming them with technical details.

Joining Up with Ross Dunn and WebmasterRadio.FM
I met Ross back in February of 2007 at SMX West and he’s a great guy, great SEO and a wonderful father. If you aren’t already keeping tabs on what Ross is doing – please connect with him on Twitter and check out his company, StepForth Web Marketing Inc..

spacer Here’s a photo of [L to R] Darrell Long, myself, Ross and David Wallace at Bennigan’s last February in Santa Clara. [Photo Credit to David Wallace]

Ross and I have obviously been pretty excited about this, and here’s his take on the development:

SEO 101 on WebmasterRadio.fm has been a big success since Jennifer Evans Laycock and I took it on and revived it but when Jennifer left due to time constraints I needed someone to fill some big shoes. Thankfully I have been around long enough to know just how good a guy and SEO Eric Lander is. So when he agreed to be my co-host on the show it felt great to know the show was once again whole. Now it is time for us to plan out some great shows!!

To compound on that, I also want to thank Jennifer for her contributions to the show (and our industry) and wish her the best with all that she has going on. We’ll certainly try to steal her back and get her back on the air for our listeners, too. :)

About Webmaster Radio
spacer Founded in November, 2004 – WebmasterRadio.FM is a free, 24/7, interactive Internet Radio Network focused on the B2B digital marketing world. We have burst on to the Internet Media scene as a community destination with 100% original programming, live broadcasts, archives, and podcast shows. We are proud to boast that our shows are hosted by the most respected names in the business world!

    WebmasterRadio.FM Links of Interest:

  • WebmasterRadio.FM Website
  • SEO 101 Program Details
  • Searchbash
  • Advertising on WebmasterRadio.FM
  • WebmasterRadio.FM’s Blog
  • WebmasterRadio.FM on Twitter
  • WebasterRadio.FM on Facebook

spacer

Filed Under: Links & Resources, Search Industry Conferences ·

An Introduction to Log File Analysis for SEOs & Webmasters

March 30, 2009 by Eric Lander · 35 Comments

It doesn’t matter what sort of web analytics suite you choose to use… If you’re not actively reviewing your log files, you’re missing out on some key data and reporting metrics. Today’s blog post is all about log files and how you can begin using them to learn more about what is happening on your web site.

My goal here is to encourage you to the point of using log files on a regular basis to better tune up your hosting and marketing platforms.

Page Tagging Vs. Log Based Analytics

When you think about web analytics, what do you really think of? The majority of people I know have told me all about their nicely designed dashboards filled with telling graphs, growth charts and tables that just wait online for them to view them whenever needed. While that’s all well and good – those reports are always coming out of page tagging analytic programs. The problem? Page tagging analytics has limitations, and some of those limitations are simply unacceptable for hardened SEOs and webmasters.

That’s not a knock on page tagging analytics either. Page tagging is a popular method of acquiring data as the ease of use and on demand availability of reporting all add up to make these tools a required resource. Log files for me simply help me go that extra mile.

I want you to guess how many of the following analytic suites are providing you with reports and data generated from, in part, log files:

  • Google Analytics
  • Omniture
  • Microsoft adCenter Analytics
  • Sitemeter
  • Quantcast
  • Compete
  • HitWise

Ready for the obvious answer?
– Zero! None of the above use data recorded by your server for statistical analysis.

Olivier Amar of CompuCall earned some kudos this morning. When I asked how many followers were not out there checking their logs – he tweeted a reply about ClickTracks – one of the few analytic suites out there for SEOs and site owners that actually integrates log files out of the box.

I don’t want to get into a whole lecture about the differences between page tagging analytics and log parsers (or hybrid solutions for that matter). What I do want you to realize is that no matter the hosting platform, there is some more useful information you could be extracting about your web site and your visitors if you can acquire the logs.

Familiarize Yourself with Log Files

Before we jump too far in it’s probably best for us to review what a server log file is, what it looks like, what data it contains, etc.

What is a Server Log File?
Wikipedia defines a server log as:

A server log is a log file (or several files) automatically created and maintained by a server of activity performed by it.

A typical example is a web server log which maintains a history of page requests. The W3C maintains a standard format[1] for web server log files, but other proprietary formats exist. More recent entries are typically appended to the end of the file. Information about the request, including client IP address, request date/time, page requested, HTTP code, bytes served, user agent, and referer are typically added. These data can be combined into a single file, or separated into distinct logs, such as an access log, error log, or referrer log.

How Do I Retrieve Server Logs?
Each hosting provider or company handles this differently. My hosting company makes it easy for me by keeping logs available via FTP on a 7 day cycle before any logs are removed. I have adapted to just pull those logs down off my server oce a week through an automated application. Set it up once, and now I can forget about the hassle.

I’ve seen other hosts that make log acquisition more… Trying. In any event, server logs have a number of different recording options, structures, formats and file types. This post is focused on using the logs you have available to you – not acquiring them. I highly recommend working with your server administrator or hosting provider to acquire access to logs if you do not have that already.

If you are ever presented with an option – push to acquire Extended Log Files and then quickly hand your hosting provider or server admin a copy of this resource from the W3C.

What Does a Server Log Look Like?
Here are five lines I pulled out of a server log file from my blog as recorded yesterday, March 29, 2009:

85.89.185.215 – - [29/Mar/2009:01:00:09 -0700] “GET /wp-content/uploads/2007/09/100cap006.jpg HTTP/1.1″ 200 46012 “www.ironworksforum.com/forum/showthread.php?p=1200167″ “Mozilla/5.0 (Windows; U; Windows NT 6.0; en-US) AppleWebKit/530.1 (KHTML, like Gecko) Chrome/2.0.169.1 Safari/530.1″

38.99.107.141 – - [29/Mar/2009:01:00:10 -0700] “GET /feed HTTP/1.1″ 302 5 “-” “Mozilla/5.0 (compatible; FriendFeedBot/0.1; +friendfeed.com/about/bot)”

193.252.149.15 – - [29/Mar/2009:01:07:24 -0700] “GET /276.html HTTP/1.1″ 200 24655 “-” “Mozilla/5.0 (Windows; U; Windows NT 5.1; fr; rv:1.8.1) VoilaBot BETA 1.2 (support.voilabot@orange-ftgroup.com)”

69.147.112.169 – - [29/Mar/2009:01:11:01 -0700] “GET /feed/rss HTTP/1.0″ 302 0 “-” “Yahoo Pipes 1.0″

66.249.72.136 – - [29/Mar/2009:01:12:28 -0700] “GET /robots.txt HTTP/1.1″ 200 508 “-” “Mozilla/5.0 (compatible; Googlebot/2.1; +www.google.com/bot.html)”

In a 12 minute span on my web site, some pretty cool things happened that I would never have known about through any page tagging analytic suites I’m using.

Let’s dissect each of these five lines and I’ll show you what I mean.

Entry #1

85.89.185.215 – - [29/Mar/2009:01:00:09 -0700] “GET /wp-content/uploads/2007/09/100cap006.jpg HTTP/1.1″ 200 46012 “www.ironworksforum.com/forum/showthread.php?p=1200167″ “Mozilla/5.0 (Windows; U; Windows NT 6.0; en-US) AppleWebKit/530.1 (KHTML, like Gecko) Chrome/2.0.169.1 Safari/530.1″

Someone is stealing my images! Like a lazy webmaster, I don’t lock down much and prevent other people from using it. In this referenced log file I now have evidence of someone using one of my images (regularly, I might add) on another web site’s discussion board.

This log file entry tells me that this person is using this image on this discussion thread.

Not cool! Now for me, bandwidth isn’t much of an issue and I don’t really mind if someone is repurposing that image. If that were protected photography though – I’d want to keep it under lock and key. More on this later.

The key here though is that the actual “page” being loaded up (the discussion board thread or user profile page) is hosted elsewhere. Since I don’t own that site, I don’t have Google Analytics code on the site and without this log file, I never would have known that this was taking place.

When you consider how much this could happen with a large web site – you can probably see how quicly this can become a big issue.

Entry #2
The next log file entry was this:

38.99.107.141 – - [29/Mar/2009:01:00:10 -0700] “GET /feed HTTP/1.1″ 302 5 “-” “Mozilla/5.0 (compatible; FriendFeedBot/0.1; +friendfeed.com/about/bot)”

As the tail end may suggest to you, this is a FriendFeed bot that’s coming through my web site and pulling a copy of my blog’s feed. Friendfeed’s bot will then see if there’s any new entries and pull them via RSS to use on their own site since I’ve allowed them to do so.

If you’re watching things like page views, this wouldn’t actually count in other analytics since again – the user requesting the data never actually came to my web site. The other issue? The “user” here is actually a bot and my guess is that if it’s like GoogleBot, it probably won’t bother to execute any javascript code that would be required for page tagging analytics to record the hit.

Entry #3
Next up:

193.252.149.15 – - [29/Mar/2009:01:07:24 -0700] “GET /276.html HTTP/1.1″ 200 24655 “-” “Mozilla/5.0 (Windows; U; Windows NT 5.1; fr; rv:1.8.1) VoilaBot BETA 1.2 (support.voilabot@orange-ftgroup.com)”

An old post on my blog on the Internet Marketer’s Charity Party at SES San Jose is being retrieved here by another bot, this time called VoilaBot. Ever heard of VoilaBot before? Sadly, I had not – which is more telling about my failures as an International SEO.

Voila is the provider for Wanadoo, which is a huge portal in France and one of the biggest european ISPs.
Voila itself is one of the best known web brands in France.

Where’d I get that information? From heini, a veteran user on Brett Tabke’s WebmasterWorld, silly.

Entry #4
Still with me? Good, because we’re going to go easy on these last two entries to review! Next is…

69.147.112.169 – - [29/Mar/2009:01:11:01 -0700] “GET /feed/rss HTTP/1.0″ 302 0 “-” “Yahoo Pipes 1.0″

This is the footprint of Yahoo! Pipes, a fairly new RSS / News Aggregator that’s actually quite cool. All that was happening here is that a user of the Pipes program was loading up (or refreshing) my blog’s RSS feed. Again – this would never show up in anything like Google Analytics or Omniture. Why not? You know this. Just read the last three log dissections. :)

Entry #5
And finally… The staple of any SEOs diet… Googlebot!

66.249.72.136 – - [29/Mar/2009:01:12:28 -0700] “GET /robots.txt HTTP/1.1″ 200 508 “-” “Mozilla/5.0 (compatible; Googlebot/2.1; +www.google.com/bot.html)”

A well behaved Googlebot, too! The above request is the mark of GOOG coming through and requesting my blog’s robots.txt file for some more direction. It’s always nice when bots do what they say and are supposed to do first, right?

Now, Onto YOUR Log Files…

You don’t really care about what’s happening here on my blog — you want to see what’s going on with your web site. So now we get to take a look at how to make these log files work for you!

What You’ll Need
1.) Server Log Files
2.) Server Log Parsing Application
3.) Curiosity

Again, I’m not helping you with item number one.

With item number two, I’d recommend WebLog Expert. It’s an application that I’ve been using for years and bought the professional version of some time ago. Considering the low cost, I’d recommend it – but there are certainly other log file analyzers available to you.

Just check out download.com or directory listings on the Yahoo! Directory or on DMOZ.

Since WebLog Expert offers a free BETA version with some filtering options though, I’ll use them for screen shots.

Here are some report ideas I’m going to demonstrate for you…

  1. Google, Yahoo & LiveSearch Spidering
  2. Stolen Content
  3. 400 Errors
  4. 300 Server Redirections

For the purpose of this demonstration I’m going to use WebLog Expert as the log file analyzer because it’s a free solution and provides some easy filtering options. The key here is in using these filters to look at very specific data.

Google, Yahoo & LiveSearch Spidering

Log files record the user agent of each request. When a human visitor visits your site, their web browser will be recorded as it is labeled. Refer back to log entry #5 from above to see how Googlebot identifies itself. It tells my web server that it’s user agent is Mozilla/5.0 (compatible; Googlebot/2.1; +www.google.com/bot.html).

In order to report on the spiders, you need to set up a filter that excludes all activity outside of the spiders. I’m going to take this one step further and show you how to set up filters in WebLog Expert that only pull activity on the big three – Google, Yahoo and LiveSearch.

Using the filters dialogue, you will need to add a new filter that includes activity based on spider name, and then select each of the appropriate spiders from the drop down list. To do this, you’ll need to set up three filters like so:

spacer

spacer

Once you set up three filters, one for each, you should see this:

spacer

And if you do, just click through the Finish button and then run you’re report. You’ll now get a wealth of information on your spidering activity.

Want to see what data is available? Click here to download the resulting report in PDF format. Here’s a hint to what you may find out… Yahoo! Slurp is sometimes a little more… aggressive than you may think:

spacer

Stolen Content

This report, even though I’m using it as an example for this post – really is one I need to take action on. Our goal here is just to find anyone out there who may be using my CSS or images on their own sites or for their own needs. If you were to run this same report, I’d suggest that as an action step – you take measure to prevent images from being hotlinked and so on.

With WLE, you’ll want to create the following filters:

spacer

Replace ericlander.com with your own domain, and, add or subtract any files you’d like to see in there. Other popular files to be stolen and reused? mp3, pdf, swf, avi, mpg, mov, and css lead the way for me.

Again, a sample report output of the above can be found here in PDF format.

400 Errors

One of the most useful reports for me over the years has been this report that only looks at 400-type responses. Now, any 400 error from your server indicates that something hasn’t been found. The most popular of which is the 404 error we’re all used to seeing – but there are other useful not found errors to note, including the following table from HTML Goodies:

  • 400 : There is a syntax error in the request. It is denied.
  • 401 : The header in your request did not contain the correct authorization codes. You don’t get to see what you requested.
  • 402 : Payment is required. Don’t worry about this one. It’s not in use yet.
  • 403 : You are forbidden to see the document you requested. It can also mean that the server doesn’t have the ability to show you what you want to see.
  • 404 : Document not found. The page you want is not on the server nor has it ever been on the server. Most likely you have misspelled the title or used an incorrect capitalization pattern in the URL.
  • 405 : The method you are using to access the file is not allowed.
  • 406 : The page you are requesting exists but you cannot see it because your own system doesn’t understand the format the page is configured for.
  • 407 : The request must be authorized before it can take place.
  • 408 : The request timed out. For some reason the server took too much time processing your request. Net congestion is the most likely reason.
  • 409 : Conflict. Too many people wanted the same file at the same time. It glutted the server. Try again.
  • 410 : The page use to be there, but now it’s gone.
  • 411 : Your request is missing a Content-Length header.
  • 412 : The page you requested has some sort of pre-condition set up. That means that If something is a certain way, you can have the page. If you get a 412, that condition was not met. Oops.
  • 413 : Too big. What you requested is just too big to process.
  • 414 : The URL you entered is too long. Really. Too long.
  • 415 : The page is an unsupported media type, like a proprietary file made specifically for a certain program…

The filter setup here is super simple. Just create this one:

spacer

And the resulting report looks like this (again, PDF!).

300 Redirections

Every SEO needs to have a grasp of 301 redirects, and reporting on the ones your server dishes out is super simple here. Just like the 400-responses, you’ll need to set up a quick filter that only pulls 300-level response codes. Easy!

spacer

The value here for an SEO is pretty obvious – so I’ll let you run with why this report is useful. To check out the sample in PDF format, just click here.

Wrapping Up…

Hopefully this post has given you some more insight on how you can begin analyzing server log files. If I’ve confused you at any point, please do drop a comment below and open up a discussion for us as others may have similar questions or hangups.

Don’t be afraid to get creative with the use of filters too with WebLog Expert or any other application that you may find yourself using. It’s very easy to use filters to extract in depth metrics like time spent on site by visitors viewing movies, path of visits referred from Digg, bounce rate for StumbleUpon referrals, etc.

Finally, this isn’t meant as a knock on page tagging analytics and the information they offer. Every successful web site marketer should rely on both regularly – but when it comes to running a clean site, don’t just assume the logs have nothing to provide to you.

spacer

Filed Under: Analytics ·

Entrepreneur Author Gives Poor Local Search Advice

March 28, 2009 by Eric Lander · 7 Comments

The Spring 2009 issue of Entrepreneur Startups features an article written by Kim T. Gordon called Big Results on a Small Budget. The focus of the short article is in providing small business owners with four marketing suggestions that are tailored to those on a tight budget.

While I was certainly happy to see search marketing involved, I don’t much agree with Kim’s suggestion:

Use local paid search. When your prospects search online, who will they find first, you or your competitors? The best way to guarantee your company appearing at or near the top of search results is through pay-per-click advertising. Being at the top of the page is critical to getting noticed, as most searchers rarely go past the first page of results. And the good news is that local paid search is often quite affordable and is readily available through Google, Yahoo and other search engines.

Paid search is an affordable option for those looking to drum up business. If you’re on a tight budget though, why not look at spending no money to capture that same audience?

I often fear that too many all in one marketing types simply suggest paid search because it’s effective before giving other forms of search a solid look. As enhanced listings mature and search engines’ users adapt their patterns to use these universal results – more free opportunities for listings appear.

If you were to perform a search on a major engine for any keyword phrase that combines geographical modifiers alongside a product or service name – you’re likely to see something like the Google OneBox results. It’s that familiar map with nice red icons that plot the nearest vendors that provide exactly what you’re after.

Best of all – Few businesses are out there taking the time to claim and enhance these free listings. That leaves you with some great opportunities to to clean up without spending a dime on search referrals.

My other issue with the author’s suggestion of local paid search is that she emphasizes reaching your targeted audience. Going into any paid search marketing campaign blindly will cost you far more than necessary. Paid search campaigns are optimized over time by evaluating clickthroughs, conversion rates and so on. You can’t afford the luxury of buying that information when you’re truly on a tigh budget.

Now, to the author’s credit, advertising in Google’s Local search results is certainly affordable. In my experience the analytical data for these campaigns suggest that having a unique icon on the map along with a more enhanced business profile will instigate more productive clickthroughs.

My last gripe is that no one reading the referenced article would ever know that such a search advertising product even exists. Furthermore, even basic local listings can be pushed offline and result in your customers generating favorable reviews, relying on Google for coupons, and ultimately generating more free business opportunities for you.

I’ve already gone through how to make the most of your local search profiles on Google in an old article on Search Engine Journal. If you’re interested, please check that out along with David Mihm’s local search ranking factors research.

I’m not against renting space. I just don’t understand why you’d want to rent space you could own for free.

spacer

Filed Under: Local Search Optimization, Search Optimization & Marketing ·

Video Feedback of Our Affiliate Summit Panel

February 18, 2009 by Eric Lander · 2 Comments

Really, videos like this prove to me precisely how important it is to go to conferences and represent your skill set and organization accordingly.

Ironically ADP gets a good plug here – as they did throughout the conference materials. Too bad they didn’t see the trip as being important enough to cover. I paid my way and was forced to also use my vacation time.

Oh, and hat tip to Oilman for posting this on his blog ages ago. Since he and I are a lot alike (postings once per millennium) you should thank him. Not me. :)

spacer

Filed Under: Search Industry Conferences ·
« Older Posts
Newer Posts »
gipoco.com is neither affiliated with the authors of this page nor responsible for its contents. This is a safe-cache copy of the original web site.