Archives for category: Uncategorized

How did WordPress win over Movable Type? Majordomo has a good review of Six Apart’s strategic weaknesses and errors, and how WordPress/Automattic exploited each of them.

In my opinion, Six Apart’s biggest error was its choice of focusing on the enterprise market just as blogging was poised to go mainstream. This was an understandable mistake. They had investors to satisfy (not necessarily a bad thing) and there was no clear path to profitability for a software platform in the mass blogging market.

This lead directly to the Six Apart’s disastrous change in their license, which created WordPress’s moment of opportunity.

I was a happy Movable Type user at the time, but I was moving in other directions. I adopted Expression Engine for my next site, because MT didn’t have the membership management tools I needed. By the time I was ready to refocus my efforts on my MT site (MediaSavvy) in 2009, it was actually easier (by which I mean “possible”) to move it from an old version to MT to the latest WordPress than it was to upgrade to the latest version of Movable Type. Things had changed a lot in four years and the game was over.

The next time you justify a strategic decision by saying “That’s where money is”, you should remember that Willie Sutton was not a brilliant strategist.

Written in reply to a friend who asked for my thoughts on AOL/HuffPo.

My first thought on hearing this was “I hope they got cash, because I wouldn’t want to be a long-term investor in AOL.” Of course, they did.

I think HuffPo is a boring collection of gas-bag blogs. I was stunned by the price and startled when I looked them up on quantcast: They’re huge and growing fast, but page views per reader are declining.

So, long term, there are two kinds of content: the AOL/Yahoo/Demand Media/Rupert’s Daily dross, and the good stuff. The first type comes with a (problematic) business model. The second type has never been much of a business, but we continue to make it. Everything in the middle (me-too content not produced at scale) has no reason to exist.

Speaking as a competitor to Patch, I hope they continue using their cash to buy stuff.

Netflix is using an API, not to make their streaming services available to developers — which would cause all kinds of problems with their suppliers — and not for a handful of special projects.

Netflix has made their API a pillar of their multi-device distribution strategy. The other pillars are, suprisingly, HTML5 and Webkit (the open source Web browser engine). Instead of creating custom applications for each new device they want to be on, they’re porting Webkit to that device, if it’s not already there, and then building their custom application in HTML5.

The result, as I’m sure you know, is that Netflix is everywhere. Their customers can watch their movies on about 200 different devices. This video (from Mashery’s Business of API’s conference) changed the way I think about the impact of API’s on publisher strategy. I recommend watching starting at about 9 minutes into the video.

But their breadth of distribution is only one headline. The other headline is that they can rapidly change their appearance on all these devices quickly, with much less programming and a single platform: HTML + CSS + Javascript.

If Netflix can solve a distribution problem like this using standard Web technologies, publishers should reconsider non-standard solutions targeted at single platforms, such as iPads or iPhones. Right now, Apple is clearly dominant in non-PC distribution platforms, but we can expect rapid innovation from both hardware and software (Android, Windows, HP’s Palm, RIM, and more) manufacturers. Market fragmentation could become the norm in the near to mid term. And these new devices will be even better platforms for reading than for video.

Highly customized, device-specific “electronic magazines” and “digital newspapers” demo well and dominate the blog-driven news cycle, but they’re hardly the flexible platforms we’ll need in the next five years.

When mulitplatform distribution becomes the norm, we know that Netflix will be ready.

How many publishers will be ready?

It’s hardly a social strategy, but posting your site’s headlines to your Twitter feed and Facebook page is an important first step for a news site.

I’m pretty sure that most WordPress and Drupal users can do this with a plugin. Things aren’t so simple for users of Expression Engine or some other content management system. You may also want to use a third-party service if you’re reaching the limits of your current system. I’ve wrestled with this problem for year or two. I thought I’d share my experience and save you some time.

My first impulse, and probably yours as well, was to try get my Facebook page to pull in my feed. There’s a way to do it, but I defy you to discover it yourself, even though it’s accessible right there on the page. Once you know the non-obvious directions, it’s easy to do. But this is a very slow way to do an update. Facebook polls RSS feeds at long and seemingly random intervals. It can take hours for a new post to show up on your Facebook page. These days, this isn’t acceptable for J. Random Blogger, let alone a news site. And, of course, that doesn’t get your feed onto Twitter.

You need a third-party service to deliver your headlines as quickly and flexibly as your news cycle demands. Most of these services use the Pubsubhubbub protocol to know the moment you’ve updated your feed. I’m not going to go into a long explanation of how it works. I’m a little fuzzy on that to be honest. But, like the http protocol, you don’t need to know a lot to use it. Superfeedr will manage your push notifications for you. It won’t deliver your feeds to your social pages, but it will make possible for other services to do so more quickly.

Twitterfeed is designed will update social sites from your RSS feed. I found it to be reliable, nicely designed, flexible, and easy to use. It accepts Pubsubhubbub notifications. But its dashboard is a little buggy in my experience, and I’m not using it as much as I have in the past. aims to be a switching yard for social updates. The good news is that it can you can update it immediately via email, allowing you to be more selective about which stories appear on your social feeds. The bad news is that (1) can’t take RSS feeds directly, making it harder to automate your workflow, and recommends Twitterfeed for RSS notifications. (2) I found the exact format of updates to my Facebook page to be unreliable — they changed over time. This resulted in a Facebook page that was a visual mess and occasionally just stupid-looking when (or Facebook) picked up the wrong image to display with a story. And (3) Ping also seems to have gone into a developmental hiatus since Seesmic bought them last March. feels like a more polished solution for posting your feeds. It accepts Pubsubhubbub notifications, is easy to use, has a flexible system for adding feeds and social destinations, provides excellent reports on items delivered and clicked, does a better job of formatting Facebook entries, and allows to to filter your posts based on their content. It can also deliver RSS output, which I used to buffer a friend’s RSS feed which my CMS started polling him too frequently.

All three services (Twitterfeed,, and look good and deal creatively with the problem of putting a simple, Web 2.0-style interface on creating complex many-to-many relationships. and also offer API’s if you have the need and the skills to do something even more custom.

I recommend trying all three (using Superfeedr) However, I’ve found that is the easiest, fastest, and most powerful way to update my social feeds from my news site.

I’ve been experimenting for years with how to create and structure community news feeds. Readers want to follow breaking news and relevant information as it becomes available — and not just on Facebook and Twitter. All this time, I’ve been looking for an easy and reliable way to merge a bunch of related RSS feeds into a single feed that would make this easier for my readers.

What seems like a simple problem turns out to be difficult to do well. The problem is that it’s an easy enough problem that everyone thinks they can solve it, so no one is willing to invest what it takes to create a quality implementation. I’ve tried a lot of methods short of writing my own aggregator, which have no interest in doing.

Speed is an issue for all aggregators, because they must recognize that a feed has been updated, read the feed, add it to the merged feed, and deliver the new feed to its subscribers. This can slow things down, because neither the aggregator nor the end user can poll too often. Also, many of these services do not scale very well and slow to a crawl as soon as they become popular.

A lot of sites aim to create merged feeds for you. About half those you’ll find on most lists via Google no longer exist. So how can you count on any of the others to be there in the long run, or to be fast and reliable enough to use in a production system?

Yahoo Pipes is ineffably cool, and a pretty good solution, but it’s slow, kind of obtuse, and unreliable. It doesn’t appear to be getting a lot of attention from Yahoo, which is a good sign you shouldn’t make it part of your infrastructure.

The Simplepie RSS parser has a nice-looking website and lots of links to interesting examples, but the project seems dead in the water. I was able to find a good example of an RSS aggregator that did most of what I was looking for, modified it to make it my own, and added it to my site. However, maintenance became a problem. First, the script slowed down my host. Shortly after that, it started to get aggressive in its reading of my friends’ sites. It was becoming clear that unless I wanted to go from being a publisher to a programmer, that rolling my own aggregator from Simplepie simply took too much work.

I finally realized I could solve this problem using WordPress, of all things. In a couple of hours, I was able to create a news aggregator using WordPress and a plugin called FeedWordPress.

My new aggregator can pull in headlines and summaries from multiple sites and deliver a combined feed. The aggregator itself isn’t intended to be a destination — right now I’m the only person who even knows where it is. It’s just a bit of infrastructure that works reliably and fast. I’m able to read the merged site’s feed with the magpie reader built into Expression Engine and deliver the results as a local news digest.

FeedWordPress has a terrible name and an ugly website, but it works really well and has plenty of options for managing the way you gather and display feeds. And it’s shareware. I sent in my donation.

Steve Yelvington is right about the five sad reasons the America press is outraged by Wikileaks.

I’d add another:

American journalists are fundamentally conservative –they hate change. The conventions of American journalism are as ritualistic and metaphorical as a Japanese tea ceremony or diplomatic protocol. Wikileaks is outside their taxonomy and it makes their brains itch.

Daniel Ellsberg’s endorsement notwithstanding, the diplomatic cables are not the Pentagon Papers. And the Pentagon Papers currently define the outer edge of acceptable journalistic behavior.

Traditional American journalism has almost always been about telling the least objectionable story. Usually it serves the powerful, sometimes it serves the public at the expense of the powerful. It is seldom willing to alienate both the powerful and the public when both are on the same page.

This is an interesting description of how is delivering local headlines to CNN. It’s from Mashery’s Business of API conference in New York last month.

I’ve never thought too highly of local aggregators like but they’re doing the right thing and linking those headlines on CNN not to themselves, but to the originating sites.

As a community news publisher I appreciate it.

Sidebar: I can also appreciate the irony of embedding this video after proudly declaring myself to be (mostly) Flash-free. Flash is still an important part of the Internet’s plumbing, but I can tell you my Macbook Pro runs better when it’s not in my browser.

I removed Flash from my Mac. I couldn’t be happier.

As a user, Flash sucks up my resources, drains batteries, enables indelible cookies, and bypasses popup blocking.

With a lot of tabs open — which is how I work — Safari would become sluggish and intermittently unresponsive. Since I removed Flash, these problems have gone away because I no longer have a dozen Flash animations running in the background. The improved privacy and reduced interruptions are side benefits. If I absolutely need Flash, I can open Chrome, which has Flash built in. But I have FlashBlock installed there by default as well.

I’ve gone back and forth on Flash. For years, I said it was junking up the web. It never really stopped junking up the Web, but in the last five years Flash made it Web video a practical, so I cut it some slack. Before Flash, Web video was a nightmare of proprietary players, constant updates, tiresome visits to the RealPlayer’s home page for yet another attempt to trick you into buying their worthless merch.

As a publisher, Flash is a ticket to misery. I wasted countless hours producing Flash videos for Coastsider, trying to get Flash players to work on my site, and attempting to embed videos from other sites.

I’m done with all that. I recommend you be done with it to.

Get ready for the post-Flash world. Millions of iPhone, iPad, Android and other mobile users are already surfing Flash-free. The thought leaders in Web and software development are beginning to see the light as well. Expect uninstalling Flash to be a trend in 2011.

Resistance is likely to come from three sources:

  • Editorial personnel who confuse Flash development with Web development: Do them a favor and get them trained in something more future-proof.
  • Advertisers (agencies, actually) love Flash: It makes them feel like artisans. The best you can do with them it to make sure they provide ads in alternative formats for serving to non-Flash users of your site.
  • Web producers who see Flash as an easy way to create fancy user experiences: This is already less common in media than it is in the production of commercial sites for restaurants and hotels.

Those who are complaining the loudest about this trend have big investments in Flash development and technology. If that describes your situation, now might be a good time to think about what you’re willing to sacrifice to keep Adobe’s Flash relevant.

You’ll need to continue to use Flash for video as long as a reasonable number of users need it, but you should be preparing now to produce video that works with HTML 5. This will mean multiple formats in the short term. You can ease this by finding a video host that can deliver modern Web video to your users, and Flash when they demand it. I’ve moved all my videos to Vimeo and I’m not looking back.

A friend who runs a local mailing list on Yahoo Groups just got the following email. We live in interesting times.

From: “”


Yahoo! is creating an exciting new opportunity for San Francisco and San Jose area residents. We’re looking for neighbors who want to participate in their communities by getting out the word on what’s happening around them. Because you have led locally focused Yahoo! Groups, we thought our new initiative might be another way for you to communicate what you care about with others.

You can be among the first to contribute to this soon-to-launch neighborhood content destination by joining the conversation today. Depending on where you live, signup at either or After you sign up (which is free and carries no obligations) on Associated Content by Yahoo!, you’ll be given $10 writing assignments that will explain more about this project and how you can get involved!

To put this in perspective, $10 per story, for a Patch editor producing 2.5 stories/day, 6 days/wk, would yield an annual income of $7,800 — no benefits.

I came into this thinking Bob Cox needed to get a life.

Bob has now convinced me that his local Patch editor’s reuse of some mug shots from his site was plagiarism. I wouldn’t (and he didn’t) call it copyright infringement. But he added value to the photos, and the local Patch editor used them and didn’t credit him.

Looks like a goof by a newbie blogger, but it should have been acknowledged by that editor’s chain of command.

Bob has made some other claims that Ms. O’Connor has refuted, but he has also proven (at least to me) that the following statement by Patch regional editor Kathleen O’Connor is simply false:

any similarity to Mr. Cox’s presentation of those public images is purely coincidental. Linking mug shots together in Photoshop (in this case, apparently doing nothing more than placing three similar sized objects in a row) is standard operating procedure for news organizations everywhere.

Patch is a large, new, decentralized operation. Everyone in it is already working too hard. There will be mistakes. But how Patch deals with those mistakes will be critical to its reputation as a news organization.

Full disclosure: I reused a mug shot from the SF Examiner just last week.  I had no problem with reusing the image, which is public property, but I made sure to credit the Examiner as the source and link to the story. After all, who wouldn’t do that?