Lee LeFever reflects on the CommonCraft style

On this week’s Innovators show I spoke with Lee LeFever who, with his wife Sachi, runs CommonCraft. You’ve probably seen their clever animated videos explaining web topics like RSS and Twitter. If not, check out their most popular one, RSS in Plain English. It’s uncannily effective, and a real inspiration to those of us who’ve struggled to explain this concept to civilians.

As I mentioned here, and as we also discuss in the interview, the CommonCraft style shares some DNA with the practice of paper prototyping as described by Bill Buxton in another interview. In Bill Buxton’s case, you abstract away unnecessary detail, and convey action, by sketching and manually animating parts of the interface you’re designing. In Lee’s case you use the same methods downstream, not to elicit the key concepts to be embodied in the yet-unbuilt software, but instead to explain those concepts after it’s shipped. Two sides of the same coin, I think.

A conversation with Tony Hey about Microsoft External Research and the new breed of e-scientists

In my Perspectives series I’ve been exploring some key relationships between Microsoft and external partners: Mercy Corps, NASA, the Library of Congress, the community of scientific and scholarly publishers, and individual scientific researchers including George Hripcsak and Dennis Baldocchi.

For this week’s show I spoke with Tony Hey, the VP of the External Research division of Microsoft Research. More than anyone else, he’s responsible for advancing what I’ll call the Jim Gray vision: All science is becoming computational and data-driven; a new breed of scientists requires a new generation of tools and services; Microsoft can and should rise to the challenge.

In coming weeks and months I’ll be reporting more of these stories of collaboration between Microsoft Research and scientific researchers. This is a treat for me because, little known fact, during the early part of my software career I took a hiatus to earn a graduate degree in science writing. I’ve always been curious about how science works, and this is an opportunity both to learn and to share what I learn.

It’s also an opportunity to expore, and to highlight, a side of Microsoft that you’ll never read about on TechMeme. When Tony Hey recently counted up the number of Microsoft Research engagements with university partners, the total exceeded 400. It’s an impressive outreach effort, and one that will benefit everyone: Microsoft, the scientists whose research work is enhanced by these projects, and all of us who depend on the outcomes of that research.

Overcoming synthetic voice shock

To prepare for my interview with Susan Gerhart I tried using text-to-speech software to read menu choices and text selections aloud. As always, I experienced the reaction that Susan, in her latest post, calls synthetic voice shock.

For those of us who don’t need to rely on synthetic voices, that reaction isn’t a problem, it’s merely a deterrent to optional use of the technology. For example, though it might be convenient to shift some material from the domain of written text to the domain of audio, the unpleasantness of synthetic voices stops me from doing that.

But the real problem, Susan explains today, is that synthetic voice shock deters people who have lost their vision, and who would benefit greatly if they could adapt to those voices. Here’s how she characterizes typical reactions:

  • I cannot understand that voice!!!
  • The voice is so inhuman, inexpressive, robotic, unpleasant!
  • How could I possibly benefit from using anything that hard to listen to?
  • If that’s how the blind read, I am definitely not ready to take that step.

She adds:

Conversely, those long experienced with screen readers and reading appliances may be surprised at these adverse reactions to the text-to-speech technology they listen to many hours a day.

How can we help people cross that chasm? Susan offers advice to four groups: “vision losers”, developers of assistive reading technologies, sighted people who are helping vision losers, and rehab trainers.

I’m in the third group. My mom’s macular degeneration is progressing, and although she’s not yet forced to rely on text-to-speech, that day may come. To those of us in this group, Susan recommends that, when evaluating applications and appliances, we need to bear in mind that voice quality is a separable concern, not directly tied to the capabilities of the software and hardware. And she suggests that, in order to help friends or family members, we might want to develop some familiarity with the range of available voices.

To that end, Susan has provided audio renderings of her blog posts, including four different versions of today’s post as read by Neospeech Kate, Neospeech Paul, Microsoft Mike, and Robotic UK Jane. None of the readings is pleasant to listen to. But Susan says:

I testify it takes a little patience and self-training and then you hear past these voices and your brain naturally absorbs the underlying content.

Those of us not compelled to learn how to “hear past” those voices might still want to try the experiment, in order to help friends and family members make the transition.

A conversation with Joshua Tauberer about GovTrack

By an odd coincidence, my interview with Joshua Tauberer, founder and operator of GovTrack, went live just as the bill I’d been tracking using that system made headlines. Had I not been doing some investigation into legislative support for alternative biomass heating systems, I’d never have known that the Foreclosure Prevention Act of 2008 — which the President is now expected to sign into law this week, and which is generally known as the bailout for Fannie Mae and Freddie Mac — is also the bill that may result in tax credits for those of us in New England who’ve been investing in those heating systems.

This has been my most comprehensive attempt to open the lid of the legislative sausage factory, peer inside, and try to understand how a specific interest of mine was being processed. What I found is that, even with power tools like GovTrack and MAPLight, it’s really hard to make those connections. That’s partly because we lack good mechanisms to track the flow of bits of legislative language through an evolving assortment of bills, and to relate those fragments to the activities and interests of their sponsors.

But it’s also because a novice who tries to read and interpret this record lacks context. I had learned along the way, for example, that one of my senators cast a crucial NO on a cloture vote that would have amended the Energy Independence and Security Act of 2007 to include $22 billion of renewable energy tax incentives — including, as one small component, the “pellet stove” provision that has now landed in the Foreclosure Act. But I wasn’t at all clear on the significance, or recent evolution, of the parliamentary maneuver known as the cloture vote.

As Mike Caulfield points out here, there’s been a recent and dramatic increase in the use of this maneuver. As a result, “what your individual Senator does on bills matters quite a lot less than you think.”

The emerging breed of Congress-tracking tools and services can’t yet capture these nuances in ways that enable an ordinary citizen, motivated by some personal interest, to dive into the process and bootstrap an understanding that will guide voting or advocacy. Recognizing this, Joshua Tauberer explains in our interview that he’s currently exploring ways to crowdsource that bootstrapping effort. So for example, the site has a feature called “Users tracking this bill are also tracking…” Because I’m tracking both H.R. 6 (Energy Independence and Security) and H.R. 3221 (Foreclosure Prevention), the connection I made between these two seemingly unrelated bills creates a link for others to follow.

Mike writes:

As more people delve into these tools, the conversation around how our process really works can occur. And hopefully it can start to occur in channels outside the poliwonkablogosphere that I love so much but that others, well, not so much.

The first reaction to tools in the hands of novices is always that the novices don’t have the proper context. But use those tools and the context will come to you.

That statement resonates powerfully with two themes I’ve touched on before. One comes from John Willinsky, the educator, reading specialist, and open access advocate whom I encountered here, here, and here. There are only two factors that govern reading success, Willinsky says: motivation and context. To illustrate these principles, he talks about medical patients who, though they are nominally grade-8 readers, routinely surprise their doctors with grade-14 mastery of medical literature. How? When you or a loved one is sick, you’re intensely motivated to understand that literature. Yes, you initially lack context, but in the era of the Net, with the help of an ad-hoc community of others in the same situation, you can bootstrap that context.

The second resonance is with this Jeff Jonas meme: Data will find data, and relevance will find the user. I highlighted the social dimension of that idea in this essay: Data finds data, then people find people. Deepak Singh says these principles “define what the modern web is all about,” and adds that “without common data to commune around, there can be no communities.”

GovTrack is one of a number of efforts to create a framework of common data around which new political communities can form. I’ve also mentioned MAPLight, one of a number of projects funded by the Sunlight Foundation, whose chief data architect, Greg Elin, I interviewed here.

Admittedly these new services don’t yet close the gap that separates an ordinary citizen who’s motivated by a particular interest from a professional denizen of the poliwonkablogosphere. But it’s at least possible to leap across the gap, and we’re on the verge of building bridges across it.

Kudos for MAPLight’s visualization of Congressional activity

In yesterday’s essay on my efforts to relate what our New Hampshire senators say about renewable energy to what they do or don’t do about it (and why), I left out a crucial step. It’s a nice illustration of the subtle and powerful data visualization provided by MAPLight.org.

I’d found references to a crucial NO vote cast by John Sununu, but was having trouble finding out for which bill, and for which step in its process, that vote was recorded. I thought the bill was H.R.6-110, and I was looking for a NO vote, cast by Sununu just before its Dec 13 passage, where the final tally was 59-40.

MAPLight’s use of a Timeline proved decisive. In this view of activity around H.R.6-110, scoped to Sununu, the circled region records two Sununu votes around that time. One red for NO, the other green for YES.

Mousing over the red NO vote got me very close to the answer:

This wasn’t quite right. Although the annotation shown in the screenshot reads Dec 07, the mouseover popup (which I couldn’t capture) says Dec 13, which turns out to be the date I was looking for.

It seems that this rendering of the timeline can’t accommodate the closely-spaced votes revealed in this details view:

December 07, 2007 Senate Motion to Invoke Cloture on… Fail 53 42
December 13, 2007 Senate Motion to Concur in the… Pass 86 8
December 13, 2007 Senate Motion to Invoke Cloture on… Fail 59 40
December 18, 2007 House On Agreeing to Senate Adt… Pass 314 100

But when I got to that view, and saw the 59-40 tally that had been reported, I knew I’d identified the crucial cloture vote.

Nicely done MAPLight!

While we’re on the subject, by the way, I wanted to pin down exactly what language was removed from HR 6 by the failure of that crucial vote. From a Congressional Research Service report entitled Energy Independence and Security Act of 2007: A Summary of Major Provisions, by Fred Sissine et al., I found this clue:

The White House threatened to veto the House-passed bill mainly because of the presence of provisions for an RPS [renewable portfolio standard] and for the repeal of oil and gas tax subsidies.

On December 7, 2007, a Senate cloture vote on the House-passed version of H.R. 6 — with provisions for an RPS and for the repeal of oil and gas subsidies — failed (52-43). After stripping out the RPS and modifying the tax package, a cloture vote on S.Amdt. 3841 failed (59-40).

To visualize what was subtracted from S.Amdt 3841 to produce S.Admt 3850, I dropped 3841 into Wikipedia, replaced it with 3580, and compared those two versions.

Here’s where the two part ways:

And here, deep in the thickets of 3841, is my beloved pellet stove provision in the left column, matched only by emptiness in the right column:

Now this surely isn’t an appropriate use of Wikipedia, so I’ve withdrawn the page I made to compare these two versions. But it’s odd, don’t you think? Why can we readily analyze the differences between two versions of a fanciful article about the heavy metal umlaut, but not so readily analyze the differences between two versions of a major piece of legislation?

Net-enhanced democracy: Amazing progress, solvable challenges

Last week I spoke with Joshua Tauberer for an upcoming interview about his project, GovTrack.us, which is one of several efforts to make official sources of information about Congressional legislation available in more useful and actionable ways.

On Sunday, my local newspaper ran an article that sent me scrambling to GovTrack, to OpenCongress, to the Washington Post’s Congressional voting database, to the Library of Congress’ THOMAS, and to a flock of other sites. Why? The newspaper article, illustrated by a photo of NH Senator John Sununu standing with the owners of New England Wood Pellet in front of a pile of sawdust that will be pelletized, said:

Sununu supported a tax credit on high-efficiency biomass fuel systems, including wood-pellet stoves. The Senate passed the measure in April, in effect encouraging people to use the appliances, just as those who take on solar or wind power receive financial incentives for doing so.

I’ve invested rather heavily in clean wood-burning technologies — first in a pellet stove, now also in a wood gasification boiler — and I’m hoping these investments will qualify for tax credits. When phone calls to the local offices of my state representatives produced no results, I resolved to research the matter myself.

Here were my questions, along with the answers I found.

Q: Which bill approved by the Senate in April contains these incentives?

A: Bizarrely but typically, it’s not an energy bill. Instead it’s a housing bill, the Foreclosure Prevention Act of 2008, otherwise known as:

Building American Homeownership Act of 2008
Clean Energy Tax Stimulus Act of 2008
FHA Manufactured Housing Loan Modernization Act of 2008
FHA Modernization Act of 2008
Mortgage Disclosure Improvement Act of 2008
REIT Investment Diversification and Empowerment Act of 2008

Q: Which appliances will qualify?

A: The bill says:

(F) a stove which uses the burning of biomass fuel to heat a dwelling unit located in the United States and used as a residence by the taxpayer, or to heat water for use in such a dwelling unit, and which has a thermal efficiency rating of at least 75 percent.

On that basis I believe that both of my appliances should qualify.

Q: What is the status of that legislation?

A: According to GovTrack:

Status:
Introduced Jul 30, 2007
Scheduled for Debate Sep 5, 2007
Amendments (235 proposed) [details]
Passed House [details] Aug 4, 2007
Passed Senate Apr 10, 2008
Differences Resolved
Signed by President
The bill may now proceed to
a conference committee of senators and representatives to work out
differences in the versions of the bill each chamber approved.
The bill then goes to the President before becoming law.
[Last Updated: Jul 20, 2008]

Cool! That’s what I needed to know. It ain’t law yet, but it might be. And thanks to a bevy of Web 2.0 tools and services, it was easy to find that out, right?

Wrong. It took me three hours on Sunday morning to sort this out. Along the way, I uncovered information that challenges the premise of the newspaper story. Entitled “Sununu touts alternative energy”, the story supports NH Senator John Sununu’s own claim to be a champion of long-term energy policy.

Here’s part of the analysis I posted in a comment to that newspaper article:

Sununu (and Gregg) were two of the votes responsible for the failure, by one vote, to include a $22 billion energy tax package in the Energy Independence and Security Act of 2007. Weirdly, that rejected tax package included Sununu’s original pellet stove provision from S1697!

So while the act became Public Law 110-140 in December, $22 billion of renewable energy incentives were jettisoned from it by Sununu. Why? The Congressional Quarterly noted that it “would have been offset in part by stripping about $13 billion in tax breaks from the oil and gas industry,” suggesting that Sununu was influenced by those interests.

Subsequently, I remembered Maplight.org where I found the breakdown for that crucial NO vote:

It’s utterly amazing, and deeply hopeful, that an ordinary citizen with no prior experience with an issue can dive in and do this kind of analysis.

And yet…I’m not really an ordinary citizen. It’s true that I’ve never before attempted this kind of detailed analysis. But I have an unusual degree of skill and perseverence when it comes to searching the Net, finding and using a variety of services, and correlating information from a variety of sources. At one point, trying to pull it all together, I had two dozen browser tabs open. That’s not normal.

It’ll take a few more turns of the crank before this kind of thing becomes routine enough for a truly ordinary citizen who cares about some issue to do the same kind of analysis.

On the first turn of the crank, we’ll need to get to the point where a Net-savvy individual can do what I did in minutes not hours.

On the second turn of the crank, we’ll need to deliver that capability into the hands of folks who aren’t freakishly Net-savvy.

I really believe that’s possible, maybe even inevitable. But challenges remain.

One of the biggest challenges, as Joshua Tauberer says in our upcoming interview and as my experience here confirms, is the analysis of the legislative texts themselves. Even with all of the excellent online services now available, it’s incredibly hard to relate colloquial discussion of legislation to the actual process and to related commentary. Every day, we read statements like this:

The Senate passed the measure in April, in effect encouraging people to use the appliances, just as those who take on solar or wind power receive financial incentives for doing so.

It took me hours to unravel that statement. For starters, it would help if newspapers would cite the legislation they mention. But that’s only a start. In this case, as is typical, the Sununu pellet stove provision began its life in one Senate energy bill that died, migrated into another Senate energy bill that died, and eventually wound up as part of a housing bill that passed. Separately, the provision was included in, then struck from, yet another energy bill that passed.

Crowdsourcing will be part of the answer. But to empower the crowds, we’re going to need power tools that help us visualize how specific bits of legislative language flow through the bewildering thicket of bills, to tie those bits of language to their sponsors, and to relate them to journalistic/blogospheric commentary.

Happily, this is just the sort of problem that should appeal to software engineers. The management of source code, like the management of legislation, is a high-stakes game. And in the software realm, we play that game really well. We take for granted the ability to reliably track changes in a large and evolving corpus of interrelated texts. When that same expertise is applied in the legislative realm, the results could be dramatic.

On a related note, I enjoyed this snippet from a Wired interview with Martin Wattenberg, whom I also interviewed here.

Wired:
Why is a numbers guy like you so interested in large textual data sets?

Martin Wattenberg:
Language is one of the best data-compression mechanisms we have. The information contained in literature, or even email, encodes our identity as human beings. The entire literary canon may be smaller than what comes out of particle accelerators or models of the human brain, but the meaning coded into words can’t be measured in bytes. It’s deeply compressed. Twelve words from Voltaire can hold a lifetime of experience.

Similarly, snippets of legislation can be the tickets to a more participatory democracy — if we can unravel and expose the contexts surrounding those snippets. It should be doable.

John Faughnan’s amazing Outlook hack (and why it matters)

Although I’ve conversed online with John Faughnan since my days at BYTE, we’ve never met, and we had not even spoken on the phone until last week when he joined me on an episode of my Interviews with Innovators podcast. It was a great pleasure to finally connect in realtime with the prolific author of thoughtful analysis and commentary on things in general, on information technology, and on resources for parents of children with cognitive or emotional-behavior disabilities.

John was a country doctor, and he retains his medical license, but he doesn’t see patients nowadays. Instead he directs the development of clinical productivity software, with particular focus on methods of knowledge representation, and on strategies for effective collaboration.

We share a passion for strategies that entail simple but often overlooked uses of common software applications. For example, did you know that it’s possible, in Outlook, to edit the subject of an email message after it’s been received, and is just sitting in your archive? Try it, and you’ll find that you can. Color me amazed. I’m just the sort of personal information management geek who’d have discovered a hack like that, but I never did.

Now, why would you want to do such a thing? It’s a defensive strategy. The message entitled “Re: Next week” probably ought to be entitled something like “Consensus reached among A, B, and C on issue X for project Y.” Which title would you rather scan, in search results, six months later?

(John would like to find, and personally thank, the developer responsible for this feature, so if you know that person, or are that person, speak up!)

You can think of this technique as a kind of enhanced tagging. It’s related to a strategy for enriching email — embodying the journalistic principle of “heads, decks, and leads” — which I described in my book and in this report.

People mainly still think of information architecture as a discipline practiced only by designers and publishers. But what John and I have always thought is that we’re all becoming designers and publishers of streams of information, that those streams can all be navigated and searched in one way or another, and that the value of those streams depends on the ability of ourselves and others to navigate and search them effectively.

We also think that effectiveness requires two things. First, obviously, software that embodies the right principles and enables the right practices. But second, a broad awareness of right principles and practices. Those, we agree in this conversation, are not necessarily intuited by Gen X, Y, or Z just because they’re so-called digital natives. This stuff needs to be articulated, and it needs to be taught.

How to wire up a timer-triggered WPF event handler in IronPython

In the last installment of my little series on turning Internet feeds into TV feeds, I had decided to use IronPython to fetch data from the Internet, but C# to drive the WPF (Windows Presentation Foundation) application whose display my local public access TV station will broadcast. This division of labor between C# and IronPython arose because the XAML that drives the display needs to be refreshed periodically, and I didn’t know how, in IronPython, to properly delegate a timer-based event handler for WPF.

In this comment, Michael Foord, author of IronPython in Action and a major contributor to the IronPython Cookbook, showed me the way. Thanks Michael!

Based on his example, I’ve rewritten the C# program shown here as the IronPython script shown below.

I haven’t yet decided which version to deploy, but I’m leaning toward the IronPython version. Not because it’s more concise. It isn’t, really. Nor because I feel any need to use the same language for both components of the solution — that is, the feed fetcher and the feed displayer. I don’t care about language uniformity for its own sake.

I am, however, thinking that the folks at the TV station may want to modify these programs themselves. They’re pretty simple, and there’s no reason they shouldn’t be able to tinker with them. From that perspective, code that can be modified with nothing more than a text editor will be more accessible than code which requires a compiler.

I’m reminded of my early days as a website operator, when I was always glad to discover that a third-party application was written in Perl, rather than in C. That meant I could, and sometimes did, tweak the application in ways that otherwise would have been difficult or even (lacking C source code) impossible.

The difference here, of course, is that all of the underlying machinery — XAML, WPF, and the entire .NET Framework — is exactly the same1 when approached from a scripting language like IronPython or a compiled language like C#. This ability to use common infrastructure from different langages — and from very different kinds of languages — has always seemed like a big deal to me, and still does.


1 The same, that is, modulo the kind of boundary-crossing issue that stumped me until Michael Foord pointed me to CallTarget0, the wrapper for creating a delegate in IronPython.


import clr

clr.AddReferenceByPartialName("PresentationCore")
clr.AddReferenceByPartialName("PresentationFramework")
clr.AddReferenceByPartialName("WindowsBase")
clr.AddReferenceByPartialName("IronPython")

from System import *
from System.Windows import *
from System.Windows.Markup import *
from System.Windows.Media import *
from System.Windows.Input import *
from System.Windows.Threading import *

from IronPython.Runtime.Calls import CallTarget0

def LoadXaml(filename):
  from System.IO import *
  from System.Windows.Markup import XamlReader
  f = FileStream(filename, FileMode.Open)
  try:
    element = XamlReader.Load(f)
  finally:
	f.Close()
  return element

class Scroller(Application):

  def tickhandler(self,sender,args):
    def update_xaml():
      self.window.Content = LoadXaml(self.xaml)
    self.timer.Dispatcher.Invoke(DispatcherPriority.Normal, 
      CallTarget0(update_xaml))

  def __init__(self):
    Application.__init__(self)
    self.xaml = "scroller.xaml"
    self.window = Window()
    self.window.Content = LoadXaml(self.xaml)
    self.window.WindowStyle = WindowStyle.None       # go fullscreen
    self.window.WindowState = WindowState.Maximized  # 
    self.window.Topmost = True                       #
    self.window.Cursor = Cursors.None                #
    self.window.Background = Brushes.Black           #
    self.window.Foreground = Brushes.White           #
    self.window.Show()
    self.timer = DispatcherTimer()
    self.timer.Interval = TimeSpan(0, 60, 0)         # refresh hourly
    self.timer.Tick += self.tickhandler              #
    self.timer.Start()                               #

Scroller().Run()

Dan Bricklin on becoming a Happy Caster

The Conversations Network is embarking on a new phase in which it will expand its ambition to capture, publish, and curate spoken-word audio from a wide range of sources. One of the challenges will be to help more people effectively capture audio to a reasonable standard of quality. Dan Bricklin, my guest for this week’s ITConversations show, has ascended that learning curve in recent years. In this conversation he explains why he’s become interested in audio recording, and what he has learned about equipment, and techniques, which can be readily transferred to individuals and organizations wanting to make decent recordings of their own events.

When I embarked on my personal audio adventure a few years ago, I naively thought that our fancy new digital technologies would make the whole process very simple. Boy, was I wrong about that. Yes, we’ve made digital photography accessible to the masses, but there was vast demand for enabling the so-called Happy Snapper to point, shoot, and take a decent photo. There’s been comparatively little demand for enabling the Happy Caster to plunk down a microphone, punch record, and capture a decent sound track.

Over the last few years I’ve slowly and painfully assimilated just a fraction of the audio lore possessed by domain experts like the Conversations Network’s founder Doug Kaye, and its senior audio engineer Paul Figgiani. So it was refreshing to hear from Dan Bricklin that it has also been a struggle for him to become competent in this domain.

I guess the demand for point-and-shoot photography will always outstrip, by orders of magnitude, the demand for plunk-and-punch audio recording. But the latter demand is growing, and in this conversation we speculate a bit on what the Happy Caster solution might be.

Mainly, though, Dan focuses on two things. First, the new opportunity to capture spoken-word events that would otherwise be lost, and publish them for audiences that didn’t attend, or couldn’t have attended, in person.

Second, the minimal setup that will enable folks who are not audio experts to accomplish that capture and publication.


PS: A bit of backstory on this recording illustrates some of the challenges of the audio domain. In my FAQ for interviewees, I invite remote interviewees to record themselves locally, then send me the track which I combine with my own locally-recorded track. Why? If you’re sending voice over the network, whether it’s POTS (plain old telephone service) or Skype, there’s a lot that can and often does go wrong. Eliminate the network and you avoid all those problems.

In principle, combining local tracks recorded separately is a great solution. In practice, it has almost never worked out, and this case was no exception.

Usually the problem is that interviewees lack the gear or knowledge required to make a decent local track. Attempts to record directly into a computer always end badly. Most people don’t own standalone digital audio recorders. In one case, a musician who routinely records his music through a mixer nevertheless produced an unusable track because he’s not used to recording his voice and overshot the limits.

In this case, Dan was quite capable of making a good recording, and he did, but things went wrong on my end. What Dan recorded was an MP3 file. What I was expecting was a WAV file, because I was going to edit the combined recording and it’s dicey to uncompress an MP3, edit, and then recompress.

Now, Dan had recorded the MP3 at a bit rate — 192kbps — that he judged would be high enough to survive an edit. But would our discriminating audio engineer Paul Figgiani agree? We weren’t sure, so I sent Paul samples of Dan’s MP3 track and the WAV file I made from the telephone track I’d recorded using the Telos. Paul’s verdict: “I think we can make the 192 kbps mp3 version work. The bit rate is high enough … lets go with it.”

So far, so good. But when I loaded up my local track and Dan’s remote track into Audition, things didn’t line up. My WAV track was slightly longer (or shorter, I can’t remember) than Dan’s MP3 track. The difference was only about 1.5 seconds over an hour-long recording, but still, it had to be dealt with.

Audition has a time-stretch feature that can be used to solve this problem. And I could swear that I’ve used it successfully before in these circumstances. But this time, I couldn’t make it work. Every time I tried to stretch the shorter clip, it snapped back to its original position. I fiddled with every approach I could think of, or could discover by searching, and finally threw up my hands and just used the original recording that had both halves of the conversation in sync. If this Audition behaviour rings a bell with anyone, I’d love to know what went wrong and how to avoid it next time.

The moral, anyway, is that if a reasonably technical guy like me is struggling to keep his head above water in this domain, it’s clear that non-geeky civilians will just drown. I’m quite curious to know when, or perhaps whether, those civilians will constitute a market that technology providers want to serve.

Homophily, anti-recommendation, and Driveway Moments

The folks at National Public Radio love to create driveway moments:

You’re driving along, listening to a story on NPR. Suddenly, you find yourself at your destination, so riveted to a piece that you sit in your idling car to hear it all the way through. That’s a Driveway Moment.

The podcasting counterpart, for me, is the Ashuelot Moment. I’m jogging along the Ashuelot River, and I’m so riveted to a piece that I take a longer route so my run won’t end before the story does.

The Long Now podcasts are my most reliable source of Moments but they’re only on a monthly cycle. TED talks are another good source, though I’ve lost track of how to subscribe to the comprehensive audio-only feed. The Conversations Network, to which I contribute a weekly show, produces occasional Moments, but a lot of the material there is so closely aligned with my own particular interests and inclinations that it doesn’t often surprise or challenge me.

Another good source is Christopher Lydon’s Open Source, which launched in 2005, suffered a setback in 2006, and then recovered in 2007. It took me a while to reconnect after the hiatus, but now I’m finding it to be more stimulating than ever.

Here’s my most recent Moment, from this Open Source show with Ethan Zuckerman and Solana Larsen. Ethan is speaking:

My hope was that with the Internet, suddenly we’re all connected, we hold hands and sing Kumbaya. And it just hasn’t worked out that way.


You loook at a site like Digg, or Reddit, these are sites that promised the future of journalism. We’d all get together and decide what’s important. But, who’s we? Or as per the Lone Ranger, who’s we, white man? Or more to the point, who’s me, white geek?

If you’re getting your news from these sites, you’re getting a very particular, tech-heavy view of politics, a fairly focused view of the world. And you start falling victim to homophily, which is what happens when all of your news and opinions are coming from people who’ve got the same background and the same values as you.

Homophily is the tendency of birds of a feather to flock together. It’s the tendency to walk into a room, find the person most similar to you, and form a bond. It’s a natural human tendency, but it’s probably worth fighting against. Homophily makes you stupid.

Of course I share tribal affiliations with Ethan Zuckerman, so I’d have been likely to find that particular show one way or another. But Global Voices Online, the project that Ethan and Solana discuss on that show, is all about resisting homophily, and enabling us to tune into global perspectives offered by people in circumstances very different from our own.

Just because we can, though, doesn’t mean we will. Homophily is a natural tendency. It’s easy and comfortable to immerse ourselves in the familiar. It’s hard and uncomfortable to seek out the unfamiliar. How do we overcome that?

Recommendation systems don’t help me much. They only suggest things similar to other things I’ve shown interest in. Increasingly that just frustrates me. The most delightful recommendations are those that connect me with things that interest me in unpredictable ways. That happens serendipitously, and I haven’t yet found a reliable way to manufacture the serendipity.

Lately I’ve started to wonder about the notion of anti-recommendation systems. One example of an anti-recommendation system is LibraryThing’s UnSuggester, which find books least likely to coincide with yours. It’s a whimsical feature that honestly hasn’t been useful to me yet, but I think the idea merits exploration and development.

Although it isn’t automated or automatable, I’d argue that the Passion Thursday series on Open Source is a kind of anti-recommender. The series includes shows about birdwatching, the pursuit of truth, poker, the potato, cursive handwriting, and the theremin, an early electronic instrument recently notable in the repertoire of the indie band DeVotchKa. The only common thread is someone’s passionate interest in something.

We’re not inclined to resist homophily and seek out otherness. But passionate storytellers can take us to places we wouldn’t otherwise go, and create Moments there.

Passion is a good way to lubricate the engine of serendipity.

Will people understand and embrace the right identity systems? Maybe yes!

In conversation with English and Welsh friends last week, the subject of Britain’s imminent National Identity Scheme came up. My friends, who are worldly and well-educated but not technical, voiced concerns about the amount of personal information that will be stored. Their understanding was that a lot of this information will be kept on the new ID card. In fact, the proposal says that only a subset will stored on the card, which will be backed by a cloud-based (and decentralized) National Identity Register. But either way, my friends’ concerns are of course valid. If governments or businesses aggregate too much personal information, accidents and abuses will occur.

At the same time, my friends do recognize the need for a strong and secure means of identification. So they’re not opposed to identity cards on principle, they just don’t want those cards to contain, or link to, extensive dossiers.

At this point, channeling Kim Cameron, I launched into an explanation of the laws of identity and the identity metasystem. Well, sort of. I didn’t say anything about cryptography, or digital certificates, or XML web services. But I did paint a picture of a world in which individuals interact with many identity providers and many relying parties, in which all actors trust one another in exactly the ways they already do today, and in which disclosure of personal information is minimal and context-dependent.

Halfway through I thought, well, this will never fly. This whole scheme is based on decentralization and indirection, and I know people don’t take naturally to those concepts.

But…they completely got it! Maybe that’s because the threat of a monolithic system leads people to appreciate the virtues of a decentralized one. Maybe it’s because ongoing experience with the Net makes people more comfortable with the principle of indirection. Maybe it’s both these factors and others as well. In any event, it was a hopeful moment. Identity geeks have struggled, for many years, not only to devise right systems, but also to motivate an understanding of what makes systems right, and why. Now that right systems are coming into existence, it’s good to see that (some) people are ready to appreciate and embrace them.

How the WorldWide Telescope works

On my Perspectives show last week, Curtis Wong and Roy Gould relate the history and educational mission of the WorldWide Telescope. On this week’s show, principal developer Jonathan Fay describes how the underlying technologies enable the WWT’s seamless view of the sky.

There were a bunch of things I wanted to know, including:

How does the WWT project build on, and extend, the SkyServer project to which Jim Gray made fundamental contributions?

What standards and protocols enable the various sky surveys to be woven together?

What’s the relationship between Deep Zoom and the WWT’s own scheme for managing and viewing tiled multi-resolution imagery?

How much of the data is stored on Microsoft servers, how much is stored elsewhere, and in what ways do the supporting data services cooperate?

Jonathan answers all these questions, and he also answers one I didn’t think to ask:

What technique is used to project the stars onto an imaginary sphere at near-infinite distance?

The answer to that last question is that a new kind of spherical projection had to be invented:

Imagine taking a round room, and trying to put a bunch of bathroom tiles on it, and grout it. The tiles seem to come together and have parallel lines for a while, but eventually it stops working well. Maybe you can take one line around the equator, but as you go up you have fewer tiles, and weird-shaped tiles, and nothing lines up.

That’s the problem we have. We’re looking at spherical data, so we had to come up with a new spherical transform that preserves the poles. In previous projects, like Virtual Earth or TerraServer or Google Earth, the poles weren’t important, because nobody lives there and nobody needs map directions for driving around there.

So we had to come up with something called TOAST: tesselated octahedral adaptive subdivision transform. It creates a 360-degree wraparound view that’s either a planet surface or the infinite sphere of the sky, and lets you represent it using a 3D graphics accelerator, very rapidly and efficiently. So we can have an image pyramid the way Deep Zoom does, and TerraServer before it, but we don’t have to give up the poles.

This transform isn’t proprietary, and in fact it’s being applied to the 50-odd full-sky surveys hosted at NASA’s SkyView virtual observatory. The implications are pretty astounding. This imagery is stored in astronomical databases using what’s called tangential projection, which suffers from polar distortion when combined into large mosaics. Now the imagery can be combined into large mosaics — or indeed a complete view of the sky — and seen without distortion. What’s more, multiple surveys can be aligned to that spherical projection. That’s why, in WorldWide Telescope, you can cross-fade between a view of the Milky Way in visible light and views in infrared or ultraviolet light.

What the WorldWide Telescope really is, Jonathan says, is a browser, like a web browser but for an information space defined in astronomical terms. Here’s how he sums up the work that was necessary to make that possible:

The vision of getting everybody access to all this astronomy data required systematic changes at every single level. We built on some things that Jim pioneered with NVO, and worked from there, but it was very systematic. How people process the data. The client to access the data. The protocols over the wire. Educating people, providing the context for it.

We put a lot of things together, but we also created a systematic model for how to do everything end to end, top to bottom, left to right. Now there may be other people who use the pieces that we’ve created, and then change them to use different data sources, different visualizations. Say someone creates a Mac client, or an iPhone client, that’s possible. Or a mobile phone version of it, or a web-based version. Over time we or others can replace various components, but as a reference model for solving all the problems in order to get the data into people’s homes and into their eyeballs — you had to solve for all of those problems, otherwise people are still blocked from being able to really explore.

For Curtis Wong, the WWT is an extension of John Dobson’s sidewalk astromony — a way to bring telescopes to the public and to enable astronomers to share their knowledge of the sky with everybody. For Jonathan Fay, it’s the perfect application of earth and sky visualization technologies he’s been developing throughout his career. Their interests and talents combined, as Jonathan says, like peanut butter and chocolate:

Curtis had been exploring how to create an educational environment with rich tools for exploring space, and he’d been collaborating with Jim Gray on TerraServer, and now he was looking for the technology to make it possible.

Here I had this technology, and was looking for somebody who was enthusiastic about having a purpose for it. So it was the peanut butter and chocolate moment.

Yum.

More ways to turn Internet feeds into TV feeds

Last week I started looking into ways to Internet feeds into TV feeds. Although I did come up with a way to turn a data feed into a video file, that wound up being overkill. It turns out that the local station is willing to broadcast the signal from a computer display. To create that signal, several folks suggested using PowerPoint, but I found that its scrolling credits feature doesn’t accommodate really long lists of credits. So I decided to try XAML, the application markup language that works with Silverlight and the Windows Presentation Foundation (WPF), in concert with IronPython.

The plan was as follows. A long-running IronPython script periodically fetches the feed from a web service, interpolates the text into a XAML template that animates the crawl, and displays the XAML in a fullscreen white-on-black WPF window.

Here’s the XAML template:

<Grid xmlns="http://schemas.microsoft.com/winfx/2006/xaml/presentation"
  xmlns:x="http://schemas.microsoft.com/winfx/2006/xaml"  
  Width="%s" Height="%s">
<TextBlock xml:space="preserve" FontSize="%s" Margin="%s,%s,0,0" 
  FontFamily="Arial">
    <TextBlock.RenderTransform>
      <TranslateTransform x:Name="translate" />
    </TextBlock.RenderTransform>
  <TextBlock.Triggers>
    <EventTrigger RoutedEvent="FrameworkElement.Loaded">
      <BeginStoryboard>
        <Storyboard RepeatBehavior="Forever">
          <DoubleAnimation From="%s" To="-%s" 
            Storyboard.TargetName="translate" 
            Storyboard.TargetProperty="Y"
			Duration="00:%s:%s" />
        </Storyboard>
      </BeginStoryboard>
    </EventTrigger>
   </TextBlock.Triggers>
<Run>
<![CDATA[
%s
]]>
</Run>
</TextBlock>
</Grid>

Some of the values depend on the number of items in the feed, so the script interpolates those values into the template. Then it formats the feed and plugs the formatted text into the template’s CDATA section. The formatted text looks like this:

EVENTS FOR MON JUN 30 2008 FROM THE ELMCITY.INFO CALENDAR

06:00 AM:  lap swim (ymca)

07:00 AM:  AA: On Awakening Group (eventful: Keene Unitarian
      Universalist Church)

After generating the XAML, the IronPython script fires up an Application object, creates a window, loads in the XAML to start the crawl, and sets a timer to refresh the XAML.

I ran into a snag when I tried to set that timer, though. There are a few different timers you might imagine using in this context, including Python’s own timer object and various timers available in the .NET Framework. All but one of these, however, will complain about invalid cross-thread access when you try to update the application’s user interface from a timer event handler.

The right timer to use, it turns out, is .NET’s System.Windows.Threading.DispatcherTimer. But when I tried it, I ran into another snag. In C#, you create a WPF-friendly timer like so:

DispatcherTimer timer = new DispatcherTimer();
timer.Tick += new EventHandler(event_handler);

event_handler is a method, but EventHandler returns a delegate that encapsulates that method. I couldn’t find a straightforward way to create a delegate, and do that encapsulation, in IronPython.

If you know how, I’d love to hear about it. Then again, it really doesn’t matter. Logically this program has two loosely-coupled parts. The engine part reads the feed from a web service and formats it as XAML. It can be a Python script that runs on a scheduled basis to fetch and format the feed.

The user interface part loads, displays, and then periodically refreshes the XAML. It can be a little C# program that runs forever, displays the animation, and refreshes the data, like so:

using System;
using System.Windows;
using System.Windows.Markup;
using System.Windows.Input;
using System.Windows.Media;
using System.IO;
using System.Windows.Threading;

namespace CalendarCrawl
{
  public class CalendarCrawler
  {
    static private Application app = new Application();

    static private StreamReader getXaml()
    {
    StreamReader sr = new StreamReader("WPF.xaml");
    return sr;
    }

    [STAThread]
    static public void Main(string[] args)
    {
        Window win = new Window();
        win.WindowStyle = WindowStyle.None;           // go fullscreen
        win.WindowState = WindowState.Maximized;      // go fullscreen
        win.Topmost = true;                           // go fullscreen
        win.Cursor = Cursors.None;                    // go fullscreen
        win.Content = XamlReader.Load(getXaml().BaseStream); 
        win.Background = Brushes.Black;
        win.Foreground = Brushes.White;
        DispatcherTimer timer = new DispatcherTimer();
        timer.Interval = new TimeSpan(0, 1, 0);       // every hour
        timer.Tick += new EventHandler(eventHandler); // wire up handler
        timer.Start();
        app.Run(win);
    }

    static private void eventHandler(Object sender, EventArgs args) 
    {
        app.Windows[0].Content = XamlReader.Load((getXaml().BaseStream));
    }
  
  }
}

It was odd how reluctantly I came to this division of labor. Evidently I still need to remind myself that in a world of loosely-coupled applications and services, when you need to get something done, There Is More Than One Way To Do It.

Here’s another way. If the engine doesn’t have to talk to the .NET Framework’s WPF machinery, there’s no need to use IronPython. Any flavor of Python makes a handy tool for talking to RESTful web services, wrangling text, and interacting with the file system.

Here’s yet another way: A Silverlight version of the user interface. It’s nice to know that option is available. However, I’m leaning toward the C# version. The target machine is Vista, it already has .NET and WPF, why use a long-running browser instance just to host this tiny little thing?

One final point is worth mentioning. XAML is really just another source language for the .NET runtime and framework, like C# and IronPython and others. You can, for example, create an application window by writing a Window tag in XAML markup, and specifying parameters as attributes. Or you can do it by invoking System.Windows.Window from IronPython or C# or another .NET language, and specifying parameters in code. The boundary between markup and code is very fluid, and you can draw the line for reasons of convenience, maintainability, and taste. It’s a very flexible system, and it becomes even more flexible when you can use a dynamic language like Python to generate the XAML, the code, or both.

From seeing to hearing: A conversation with Susan Gerhart about assistive technologies for the sight-impaired

For many of us, the podcasting revolution has opened up the audio channel as a new option for receiving information that we might otherwise read. But for the sight-impaired, like Susan Gerhart, who joins me for this week’s ITConversations show, the audio channel isn’t optional. Her myopic retinal degeneration has forced her to shift almost entirely into audio mode in order to read, and to work on the computer.

As a lifelong technologist, Susan is a capable user and evaluator of software and computational devices. When she entered the world of assistive technologies — including the NVDA screen reader, the LevelStar Icon, the Kurzweil NFB Reader — she decided to share her experiences on a blog. In our interview she summarizes what she’s learned so far about using these technologies to adapt to her changing vision.

From PowerPoint to IronPython/XAML

As per the comments on yesterday’s item about creating a video crawl for local TV, it turns out there’s no need to produce a video file. Instead it’ll be OK to use a computer display directly. The computer could be running, for example, a PowerPoint slideshow in a loop.

Here’s the apparently standard recommendation for making scrolling credits in PowerPoint. It was written for earlier versions, but seems applicable also to the current 2007 version:

Create movie-style crawling credits in PowerPoint presentations

  1. In a PowerPoint presentation, create a new slide for credits or any other list that you want to scroll from bottom to top.

  2. Type your credits or other text. Don’t worry about text running off the bottom of the slide. In fact, it should run off the bottom if you are going to have enough text to make a crawl effect work well.

  3. Right-click the text, and on the shortcut menu, click Custom Animation.

  4. Select the text that you want to scroll. In the Custom Animation task pane, click Add Effect. Point to Entrance, and click Credits.

  5. Click Play to see how the effect will look on-screen.

  6. Move the text block completely off the top of the slide. When you play your presentation, the text will crawl or scroll from the bottom of the screen and disappear off the top.

But it doesn’t look like you can get more than three screenfuls of data into the crawl. For example, I made a textbox with 200 lines of text numbered accordingly. Then I animated it using several varations on this technique.

First I put the top of the textbox at the top of the slide, like so:

The effect: Line 0 crawls into view from the bottom of an empty slide, and the crawl ends with line 25 at the top and line 50 at the bottom.

Next I put the top of the textbox at line 25, like so:

The effect: Line 0 appears at the top of the slide, the crawl ends with line 50 at the top and line 75 and the bottom.

Is there a way to include more than three screenfuls of data in the crawl? If not, it looks like it’d be necessary to create a series of slides, each with two screenfuls of data. The first slide would need to have its first line of data at its top. But the second and following slides would need to have their middle lines of data at their tops. Gnarly.

I’m sure that could be done, but why bother? Absent a requirement to produce a video file, there a zillion ways to make text crawl up a computer screen. This might be a good opportunity to explore the combination of IronPython and XAML.

Turning Internet feeds into TV feeds

I’ve cobbled together a way to turn an Internet data feed into a video crawl that can run on my local public access cable TV channel. Before explaining how, I need to explain why. Here’s the short answer: As much as I want everyone to use the Internet for all it’s worth, most people don’t yet.

A couple of years ago, I was campaigning in my community to open up the parent portal into PowerSchool, a student information system that was being used internally by teachers and administrators but wasn’t available to parents via the Internet. At one point I made a screencast that addressed the perceived risks, and showed the compelling benefits, of opening up the portal. The screencast was published on the Internet, available to the whole world, and the whole world includes Keene NH, so that ought to be a good way to bring my message to the community, right?

Wrong. Nobody watched it.

A while later, it hit me. There still aren’t many folks here who are inclined to receive a message like that from InfoWorld.com, or from YouTube, or from any other Internet destination I might use. But there are significant numbers who tune into the local public access station. Why not show the screencast there?

So I dubbed it onto a MiniDV tape, took it down to the station, and gave it to the executive producer.

Him: What’s this?

Me: A demo and discussion of the PowerSchool software. Will you run it?

Him: Sure.

And lo, a couple of weeks later, I heard from the assistant superintendant of schools. He thanked me for applying the external pressure that they’d been needing in order to break through an internal logjam, and he invited me into the beta program. Now, two years later, it’s fully deployed and making a big difference.

Meanwhile, I’ve been working on a community information project that’s all about feeds and syndication. But slow learner that I am, I continue to invite people to use Internet feeds and Internet syndication. And people continue to mostly decline the invitation.

For example, I’ve been working on calendar syndication. The syndication flows two ways. First, inward. The service pulls events from various local websites, and I’m working with the proprietors of those sites to clarify why and and to publish true calendar feeds.

Second, it syndicates outward. With a JavaScript call, you can include the day’s events in another website, like CitizenKeene or Cheshire TV.

But this is all still just Internet stuff. And as we’ve seen, the community doesn’t (yet) tune into the Internet for local information. It does tune into public access cable TV.

So why can’t Internet data feeds show up there?

Well, of course, they can. Here’s a prototype video crawl (the link goes to an animated gif, just for convenience) made from yesterday’s combined calendar. We’ll need to work out the details of format and workflow, but I think it’ll work. And it seems like a great way to connect two worlds.

Calendars are just part of the story. Consider, for example, the public library’s RSS feeds announcing new books and DVDs. I’m one of probably a handful of subscribers to those feeds. Now imagine that the feeds showed up as a video crawl on TV. I bet a lot more folks would find out about new books and DVDs. And maybe, just maybe, the reception of that feed via TV would lead to discovery and use of the more convenient and powerful Internet feed.

We’ll see. Meanwhile, below the fold, I describe the method I’ve come up with to do this. The paint isn’t dry, and I’ll be very interested in comments and suggestions.

… the fold …

Our public access TV station, as may be typical (though I dunno), is a mostly Windows-based operation. As is surely typical, there’s little money to spend, either on people to produce these feeds interactively or on software to produce them automatically. So the requirements seem to be:

  1. Windows-based
  2. Cheap or free
  3. Fully automatable

My first idea was to leverage SMIL. I knew it would be easy and free to transform a feed into markup that can be played by Real, or QuickTime, or Windows Media. And I hoped it would also be easy and free to render that markup into a video format. But there I ran aground. If there’s a free, or at least cheap, SMIL renderer that can be scheduled to run automatically, I’d like to know about it, because that’d probably be the ideal solution. But I haven’t found one.

The next idea was to produce the animation frame by frame. And that’s what I’m actually doing for now. It sounded a lot harder than it turned out to be. After installing the Python Imaging Library, it was possible to write this very concise frame generator:

import Image, ImageFont, ImageDraw

s = """
EVENTS FOR WEDS JUNE 30 FROM ELMCITY.INFO (HTTP://ELMCITY.INFO/EVENTS)

06:00 AM lap swim  (ymca) 
07:00 AM Cheshire Walkers: Indoor Walking Program (eventful: Keene Recreation Center) 
...
Trainers Academy - Level II (eventful: Monadnock Humane Society) 
TOR 7pm (swamp bats) 
"""

lines = s.split('\n')

def frame(index,top):
  image = Image.new('RGB',(720,480),(0,0,0))
  draw = ImageDraw.Draw(image)
  font = ImageFont.truetype("arial.ttf", 18)
  for line in lines:
    draw.text((10, top), line, (255,255,255), font=font)
    top += 25
  image.save('cal%0.3d.gif' % index)

top = 450
for index in range(len(lines)*8):
  print index,top
  frame(index,top)
  top -= 4

This yields a sequence like cal000.gif…calnnn.gif.

I wasn’t sure how to make a video directly from that sequence, but I knew that ImageMagick could turn it into an animated GIF, like so:

convert -adjoin cal???.gif animation.gif

So I did that, and went looking for ways to convert that into a video format. ffmpeg will do it, but the results weren’t pretty, and ffmpeg can be a dicey thing to ask people to install. QuickTime, I found, did a better job. You’d need QuickTime Pro for Windows, which isn’t free, but $30 won’t break the bank.

Now the question became: How to automate the QuickTime conversion? I installed the QuickTime SDK, went looking for examples, and found just what the doctor ordered. Thanks, Luc-Eric!

Luc-Eric’s JavaScript solution, which runs on the Windows command line courtesy of the Windows Script Host, turned out to provide a double benefit. In addition to showing how to automate the conversion of a batch of GIFs to an AVI, it showed me that there was, in fact, no need to produce an intermediate animated GIF. You can just point QuickTime at the sequence in the same way that you can point ImageMagick at the sequence. I hadn’t known that! So ImageMagick dropped out of the toolchain, and there was one less component to require the station to install.

So that’s where things stand. I’m pretty sure there’s a better way to meet the requirements, and I’ll be delighted to discover it. But maybe there isn’t, in which case it looks like this will work.

Either way, it’s the end result that will — or maybe won’t — matter. We’ll do the experiment, and we’ll find out.

A conversation with Jean-Claude Bradley about open notebook science and the educational uses of Second Life

On this week’s ITConversations show I finally got to meet Jean-Claude Bradley, the Drexel chemistry professor who coined the phrase open notebook science and who champions the principles behind it.

There were a couple of surprises for me. First, I was intrigued to learn about Jean-Claude’s vision for mechanized research. I’ve always thought of open notebook science as a way to speed up the iterative cycle of research and publication, and to engage more human minds in collaboration. Of course Jean-Claude thinks so too. But he also thinks that when data are published in accessible formats, and exposed to computational processes running in the cloud, we’ll be able to automate certain aspects of research.

It reminds me of George Hripcsak’s effort to mechanize the interpretation of electronic health records. In general, we’re collecting way more data than the collectors can analyze. Crowdsourcing is one solution to this problem. Mechanization is another. We’ll need both.

The other surprise was hearing about Drexel’s fairly aggressive use of Second Life. I’ve been an amused skeptic on that front, but Jean-Claude’s passionate advocacy requires me to rethink that stance.

What didn’t surprise me, but might well surprise tuition-paying parents of Drexel students, was Jean-Claude’s attitude toward the classroom. He mostly doesn’t see a need for it. The content delivery aspect of education, he feels, is best handled in other ways, including screencasts and podcasts as well as traditional texts. There can, and should, be a range of sources, to accommodate the differing inclinations of learners. And teachers need to be competent producers and orchestrators of those sources. But for Jean-Claude, the best way to engage directly with students is to meet with individuals, not with whole classes.

Now admittedly, a chemistry class doesn’t invite and thrive on group discussion in the same way that, for example, a literature class does. And yet Jean-Claude says that a literature class was one of the models for his use of Second Life. When group interaction is central to the educational experience, he thinks that virtual environments — though he doesn’t require their use — may outperform real ones.

I remain skeptical on that point, but I’m always open-minded, so I hope Jean-Claude will take me up on my offer to visit one of his virtual environments and document the interactions that happen there.

The story of the WorldWide Telescope

My guests for this week’s Perspectives are Microsoft researcher Curtis Wong and Harvard-Smithsonian science educator Roy Gould. At Ted 2008 they jointly delivered the first preview of the WorldWide Telescope, an elegant and powerful application for exploring the sky and weaving narratives about it. In this extended interview, you can hear (or read) the whole story behind the WWT.

I’d known that the WWT was based on Jim Gray’s work, and also that it was dedicated to him. I’d also heard several of the talks he’d given about SkyServer, SkyQuery, and the SQL and XML web services technologies powering those projects.

What I hadn’t fully grasped, until I began preparing for the interview with Curtis and Roy, was Jim Gray’s larger vision for that work. In 2002, with Alex Szalay of Johns Hopkins, he published a paper entitled The World-Wide Telescope: An Archetype for Online Science. Here’s the abstract:

Most scientific data will never be directly examined by scientists; rather it will be put into online databases where it will be analyzed and summarized by computer programs. Scientists increasingly see their instruments through online scientific archives and analysis tools, rather than examining the raw data. Today this analysis is primarily driven by scientists asking queries, but scientific archives are becoming active databases that self-organize and recognize interesting and anomalous facts as data arrives.

Although the WWT isn’t an instrument for professional scientists, Roy Gould thinks it will be used by citizen scientists to collaboratively search the fast-growing corpus of sky imagery. That is, of course, a poignant echo of the collaborative search for Jim Gray when his sailboat went missing.

But for Curtis Wong and Roy Gould, who grew up in Los Angeles and New York, respectively, where neither had access to the dark night sky, the WWT is first and foremost a way to reacquaint our society with the night sky, and to teach us about the universe.

Roy Gould says that when his team surveyed high school students around the country, they found that a majority believe that stars reside within the orbit of Pluto. They also believe that galaxies are closer than stars, because “stars are just point sources, no matter what the magnification, so they must be very far away, whereas galaxies, whatever they are, look big, so they must be closer.”

To fulfill its educational mission the WWT delivers seamless navigation of the sky, contextualized in a variety of ways. Objects are described onscreen, and linked to sources on the web. When you find your way to a stellar neighborhood, thumbnails of the objects in that neighborhood invite you to explore images from a variety of catalogs: the Sloan Digital Sky Survey, Hubble, Chandra.

What’s more, the imagery is correlated so you can see the same object in any of the wavelengths of light used to observe it. If you look at the Milky Way in the standard view, and then switch to infrared, a band of incandescent whiteness emerges from the cloud of stars.

You can use the WWT to explore the sky randomly, but most people will enjoy taking one of the guided tours. Curtis Wong’s lifetime of experience as a creator of interactive multimedia is distilled into this feature of the WWT. Tours are slideshows that move from one object in the sky to the next, and may be annotated with text, spoken-word audio, and music. But at any point you can pause the tour — or hop off the bus, as Curtis says — and explore the neighborhood on your own.

The WWT isn’t just a player of tours, it’s also an authoring tool for creating them. You create slides, navigate to objects in the sky, annotate them, and save the results in an XML format that you can reuse and share.

Like images from catalogs, tours are contextually available. So if you happen upon the Ring Nebula while exploring randomly, and if there’s a tour that mentions the Ring Nebula, then that tour will surface.

Curtis envisions a hypermedia web of sky narratives. For him, this storytelling aspect really is the heart of the project. In the interview he reveals for the first time that an early prototype for the WWT, shelved years ago, was to have been called John Dobson’s Universe.

Dobson, a leading amateur astronomer and innovative telescope builder, founded San Francisco Sidewalk Astronomers, a group that encourages telescope owners to take their telescopes out in public and share their knowledge of the sky. The WorldWide Telescope is poised to carry on that great tradition, and take it in some amazing new directions.

A conversation with George Hripcsak about electronic health records and clinical truth

George Hripcsak, professor of biomedical informatics, is one of the recipients of a Microsoft Research grant to support work on the computational challenges of genome-wide association studies. These studies involve scanning complete human genomes, and looking for correlations between certain markers of genetic variation and certain diseases.

Doing that correlation is a computational challenge, but as I learned in my interview with George Hripcsak for Perspectives, that isn’t the challenge his research addresses. Instead he’s tackling a different challenge: mining electronic health records to figure out what they say about the diseases patients may have.

Why? Suppose you’ve sequenced the DNA of thousands of people for a study. If you’re trying to correlate genetic markers with disease, you need to know what diseases those people have. George calls this “collecting the phenotype” — that is, the expression of the genes responsible for diabetes, or a tendency to complications in labor, or whatever.

Traditionally that’s done by interviewing patients, a painstaking process that doesn’t scale. Given electronic health records, how much of this phenotype collection can be done automatically, and to what level of accuracy? That’s a different kind of computational challenge.

There are basically two ways to go. You can try to templatize the process of clinical data collection, so that health records can be harvested more effectively by researchers. Or you can try to understand the language that clinicians actually use when they describe patients.

For a decade now, George Hripcsak and his colleagues have been pursuing the latter approach, using a system for understanding natural language called MedLEE, which was developed at Columbia.

Ultimately I believe, as George Hripcsak does, that we’ll need a hybrid system that makes use of both structured templates and natural language understanding. But given that health records must primarily serve patient care, and can only secondarily serve research, I like how he harmonizes those objectives:

To the degree we make documentation efficient in serving health care, I think it’ll also be more accurate for the sake of research. If you’re filling out a record for the sake of billing, you’ll have an incentive to use diagnosis codes that optimize billing. Does that then reflect clinical accuracy? And would that then be useful for research?

The important thing is to be grounded in the clinical truth. Put health care first, and then use new computational methods to extract accurate information.

Amen.

Future of the Conversations Network

As recently announced by Doug Kaye, the Conversations Network is embarking on a new phase. The existing channels, including ITConversations and Social Innovation Conversations, will continue. But rather than creating more such channels, the Conversations Network wants to help individuals and organizations capture and publish their own spoken-word audio, mainly in the form of events that are experienced only by attendees but that could be experienced by anyone, anywhere.

This new mission dovetails with PodCorps, a matchmaking service that connects event producers with volunteer stringers who can record those events. When it launched I wrote:

There’s a huge opportunity here to transform communication patterns in a fundamental way. Checking my local events calendar, for example, I see that the following event is scheduled for tonight at the local college:

Mon., Apr. 16
7 to 8:30pm
Pond Side 2 located on Bruder St – Keene State College

Building Smart – Highlighting Local Best Practices

Come and join us in discussing the challenges and successes of implementing innovative building materials, technologies, and design solutions into the built environment.

The information exchanged at that meeting, and at countless meetings like it, has historically been available only to those who attend. There are a million reasons why local folks who might want to attend nevertheless cannot: no babysitter, schedule conflict, etc. And of course remote folks have no opportunity to attend, even though the information exchanged might be highly relevant to them.

Assuming that more of these kinds of events become available, how will we find them? Doug writes:

We will do this using a social-networking model, which allows anyone to post links to recordings he or she finds, to build collections or playlists of their favorite recordings, to share those playlists with others, and to rate and comment on playlists or individual recordings posted by others.

In other words, Webjay for spoken-word audio. It’ll be interesting to see how this unfolds.

In my interview with Webjay’s creator, Lucas Gonze, we talked about some of the reasons why the curatorial model that Webjay promoted hasn’t yet succeeded. One of them, amplified in this comment by Greg Borenstein, is the fear, uncertainty, and doubt that pervades any distribution of — or even just linking to — MP3 files.

That kind of FUD shouldn’t be an issue for spoken-word audio that is explicitly free and legal. So I hope that we can evolve a culture of uninhibited collaborative curation. We’ll see.

I’ll also be curious to see what kinds of new channels and shows may arise from this effort. That isn’t the primary focus. Rather, the idea is to capture, share, and find recordings of events that have already been planned, organized, and held. The Conversations Network mainly seeks to enable the curation of those events. So someone might, for example, assemble the best recorded material in the alternative energy genre, from a variety of sources. I’d like to subscribe to that curator.

But there’s another kind of curation. It’s what I do when I select, from among the many people and ideas that I encounter, those I’ll feature on my two series of interviews: Interviews with Innovators and Perspectives. The world is full of interesting people and ideas, and we may also see the emergence of curators who select and highlight them in original ways. I’d like to subscribe to those curators too.

exchange2ical available on CodePlex

The Exchange-to-iCalendar script that I mentioned here is now published to CodePlex. It’s intended for organizations that run Exchange and would like to publish selected calendars in iCalendar (aka iCal, or .ICS) format without having to rely on a client machine running Outlook 2007.

I’ve never run a real Exchange server, so I’m wide open to suggestions as to how to actually publish the ICS file created by this little IronPython script. Right now, it just emits that file. For user Jon on Exchange host Zanzibar, you would do something like this:

ipy Zanzibar Jon > jon.ics

There are lots of ways jon.ics could get pushed to a web-accessible location, but I’m not sure what the default should be, or whether to do a filesystem operation, or an FTP transfer, or something else.

My idea is that you’d schedule this command to run on a regular basis, and that it would run under an account that has the necessary privileges to access the specified user’s calendar. But again, I’m not an Exchange admin, so if that sounds like the wrong thing, let me know what the right thing would be.

As for the iCalendar output, this script currently does the Simplest Thing That Could Possibly Work. It doesn’t, for example, try to “fold” long lines in the output (e.g., event summaries and unique IDs), which I gather the spec recommends but does not require.

There’s only been minimal testing. I’ve run it against a couple of different Exchange servers (2003 and 2007), validated the ICS output using this handy validator, and verified that the resulting files — containing both individual and recurring events — can be successfully imported, or subscribed to, in Outlook 2007, Google Calendar, and Apple iCal.

If you have a need for such a thing, try it and let me know how it goes.

A quiet retreat from the busy information commons

Nick Carr’s essay in the current Atlantic Monthly crystallizes a lot of what I’ve been feeling for a couple of years about how our use of the Net is changing us. Not co-incidentally I read the essay in the printed magazine whose non-hypertextuality I experienced as a feature, not a bug. (Nick writes: “Unlike footnotes, to which they’re sometimes likened, hyperlinks don’t merely point to related works; they propel you toward them.”)

Although Nick doesn’t invoke Linda Stone’s touchstone phrase “continuous partial attention,” that’s the effect he’s noticing and reacting to. Whatever you think about the consequences, it’s clear that the Net doesn’t invite sustained attention, deep reading, and quiet contemplation.

Of course neither did cable television which, in the pre-Internet era, taught us to click-slice our attention much more radically than was formerly possible. But TV was just entertainment. You didn’t have to turn it on to do your job. Most of us do have to turn on the computer. As Paul Graham points out in Disconnecting Distraction, that’s a problem. He solves it by disconnecting his main work computer from the Net. But as work increasingly entails the use of software, and as software migrates into the cloud, I wonder how feasible that will be for most so-called knowledge workers.

Although Nick acknowledges that he’ll be branded “a Luddite and a nostalgist” I think his essay is spot on. The Net surely is rewiring our brains. What should we think about that, and what should we do about it? These are important questions.

Since we can’t turn back the clock, and we wouldn’t want to even if we could, I guess we’ll have to learn to live happily and productively in the noosphere we are creating.

Part of the answer is to develop — and teach — strategies that enable us to graze on the information commons in the most effective ways. I work hard at that.

But we also need strategies that enable us to retreat from that commons, and to experience sustained attention, deep reading, and quiet contemplation.

On that front, technology sometimes gives back with one hand what it takes away with the other. I’ll admit that it’s sometimes been a struggle for me, in recent years, to find that retreat in books, particularly fiction, and particularly in printed form. But portable long-form audio has been transformative. This weekend, on several multi-hour bike rides, I listened to the LibriVox version of Jules Verne’s Around the World in Eighty Days, beautifully read by Mark Smith (thanks Mark!).

It’s a great tale! I’d never have found the hours of couch time to re-read it, but I loved listening to it on my rides. (It was also nice to recall that I had written the software that simplifies downloading the book to a podcatcher.)

As I’ve said before, the quality of sustained attention that I find I can bring to long-form audio has been, for me, an unexpected benefit of great value. I’m sure our technologies are rewiring are brains in all sorts of ways, for better and worse. Often, those changes propel us into new and uncharted territory. But they can also help us reactivate ancient traditions, like oral storytelling, and rediscover their powerful neural effects.

A conversation with Harry Lewis and Ken Ledeen about technology, society, and Hedy Lamarr

On this week’s Interviews with Innovators show I spoke with Harry Lewis and Ken Ledeen, two of the three authors of the forthcoming book Blown To Bits: Your Life, Liberty, and Happiness After the Digital Explosion. The book explores why information technologies continue to produce surprising outcomes, and how society responds to them.

One of the threads running through the conversation is a perennial topic among thoughtful technologists: How much can most people understand about the underlying principles of information technology, and how much should they need to understand?

The authors believe that those principles are more explainable than we think, and as evidence they cite the remarkable tale of Hedy Lamarr, the famous Hollywood actress who — in collaboration with the musician George Antheil — invented and patented the idea of frequency hopping. Their goal was to help American torpedo guidance systems resist jamming by the Nazis. Now, of course, it has become one of the foundations of wireless communication.

Ken Ledeen tells the story at length, and with great animation, in our interview. Coincidentally, today’s NYTimes has a review of “an imaginative, two-character multimedia 80-minute play, ‘Frequency Hopping'”, that brings that same story to life.

We need to be careful about the lessons we draw from this tale. Although Lamarr the actress and Antheil the musician are usually referred to as the “improbable inventors” of frequency hopping, it didn’t happen by accident. They may have been scientific and engineering amateurs, but they were inspired and talented amateurs. And as Ken Ledeen explains, Hedy Lamarr’s social circle included a number of leading professionals from whom she absorbed a great deal of knowledge and understanding.

Still, if a pair of inspired and talented amateurs could invent such a thing, maybe we should have more faith in the ability of most people to understand its basis, and to reason about its implications.

Serious uses for YouTube’s new popup video feature

I’m loving YouTube’s new video annotation feature, which Phil Shapiro alerted me to. Lots of people are going to have lots of fun with that. If you remember when MTV first started doing popup video, you’ll have some idea how much fun.

But from Phil’s perspective and mine, this is a seriously useful tool as well. He’s planning to annotate screencasts with it. And I found a great use for it here.

That short video features Bob Coffey, the senior climber at our YMCA. When I made and posted the video, I wasn’t quite sure how senior Bob was so I didn’t say. Yesterday I remembered to ask. Turns out he is 79.

It’s painful to add new information to a video. Opening up the raw file (if you even kept it around), adding a caption, recompressing, reuploading — it’s too much overhead, and unless there’s a compelling need you’re just not going to bother.

Of course you can update the textual wrapper, and alter the title or description. But in this case, I didn’t want to that. The information is much more effective when inserted midstream. After he’s scampered halfway up the wall, the popup annotation saying “Bob Coffey is 79 years old” makes the point more subtly and powerfully.

The point, by the way, is that we can do more, physically, at all ages, than we think. I’ve known a few people over the years who have redefined what’s possible, and it’s always an inspiring thing to see.

Questions for Exchange admins about public calendars

To complement my series on client-side calendar publishing, I’ve been looking for a way to push ICS files from Exchange. Why? A couple of local organizations with calendars I’d like to include in my calendar syndication project are running Exchange. It’s true that individuals within those organizations can use Outlook 2007 to publish calendars to the Internet. But companies like to manage these processes centrally. If the city wants its recreation center to publish summer activities, the city’s IT department should be able to do that whether or not the rec center’s desktop machine is alive.

Several folks suggested that the WebDAV interface to Exchange is the way to go, and after noodling on that for a while I’ve come up with a solution that seems to work. (I’ve tested it on a server here in my lab, and against my account on a production server.) I’m planning to release it as a CodePlex project, but as a Microsoft employee I have to cross a few t’s and dot a few i’s to get that done.

Meanwhile, since I’ve never administered Exchange, I have some questions for those who have. What I’ve got is a 100-line IronPython script that can be run on a scheduled basis. You give it a hostname and an account name, it reads events from that account’s calendar, and emits an ICS file.

First question: How does this fit in with your workflow? If you’re a town government, how would you like to manage your public calendar from Exchange? For example, would employees in various departments share one account that corresponds to that calendar? Would you rather divvy things up and then syndicate a set of calendars?

Second question: How will you want to publish the ICS file? My script only creates it. To publish the file to the Net, it could be copied to a directory on a webserver, or it could be uploaded to lots of places in lots of ways. It’s good to have choices, but it’s also good to have a sensible default, and I’m not sure what that should be.

Third question: Would you deploy this thing? I hope the small footprint will help. Being wary myself of solutions that haul in boatloads of requirements, I’ve whittled them down. You only need IronPython, plus of course the .NET Framework which it requires.

Originally, I was also using a few things from standard Python: command-line arguments, regular expressions. But it struck me that requiring all of standard Python for just those things was overkill, so I switched to their IronPython/.NET equivalents.

IronPython itself, resting as it does on the .NET foundation, is a tiny installation. Even so, it’ll be a new and unfamiliar thing to most people. Would an Exchange admin be willing to install IronPython on a production server and allow it to do the things this script does?

A conversation with John Buckman about Nazi invaders from the moon (and other things)

My guest for this week’s Interviews with Innovators is John Buckman, a serial entrepeneur with a passion for the world-changing possibilities of online communication. I was a customer of his first company, Lyris, whose email list manager I once deployed for a client. A few years later we met at SXSW, where John showed me his new project, Magnatune, and explained why he created it.

In this interview we discuss Lyris and Magnatune, plus some recent interests: BookMooch, an online book exchange, and Wreck a Movie, a “Web platform that is designed to harness the power of passionate Internet communities for creating short films, documentaries, music videos, Internet flicks, full length features, mobile films and more.”

This platform for crowdsourced film production, which is an outgrowth of the collaboration that led to Star Wreck, is now being used to create a new film called Iron Sky. You gotta love the tagline: “In 1945 the Nazis fled to the moon. In 2018 they are coming back.” As John says, who wouldn’t want to see Nazi invaders from the moon?

His ideas about how to do crowdsourcing closely parallel those of Yochai Benkler, as channeled by Bruce Sterling in his SXSW 2007 keynote, which includes a digression on this theme that begins thusly:

Socially motivated commons-based peer production: How to do it. You don’t just open up a website and invite comments. It actually has to be engineered with some thought and care. And I want to explain how this guy [Yochai Benkler] thinks it’s done. First you have to divvy up the work, because there’s a lot of it, and nobody wants to do it, and you’re not paying them, and you’re not ordering them, because you’re not the market and you’re not the state, so you can’t pay them and you can’t draft them. So you have to divide it up. It has to be granular, modular, and integratable. Granular means that even if I contribute for 5 minutes, I’m going to contribute something of merit.

Software is part of the answer. You need a framework within which to define tasks, parcel them out, and gather back results. Amazon’s Mechanical Turk is the best-known commercial example, but there are noncommercial precedents like distributed proofreaders and LibriVox. It’ll be interesting to see how Wreck a Movie follows and extends those examples.

John affirms Yochai Benkler’s point about granularity, and stresses the importance of precise task definitions.

If you say, nebulously, “Help us get press,” then nobody does anything. If you very specifically say this is the message, post to these press outlets, and name three URLs, that’s a well-defined thing. You get much stronger response. People need direction.

Also, I read recently that membership in volunteer organizations is down, and has been for a decade, but the amount of time people put into volunteer tasks is up. Why? People are less interested in open-ended commitments, instead they’re interested in knocking off specific tasks, and feeling a sense of accomplishment for finishing them.

As Bruce Sterling says, you have to engineer that scenario with thought and care. It’s a kind of engineering that is both social and technical, and some of the leading practitioners are game designers. For more on this theme, listen to Ned Gulley’s reflections on the MATLAB programming contest he’s been running for a number of years.

LiveMesh and FeedSync: software “above the level of a single device”

When David Stutz left Microsoft, he wrote a parting essay that invoked a new kind of Internet-oriented operating system characterized by “software that runs above the level of a single device.” Tim O’Reilly echoed that phrase here, and often used it to help explain what he meant by Web 2.0.

The recently-announced LiveMesh is a nice example of software that runs above the level of a single device. It runs symmetrically on all your computing devices, in the part of the cloud that’s associated with your devices, and in other parts of the cloud where services you transact with are running. This entire constellation is the LiveMesh platform which, as Ray Ozzie recently explained to investors, is an answer to this question:

What would an OS look like in a world of multiple devices, in a world where instead of the computer being at the center, you are at the center?

The platform’s connective tissue, as discussed in my interview with Ray Ozzie, is FeedSync, a synchronization system based on the same simple technology that powers the blogosphere: XML feeds of items, in RSS or Atom formats. Whether they represent big chunks of information like documents and media files, or small scraps of information like calendar events and status messages, LiveMesh objects are made up of feeds. All these objects synchronize across your mesh of devices and services using the same openly-specified FeedSync mechanism.

That openness is another key characteristic of an Internet operating system. So it’s nice to see that FeedSync isn’t only being applied in the context of LiveMesh. This week’s Perspectives interview, with Barbara Willett of Mercy Corps and Nigel Snoad of Microsoft Humanitarian Systems, details Mercy Corps’ use of FeedSync to collect, synthesize, and share information about the management of agricultural development programs in Afghanistan.

In this case, the synchronized data sources are humble Access databases for which Nigel and his team have developed a FeedSync adapter. They’ve also built adapters for spreadsheets and — what should be very interesting to Ken Banks and Joel Selanikio — for SMS messaging systems.

I love stories about pragmatic solutions that find new ways to use existing, simple, and widely-deployed technologies. This is clearly one of those. But it also illuminates an aspect of the LiveMesh platform that hasn’t yet been widely noticed or appreciated. One of its keystones, FeedSync, is an open and general-purpose building block that can be used by anyone, for any purpose.

If the Mercy Corps solution interests you, Nigel says that the toolkit he and his team built for them will be openly released in a few weeks. Watch the FeedSync blog for details.

Drafting on bloggers

The word drafting has many meanings but the one I’m interested in here comes from bicycling. When you ride closely behind another rider, you’re drafting. The leader pushes the headwinds out of the way, and the follower doesn’t have to push so hard.

Blogging can work that way too. I thought of this when I reallized that one of the benefits of subscribing to James Fallows is that I’m drafting on his interest in the fledgling air taxi industry. My interest in that topic is more than casual. I’ve interviewed DayJet founder Ed Iacobucci, for example. But James Fallows is way more deeply invested than I am, having written the seminal book on the topic, Free Flight.

I can do a pretty good job of tracking developments on that front by scanning the news, or better yet by subscribing to searches for terms like DayJet and Eclipse 500. But the best way is to draft on a blogger who is authoritative on the topic.

I don’t need to see every news story about DayJet, and pushing them all out of the way in order to focus on the ones that really matter is like pushing a headwind. But James Fallows is already motivated to push that headwind, so I can just draft on him. That way I get just the right air taxi newsfeed, with a dollop of expert analysis on top.

Happily, the analogy breaks down in a couple of ways. A cyclist can only draft on one other cyclist, and it’s a one-way relationship. The follower can’t simultaneously lead. With blogging, I can draft on many peoples’ interests, and many people can draft on mine, and sometimes the leader/follower relationship is reciprocal — I draft on you for topic A, and you draft on me for topic B.

For our purposes here, we can define blogging broadly to include the conventional format, but also microblogging formats like del.icio.us crumbtrails and Twitter tweets. Drafting, in the sense I mean here, can happen in any publish/subscribe medium.

In search of an ICS publishing component for Exchange

As part of my project in community calendar syndication, I would like to find a way to push an Exchange calendar to a web-accessible ICS file. Although that isn’t a native function of Exchange, I’m sure it can be accomplished by way of the Exchange API, as an add-in or a scheduled server process. For maximum breadth, I guess the relevant API would be Collaboration Data Objects (CDO) rather than the newer Exchange Web Services.

I asked some folks on the Exchange team and none were aware of a component that does this. So I’ve set up an Exchange server in my lab, rolled up my sleeves, and am ready to dive in and see what can be done. But it never hurts to ask. Have I overlooked an existing off-the-shelf solution?

Free online calendar publishing, part 3: Apple iCal

This post is part three of a series in which I’ll summarize what I know about publishing calendars openly on the web, for free, using popular calendar applications including Outlook, Google Calendar, and Apple iCal.

Apple iCal

If you have a .Mac account you can publish your calendar there, but
.Mac isn’t free, and the purpose of this series is to showcase free
calendar publishing options.

The solution here is to find a free service that uses the same
protocol, and the same kind of server, as the .Mac service uses.
The protocol is called WebDAV, and the server is a
special-purpose web server commonly used for calendar publishing.

It’s possible that your ISP already offers a WebDAV server you can
use, for no additional charge. But for many people it would be ideal if there were a free service available for this purpose. One such service is iCal Exchange.

Here’s the signup page:

After completing the form you’ll land here:

For private sharing you can create passwords and use the private URL, but our goal here is public sharing so the public URL — in this case, http://icalx.com/public/judell — is the one you’ll use.

Now switch to iCal, select Calendar, select Publish, and switch the Publish option from the default — .Mac — to Private Server. Paste the public URL you just created into the Base URL, and enter your iCal Exchange credentials in the Login and Password fields.

Now click Publish. Here’s the outcome:

The web address for your calendar is the public web address that iCal Exchange gave you, plus the name you gave your calendar (in this case, Jon), plus the .ics extension. For this example, it adds up to webcal://icalx.com/public/judell/Jon.ics.

This solves half of the problem. Your calendar is now published in a way that enables individuals to subscribe to it. It’s also available for syndication by online services like http://elmcity.info/events.

But the other half of the problem remains unsolved. In parts one and two of this series, we saw that the free calendar hosting options offered by Microsoft and Google provide links to hosted calendar viewers.

There are a variety of other hosted viewers, but as yet I’ve not found one that’s free, and can render any public calendar given a public URL like the one shown here. If such a service does exist, I’m hoping this entry will help me find it.