Motivating people to write the semantic web: A conversation with David Huynh about Parallax

On this week’s Innovators show I got together with David Huynh, whose work with MIT’s Project SIMILE wowed me last year. David recently joined Metaweb. His first project there, Parallax, creates a new way to browse Freebase, the structured wiki that also wowed me earlier last year.

What struck me about SIMILE and Freebase was the way in which both projects cut through the fog of semantic web technologies and terminologies and got down to brass tacks: How do you get people to want to contribute structured knowledge? You have to appeal to natural instincts and, as I explored in my writeups of both projects, they do.

In the case of SIMILE, when somebody sees an interactive, data-rich Exhibit, the natural response is: “Cool! How do I get one of those for myself?” The answer is: “Pretty straightforwardly, by cloning the example you see and then massaging your data into an easily-written format.” The hidden agenda is: “You may not know or care, but once your data is in that format, it can federate.”

In the case of Freebase, the reciprocal nature of data relationships creates a kind of social glue. People who contribute to Wikipedia, or Freebase, or to the web in general, hope those contributions will be read and appreciated. The genius of Freebase is that, when I define a relationship between one of my records and one of yours, both come into view. I may notice something missing from yours and add it. You’ll in turn notice my contribution and may reciprocate. As we advance our own interests, we naturally find ourselves advancing others too.

Parallax, as you can see in the screencast David has embedded on the project’s home page, is a new way to explore Freebase. In the standard interface, you have to do some digging to trace connections to related sets of information. Or you might even have to drop into the API to do programmatic search. Parallax brings those relationships to the surface. One of the examples in the screencast asks and answers the question:

What were the schools attended by children of Republican presidents?

This boils down to a query that finds a set of Presidents, then a set of Republican Presidents, then a set of children, then a set of schools. With Parallax you perform that query interactively, by following links that surface as you pivot from set to set.

This is a great way to browse the structured corpus, but how does it motivate people to provide more and better contributions? Here’s one way: By exposing the completeness — or incompleteness — of sets viewed in relationship to one another.

As I was browsing the set of U.S. Presidents, for example, Parallax surfaced the connection Works written. But there wasn’t much there: Jefferson and Adams for the Declaration of Independence, Madison for the Federalist Papers, and a few recent books by Jimmy Carter and Bill Clinton.

So I created an item in Freebase for Richard Nixon’s Six Crises, linked it to Nixon’s record in Freebase, and went back to Parallax. Sure enough, there were Nixon and Six Crises. The set of books written by U.S. Presidents had increased by one. Along the way, a new book record was created in Freebase, and an existing person record was enhanced.

As I mentioned to David in our interview, this strikes me as a really powerful way to motivate contributors. In Wikipedia there’s no easy way to observe an implicit set. You can only look at explicit sets, like the lists of Presidential ages and religious affiliations. Somebody might decide to make an analogous list of Presidential books, but that would be much more likely to happen if the partial list that’s already implicitly in Wikipedia could be brought into focus.

When shown partial patterns, people naturally want to complete them. Parallax looks like a great way to tap into that instinctive urge.

Collective mind probes

If you hang out in the blog-twitter-sphere you’ve doubtless seen recent mention of brand tags, Noah Brier’s “collective experiment in brand perception”. The site shows you brand logos and asks you to tag them with the first word or phrase that comes to mind. Then it shows you how others tagged those brands.

It’s addictive. Here, for example, are the top tags for Jet Blue and United Airlines:

Jet Blue: cheap, plane, airline, fly, blue, airplane, tv, cool, fun

United Airlines: airlines, airline, airplane, fly, plane, air, old, bankrupt

These slices of collective perception are remarkable indicators. Whose perceptions do they indicate, though? Not only, as I guessed, denizens of the blog-twitter-sphere. Look at what the brand tags crowd says about Twitter, TechCrunch, and ReadWriteWeb:

twitter: useless, annoying, stupid, web 2.0, fun, bird, blog, ?, tweet, internet, pointless

techcrunch: ?, blog, tech, technology, nothing, no idea, geek, computer, boring

readwriteweb: ?, nothing, huh?, no idea, rww, internet, what?, unknown, boring

Evidently the site is tapping into a broader population that doesn’t know about Twitter, TechCrunch, and ReadWriteWeb. From that perspective, it’s interesting to see which emerging tech brands have crossed over. Flickr, for example:

flickr: photos, pictures, photo, cool, picture, pics, fun, awesome

What’s a good result for your brand? Here are two tests:

  1. Does the list begin with a question mark? That’s bad.
  2. Do the words “fun” and/or “cool” and/or “awesome” and/or “amazing” occur near the top of the list? That’s good.

You won’t be shocked to discover that Microsoft passes test #1, but test #2 not so much, though I predict that will change, and I intend to be part of that change.

Likewise, you won’t be surprised to learn that Google and Apple pass both tests with flying colors. Less predictably, though, the tag sets for Google and Apple reflect a fascinating ambivalence. The seventh and eighth tags for Google, for example, are “god” and “evil”.

The brand tags experiment itself is fun and cool and amazing. Also disturbing. See the Oprah entry, for example. I’m not sure how long the fun will last without spam controls.

Such controls could be implemented. But if an experiment like this lasts long enough to be gamed, and then to evolve countermeasures, it may have outlived its usefulness. Perhaps the most effective ways of probing the collective mind will be short-lived, non-replicable hacks that yield brief flashes of insight.

The New Yorker, by the way, recently reported on another fascinating probe. In Obama’s In-box, Charles Bethea tells the story of Guru Raj who, on a lark, registered the email address barackobama@gmail.com back in 2004 when Gmail was new. Raj never replies to any of the mail received at that address, but he used to read it, and still does from time to time. Here’s how he describes the experience:

“It just became an interesting portal into Americana.”

The “myth” of free standards

In the domain of Internet software world we take it for granted that standards documents are available for anybody to read. But in other domains that’s not the case, as I was reminded when I landed on this URL, where a copy of EN 303-5 — a European standard that governs “heating boilers for solid fuels, hand and automatically fired, nominal heat output of up to 300 kW” — is offered for £180.

Here’s how I got there. I’ve imported a wood gasification boiler from Europe, and I need to show my city that it’s a listed and rated device. Well it is, but it doesn’t carry ASME and UL certifications, it carries a TUV certification. I wanted to show what that means. The Declaration of Conformity that ships with the device lists a batch of relevant standards documents. I looked them up, and they’re all online but for similar prices. It’d cost me a couple of thousand dollars to read them all. That’s not going to happen.

I wound up printing out copies of the three Directives that govern the application of those standards, which have names like 97/23/EG, 73/23/EWG, and 90/336/EWG. These, I hope, will show the code enforcement officer that my boiler is tested and certified according to requirements comparable to those in the U.S.

But it made me stop and think. Broadly speaking, open access to standards documents is the exception, not the rule. In The Myth of Free Standards: Giving Away the Farm, Andrew Bank — writing for Techstreet, a Thomson Reuters company that sells copies of industry codes and standards — argues that’s a good thing. Acknowledging “a growing consensus among standards developers and users contends that making standards available at no cost will further their use and development”, he argues that a fee-based publication system:

  1. Underwrites the expensive process of standards development
  2. Maintains the perceived value of standards
  3. Treats standards like all other required inputs to product development
  4. Provides copyright protection
  5. Subsidizes membership in standards organizations

I don’t think any of these arguments will stall the growing consensus in favor of open access. But it’s interesting to review them, reflect yet again on how the Internet realigns publishing businesses, and imagine what new structures will emerge.

Paving the cowpaths that lead to Government 2.0

Government 2.0 was the theme of couple of recent podcasts. On Phil Windley’s Technometria show, Britt Blaser discussed his Independence Year project, which places “the organizing tools of the best political campaigns” into the hands of citizens who want to organize themselves. And on my own show I spoke with W. David Stephenson, a veteran political operative who has recently focused on the same theme.

On both of these shows, I harped — too much — on how online tools and services do, and don’t, enable citizens to make sense of, and engage with, the processes of governments. But of course, as Adina Levin helpfully reminds me, it isn’t all, or even mostly, about tools and services. Adina’s comment:

Call the Senator’s office. And don’t just talk to the person who answers the phone. They are motivated to take your comment and end the call. Find the staffer who specializes in the policy area. Get their name. Get their email address. Contact him or her directly. They’ll know, and be able to answer the content and procedure questions if you ask nicely and intelligently.

Once you find the information, posting to your blog, hyperlocal news outfit, or political group blog will spread the information and increase the social context about the sausagemaking process.

The net enables new tools, but it also makes this sort of older technique informative for more people.

Elsewhere she adds:

Udell points to tools like GovTrack which are attempting to create a substrate for communities following bills. I’m seeing a trend that is fascinating and a little bit lower tech. National blogs like FireDogLake, local blogs like TransBayBlog, social network communities like the Get Fisa Right network in MyBarackObama.com, provide their communities with more detailed context on the dynamics of legislation and the process of adding ingredients to the sausage. In the context of a community, members learn more about the legislative process than civics 101 class, or than getting email from the Sierra Club.

Points extremely well taken. Those of us wired to imagine and build advanced tools and services are invariably surprised when The Simplest Thing That Could Possibly Work turns out, again and again, to actually work.

One of my favorite examples here is LibriVox, a hugely successful collaborative project that to this day runs mainly on a PHP-based bulletin board. It’s a perfect illustration of what Clay Shirky points out in this interview:

A brutally simple mental model of the software that’s shared by all users turns out to be a better predictor of adoption and value than a completely crazy collection of features that ends up being slightly different for every user.

Of course tools and services matter. The trick is to introduce them in incremental and appropriate ways. On this forum page at LibriVox you can see a nice example of that incremental approach. It’s the Completed Projects page for Edgar Rice Burroughs’ Pellucidar. No database-backed workflow software created that forum page. There’s just a convention, evolved organically by LibriVox participants working together, about using certain forum topics to represent certain phases of workflow.

As I understand it — and Hugh McGuire will correct me if I’m wrong — the project ran for a while with no other database than the one supporting phpBB. Participants simply agreed to use that database according to certain agreed-upon patterns.

Later, a separate database was developed to maintain the catalog in more flexible and powerful ways. But look at what’s happening on that forum page I mentioned. A project view from the database is wrapped in an iframe and included within the bulletin board page. It’s a sweet example of how to inject a more advanced service non-disruptively into a simple system.

In the realm of Government 2.0, as elsewhere, advanced tools and services can play crucial roles, but only if we can weave them into the patchwork quilt of brutally simple systems that people understand and use.

A tale of two cities: crime reports in San Francisco and Washington DC

Consider two analytical reports from CrimeReports.com.

San Francisco, last 30 days

Washington DC, last 30 days

A number of oddities leap out, but most of all look at the scale on the X axes. It runs to 30,000 for San Francisco, and only 800 for Washington DC. In absolute terms the numbers of thefts reported are comparable: 732 for Washington, 952 for San Francisco. But some non-reported categories from Washingon skew the comparison wildly. The numbers for Traffic and Proactive Policing, for example, are zero for Washington, but 17264 and 28356 for San Francisco.

Now to be sure, CrimeReports offers this disclaimer:

This information is for interest only, and is subject to error and change based on the validity of the data from the Department.

Nor does the service invite or support this kind of comparison. To make it, I had to combine views taken separately from the two jurisdictions.

Apparently there’s a problem with the Washington DC data. There cannot have been no traffic offenses or burglaries in the past month. But in this case, we can check for ourselves. The DC data is available to anyone.

The San Francisco data certainly looks saner. But if you wanted to verify it, you couldn’t. The San Francisco data isn’t similarly available.

Here’s how I concluded the writeup of my interview with Greg Whisenant, the founder of CrimeReports.com:

What worries me a little, though, is that CrimeReports implementations don’t (so far) yield up feeds of the underlying data. I understand the reasons why not. But I think it’s crucial that citizens will come to expect such access, and will be encouraged to make effective use of it.

It’s not Greg’s call whether to publish that data or not. Jurisdictions decide for themselves. When data is publicly available, CrimeReports.com can base its maps and analytics on the same information available to everybody else. Otherwise, and almost universally so far, CrimeReports.com plugs into police databases behind the firewall, and produces maps and analytics, but citizens still can’t get to the underlying data.

Greg says he is perfectly willing to provide such access, if police departments want him to. So far they don’t, though he expects this will change slowly over time. Meanwhile, nobody else can get to that data. That means, among other things, that in San Francisco, EveryBlock has been frustrated in its efforts to include block-level crime data in its hyperlocal news feeds.

There are all sorts of reasons why police departments won’t want to publish raw data, and I can sympathize with some of them. Privacy, even when data are aggregated to block level, is a concern, given the many surprising ways in which deanonymization can occur. And misinterpretation is another valid concern.

Personally I think that the benefits of transparency outweigh these risks. But it doesn’t matter what I think. Open access will happen because an administration (like DC) decides on that policy, or because a groundswell of citizen sentiment demands it. I haven’t seen an example of citizen demand driving the decision, but maybe San Francisco will be a bellwether on this issue as it has often been on others.

Should I visualize more weather data, or just build an ark?

Last July the crazy weather prompted me to examine my subjective notion that things were, indeed, going crazy. However, a cursory analysis of NOAA’s historical temperature and precipitation data, available back to 1921 for the Concord, NH station that’s closest to me, didn’t seem to reveal any notable patterns.

Well, we’re having another crazy summer, and I took another look at that data. This time around I used Excel to do some deeper massaging, and here’s what I’ve found:

Monthly precipitation for Concord NH, 1921 to 2008

(top 15%)

rank jan feb mar apr may jun jul  aug (1-10)
1 1979 2008 1937 2007 2006 1944 1938 2003
2 1978 1981 1936 1936 1984 2006 1940 1973
3 1935 1951 1953 2004 1954 1998 1988 1953
4 1958 1984 1983 1933 1940 1922 2008 2008
5 1999 1970 1942 1996 1945 1982 1928 1946
6 1936 1971 1925 1983 1922 1935 1967 1980
7 1953 1955 2008 2005 1968 1945 1951 1937
8 1996 1929 1933 1951 1937 1968 1981 1977
9 1986 1969 1994 1940 1983 1987 2007 1969
10 1952 2003 1922 2000 2004 2001 1972 1943

I ftp’d the data from ncdc.noaa.gov/pub/data/ghcn/daily/all/USC00271683.dly, and used Python to parse the fixed-length records into a tab-delimited format like so:

year month day precipitation

Then I brought it into Excel. I’m no kind of Excel wizard, so there are surely better ways to proceed, but here’s what I did. I started with a pivot table where values are the amounts of precipitation, row labels are years, and report fields are month and day. Next I reverse-sorted the rows by amount of precipitation, and restricted them to the top 10 years. Then I filtered month by month — and in the case of August, by just the first 10 days for which 2008 data is available — and copied each month’s top 10 years of precipitation into the final table. Finally I selected that table and used conditional formatting rules to colorize the top 15% of year/month combinations.

Wow. It really has been a snowy and rainy decade around here. And 2008 takes the cake.

Are there easier ways to do this with NOAA’s climate data? From time to time I look, and I never find anything. Maybe I should turn this spreadsheet into an application that’ll search NOAA for your local station, grab your temperature and precipitation history, and visualize it like this.

Or maybe I should just go and build an ark.

Seeing more Olympics means seeing the Olympics differently

Thanks to the expanded coverage at nbcolympics.com, we’re not just seeing more sports, we’re seeing sports differently. The podium training for men’s gymnastics, for example, was a behind-the-scenes view that you’d never see on TV.

Now that we’re into the team competitions, another new perspective is emerging — and it’s been a revelation. As we see all the performers, not just the top echelon, I’m struck by the number of blown routines. I’ve felt for years that the sport has been emphasizing extreme difficulty at the expense of form, grace, amplitude, style, and rhythm. Watching the whole competition on high bar has confirmed that. I think you should approach an apparatus with a high degree of confidence that you can complete your routine. It shouldn’t be a crapshoot. And yet, apparently, that’s what it has become for many of these performers.

Did audiences vote for this outcome? If so, they’ve probably done so without full knowledge of the consequences. It’ll be interesting to see whether an expanded view might change attitudes.

As things stand, it’d be hard to compile a full analysis of all the routines, and to annotate the outcomes. It’s great that NBC is making these long feeds available, but there’s a lot of dead time in a 2.5 hour team competion on high bar, much of it spent watching judges discussing what they’ve seen and deciding how to score it. NBC doesn’t have the resources or motivation to distill and annotate all the footage it’s providing.

But fans and practitioners do. In high school we studied Super-8 films that our coach shot at an Olympics. With the video feeds now coming online, along with link-based editing that creates annotated playlists and a MTurk-like task allocation to distribute the work of annotation, that analysis could become a crowdsourced activity.

So this is the Olympics where I finally get to at least see so much of what formerly was hidden, and it’s a great thing! For the first time, online video has displaced TV as the viewing experience of choice.

Next time, I hope and expect, I’ll get to participate in the co-creation of views that will add value to the expanded coverage, and that might support a conversation about what kinds of gymnastics people really want to see.

A space elevator might arrive sooner, and cost less, than you think

I’ve always thought that Arthur C. Clarke’s space elevator idea was intriguing, but until recently I never thought much more about it. Then I heard that Microsoft was co-sponsoring, and hosting, the 2008 Space Elevator Conference. This annual event isn’t a science fiction convention. It gathers a diverse group of passionate scientists and engineers who have, in recent years, morphed Clarke’s original vision into something that could cost less, and arrive sooner, than you think.

I wound up doing two separate interviews on this topic. On my Perspectives show, I spoke with Ted Semon, a retired software engineer who runs a blog that chronicles the movement to develop this new version of Clarke’s idea. And on Interviews with Innovators, I spoke with Maurice Franklin, a Microsoft performance and scalability engineer whose passion for this topic brought the conference to Microsoft this year.

I have no idea whether a space elevator will be built in my lifetime, if at all. And while it would surely be a game-changer, I’m not sure just which games it would change, or how. Energy? Economy? Climate? All are conceivable.

If we could in fact do this for $10 billion (a small fraction of the cost of the Iraq war), and do it in 20 years (a small multiple — I hope! — of the final duration of the Iraq war), should we?

Again, I don’t know. But it would be an incredible option to have on the table. It’s fascinating to learn how an international coalition of innovators is working to create that option.

The fact that Microsoft hosted and co-sponsored this conference, by the way, does not imply any kind of corporate endorsement of the space elevator effort. This was purely a grassroots thing. Maurice Franklin wanted to help, George Spix thought it was an interesting event to bring to campus, and so it happened.

Using technology to explain technology

Recently I gave a presentation, on a topic that is web-oriented and somewhat technical, to a group that is very much non-technical. I knew slides wouldn’t be an option, because there wouldn’t be a projector in the room. But arguably a slideshow would be the wrong thing anyway. I wanted people to focus on me, and on the message I was delivering, not on a slideshow.

So I wound up writing a twelve-page Word document that provided me with speaking notes, and also served as the handout I left behind. For me, it was an odd reversal of my usual procedure, and a bit of a revelation about Word’s ability to wrangle web content into a nicely printable form.

Typically I make slides, often heavy with web content, that I can show and then publish to the web for all the usual reasons: linking, search, bookmarking, syndication, online social interaction. Printing has been a low priority because I have mostly spoken to groups of digital natives who, like me, rarely bother to print.

In this case, I didn’t expect much follow-on linking, search, bookmarking, or online social interaction. It was more likely that copies of the printed document would get passed around, or maybe downloaded and printed.

But much of the content of that printed document still had to come from the web. My first instinct was to embed screenshots of web pages into the Word doc. But if you do it that way, you can’t optimize text for readability in the printed output. So I wound up copying chunks of HTML (along with images) from web pages, and pasting them into Word. I guess I haven’t tried this method in a long while. The translation isn’t perfect, but for simple layouts — and even for some not-so-simple table- and CSS-based layouts — it worked pretty well!

There remained the question of how to publish this printable document for downloading. I figured that PDF would be the most comfortable option for this group, so I used Word 2007’s PDF exporter — for the first time, actually — and that worked nicely too.

This experience highlighted an interesting way to use Word. I’m sure folks will chime in with a variety of other tools and methods, and that’ll be interesting too.

The real lesson for me, though, is one that I’m continually trying to learn these days. I need to make a cleaner separation between the technologies I use to explain things, and the technologies I’m trying to explain. Sometimes they’re appropriately the same. But not always.

Behind the scenes: podium training for men’s gymnastics in Beijing

As a former gymnast, I’ve been quadrennially frustrated by TV coverage of the Olympics. I don’t need the “up close and personal” portraits of a few popular gymnasts. And I don’t need Elfi Schlegel and Tim Daggett to point out some subtle things I already know, or to state the obvious that anyone can see for themselves (“oops, a little hop on the landing”). Most of all, I don’t want to be restricted to the minimal subset of gymnastics performances that was historically all that got televised.

I’ve always dreamed of being able to watch raw footage of every performance, and even — why not? — all the warmups.

Finally that moment is at hand. A colleague writes:

It is midnight and we now broadcasting the gymnastics trials live from Beijing.

http://www.nbcolympics.com —  click here to watch men’s gymnastics trials, then the woman’s trials & if you want the soccer game at 5am.

It has been a crazy ride!

As I write this entry on one monitor, a Silverlight viewer on my second monitor is displaying a 2.5-hour replay of the podium training for men’s gymnastics. That’s the one chance for the teams to practice on the floor of the venue, before competition begins.

It’s like nothing you’ve ever seen on TV. The gymnasts are in their competition uniforms, but the hall is mostly empty — except for other gymnasts. You see the warmups, and you see the practice routines. And you hear whatever there is to hear. At the moment, I’m hearing members of the US team calling out “Go Raj” as Raj Bhavsar practices his parallel bar set. I’ve been there. I know what goes on behind the scenes. Well, the high school and college scene, that is, not the Olympic scene, but … tuning in like this is a dream come true.

If you’re not deeply into men’s gymnastics this won’t matter to you. But it’s certain that, for whatever Olympic sports you do care about, the televised coverage has in the past always fallen far short of what you would have liked to be able to see.

In an NBC press release, Dick Ebersol said:

For the first time, the average American will be able to create their own unique Olympic experience whether at home, at the office or on-the-go.

It’s staggering to me to be involved in a Games where we are producing 2,900 hours of live coverage – especially from an Olympics half-a-world away. It’s more live coverage from a single Olympics than the total of all previous Summer Olympics combined. The enormity of what we’re doing just blows me away.

Me too. This is the Olympics I’ve been waiting for since the dawn of the Internet.

Go Raj!

Social dimensions of P2P and the semantic web

At the turn of our new century, during the heyday of Napster, the infotech conversation was all about peer-to-peer, or P2P, technology. There were elite gatherings of leading influentials to assess its potential, conferences to bring together ideas and money, new publications to chronicle its development.

In September 2000 Dave Winer wrote a touchstone essay called The P in P2P. “It’s not about P2P, specifically, that people should be getting excited,” Dave wrote. “The killer app of P2P is … People!”

When Groove 1.0 launched I was asked by a media producer to explain why Groove was an important P2P application. Channeling Dave, I said that Groove was important because it empowered people to work together effectively, not because it was a poster child for P2P.

Back then, we lacked the cloud infrastructure to support the loosely-coupled networks of personal services I imagined. But that infrastructure is emerging now. I see it, for example, in the P2P capabilities of Live Mesh.

Nowadays, of course, we rightly no longer regard P2P as a silver bullet. But the infotech tribe is wired to seize on some silver bullet, and so a new one is being loaded into the chamber. It bears an inscription: “the semantic web”.

For good reason! We experience, many times every day, the transformative power of search based on nothing more than the ability to ask and answer the question: “What document contains the word X?” Questions can be a whole lot more interesting than that, and answers can be a whole lot more useful.

On the semantic web, the notion is that queries will traverse ad-hoc collections of data in which the relationships among items of information — within and across datasets — are made explicit. To use the phrase that comes up again and again in these discussions, we’ll be able to “reason over” this web of linked datasets.

We’ll get there, I hope, but let’s not conflate means with ends. During the initial P2P craze we acted as if the P2P web was a goal. Today we see that it’s one of many architectural styles. We use it where appropriate, in concert with other styles, and we’re a bit clearer — I hope — that the killer apps are the ones people make when living and working in a networked world that includes the possibility of P2P.

So it will go with the semantic web, I predict. At the moment it seems like a goal, a grand challenge. And of course it is. But the killer apps will be those that people make when using — and here’s a twist, when creating — the semantic web.

That’s a key distinction. Users of the P2P web don’t, for the most part, build its enabling infrastructure. But users of the semantic web will. Yes, networked computers will weave the basic fabric of linked data. But we’ll need to do the context assembly that makes the semantic web useful to us.

We may be P2P’s killer app, but we’re going to be the kernel of the semantic web’s operating system. That’s why the engineering of that operating system will have to be as much social as technical, and maybe more so.

Lee LeFever reflects on the CommonCraft style

On this week’s Innovators show I spoke with Lee LeFever who, with his wife Sachi, runs CommonCraft. You’ve probably seen their clever animated videos explaining web topics like RSS and Twitter. If not, check out their most popular one, RSS in Plain English. It’s uncannily effective, and a real inspiration to those of us who’ve struggled to explain this concept to civilians.

As I mentioned here, and as we also discuss in the interview, the CommonCraft style shares some DNA with the practice of paper prototyping as described by Bill Buxton in another interview. In Bill Buxton’s case, you abstract away unnecessary detail, and convey action, by sketching and manually animating parts of the interface you’re designing. In Lee’s case you use the same methods downstream, not to elicit the key concepts to be embodied in the yet-unbuilt software, but instead to explain those concepts after it’s shipped. Two sides of the same coin, I think.

A conversation with Tony Hey about Microsoft External Research and the new breed of e-scientists

In my Perspectives series I’ve been exploring some key relationships between Microsoft and external partners: Mercy Corps, NASA, the Library of Congress, the community of scientific and scholarly publishers, and individual scientific researchers including George Hripcsak and Dennis Baldocchi.

For this week’s show I spoke with Tony Hey, the VP of the External Research division of Microsoft Research. More than anyone else, he’s responsible for advancing what I’ll call the Jim Gray vision: All science is becoming computational and data-driven; a new breed of scientists requires a new generation of tools and services; Microsoft can and should rise to the challenge.

In coming weeks and months I’ll be reporting more of these stories of collaboration between Microsoft Research and scientific researchers. This is a treat for me because, little known fact, during the early part of my software career I took a hiatus to earn a graduate degree in science writing. I’ve always been curious about how science works, and this is an opportunity both to learn and to share what I learn.

It’s also an opportunity to expore, and to highlight, a side of Microsoft that you’ll never read about on TechMeme. When Tony Hey recently counted up the number of Microsoft Research engagements with university partners, the total exceeded 400. It’s an impressive outreach effort, and one that will benefit everyone: Microsoft, the scientists whose research work is enhanced by these projects, and all of us who depend on the outcomes of that research.

Overcoming synthetic voice shock

To prepare for my interview with Susan Gerhart I tried using text-to-speech software to read menu choices and text selections aloud. As always, I experienced the reaction that Susan, in her latest post, calls synthetic voice shock.

For those of us who don’t need to rely on synthetic voices, that reaction isn’t a problem, it’s merely a deterrent to optional use of the technology. For example, though it might be convenient to shift some material from the domain of written text to the domain of audio, the unpleasantness of synthetic voices stops me from doing that.

But the real problem, Susan explains today, is that synthetic voice shock deters people who have lost their vision, and who would benefit greatly if they could adapt to those voices. Here’s how she characterizes typical reactions:

  • I cannot understand that voice!!!
  • The voice is so inhuman, inexpressive, robotic, unpleasant!
  • How could I possibly benefit from using anything that hard to listen to?
  • If that’s how the blind read, I am definitely not ready to take that step.

She adds:

Conversely, those long experienced with screen readers and reading appliances may be surprised at these adverse reactions to the text-to-speech technology they listen to many hours a day.

How can we help people cross that chasm? Susan offers advice to four groups: “vision losers”, developers of assistive reading technologies, sighted people who are helping vision losers, and rehab trainers.

I’m in the third group. My mom’s macular degeneration is progressing, and although she’s not yet forced to rely on text-to-speech, that day may come. To those of us in this group, Susan recommends that, when evaluating applications and appliances, we need to bear in mind that voice quality is a separable concern, not directly tied to the capabilities of the software and hardware. And she suggests that, in order to help friends or family members, we might want to develop some familiarity with the range of available voices.

To that end, Susan has provided audio renderings of her blog posts, including four different versions of today’s post as read by Neospeech Kate, Neospeech Paul, Microsoft Mike, and Robotic UK Jane. None of the readings is pleasant to listen to. But Susan says:

I testify it takes a little patience and self-training and then you hear past these voices and your brain naturally absorbs the underlying content.

Those of us not compelled to learn how to “hear past” those voices might still want to try the experiment, in order to help friends and family members make the transition.

A conversation with Joshua Tauberer about GovTrack

By an odd coincidence, my interview with Joshua Tauberer, founder and operator of GovTrack, went live just as the bill I’d been tracking using that system made headlines. Had I not been doing some investigation into legislative support for alternative biomass heating systems, I’d never have known that the Foreclosure Prevention Act of 2008 — which the President is now expected to sign into law this week, and which is generally known as the bailout for Fannie Mae and Freddie Mac — is also the bill that may result in tax credits for those of us in New England who’ve been investing in those heating systems.

This has been my most comprehensive attempt to open the lid of the legislative sausage factory, peer inside, and try to understand how a specific interest of mine was being processed. What I found is that, even with power tools like GovTrack and MAPLight, it’s really hard to make those connections. That’s partly because we lack good mechanisms to track the flow of bits of legislative language through an evolving assortment of bills, and to relate those fragments to the activities and interests of their sponsors.

But it’s also because a novice who tries to read and interpret this record lacks context. I had learned along the way, for example, that one of my senators cast a crucial NO on a cloture vote that would have amended the Energy Independence and Security Act of 2007 to include $22 billion of renewable energy tax incentives — including, as one small component, the “pellet stove” provision that has now landed in the Foreclosure Act. But I wasn’t at all clear on the significance, or recent evolution, of the parliamentary maneuver known as the cloture vote.

As Mike Caulfield points out here, there’s been a recent and dramatic increase in the use of this maneuver. As a result, “what your individual Senator does on bills matters quite a lot less than you think.”

The emerging breed of Congress-tracking tools and services can’t yet capture these nuances in ways that enable an ordinary citizen, motivated by some personal interest, to dive into the process and bootstrap an understanding that will guide voting or advocacy. Recognizing this, Joshua Tauberer explains in our interview that he’s currently exploring ways to crowdsource that bootstrapping effort. So for example, the site has a feature called “Users tracking this bill are also tracking…” Because I’m tracking both H.R. 6 (Energy Independence and Security) and H.R. 3221 (Foreclosure Prevention), the connection I made between these two seemingly unrelated bills creates a link for others to follow.

Mike writes:

As more people delve into these tools, the conversation around how our process really works can occur. And hopefully it can start to occur in channels outside the poliwonkablogosphere that I love so much but that others, well, not so much.

The first reaction to tools in the hands of novices is always that the novices don’t have the proper context. But use those tools and the context will come to you.

That statement resonates powerfully with two themes I’ve touched on before. One comes from John Willinsky, the educator, reading specialist, and open access advocate whom I encountered here, here, and here. There are only two factors that govern reading success, Willinsky says: motivation and context. To illustrate these principles, he talks about medical patients who, though they are nominally grade-8 readers, routinely surprise their doctors with grade-14 mastery of medical literature. How? When you or a loved one is sick, you’re intensely motivated to understand that literature. Yes, you initially lack context, but in the era of the Net, with the help of an ad-hoc community of others in the same situation, you can bootstrap that context.

The second resonance is with this Jeff Jonas meme: Data will find data, and relevance will find the user. I highlighted the social dimension of that idea in this essay: Data finds data, then people find people. Deepak Singh says these principles “define what the modern web is all about,” and adds that “without common data to commune around, there can be no communities.”

GovTrack is one of a number of efforts to create a framework of common data around which new political communities can form. I’ve also mentioned MAPLight, one of a number of projects funded by the Sunlight Foundation, whose chief data architect, Greg Elin, I interviewed here.

Admittedly these new services don’t yet close the gap that separates an ordinary citizen who’s motivated by a particular interest from a professional denizen of the poliwonkablogosphere. But it’s at least possible to leap across the gap, and we’re on the verge of building bridges across it.

Kudos for MAPLight’s visualization of Congressional activity

In yesterday’s essay on my efforts to relate what our New Hampshire senators say about renewable energy to what they do or don’t do about it (and why), I left out a crucial step. It’s a nice illustration of the subtle and powerful data visualization provided by MAPLight.org.

I’d found references to a crucial NO vote cast by John Sununu, but was having trouble finding out for which bill, and for which step in its process, that vote was recorded. I thought the bill was H.R.6-110, and I was looking for a NO vote, cast by Sununu just before its Dec 13 passage, where the final tally was 59-40.

MAPLight’s use of a Timeline proved decisive. In this view of activity around H.R.6-110, scoped to Sununu, the circled region records two Sununu votes around that time. One red for NO, the other green for YES.

Mousing over the red NO vote got me very close to the answer:

This wasn’t quite right. Although the annotation shown in the screenshot reads Dec 07, the mouseover popup (which I couldn’t capture) says Dec 13, which turns out to be the date I was looking for.

It seems that this rendering of the timeline can’t accommodate the closely-spaced votes revealed in this details view:

December 07, 2007 Senate Motion to Invoke Cloture on… Fail 53 42
December 13, 2007 Senate Motion to Concur in the… Pass 86 8
December 13, 2007 Senate Motion to Invoke Cloture on… Fail 59 40
December 18, 2007 House On Agreeing to Senate Adt… Pass 314 100

But when I got to that view, and saw the 59-40 tally that had been reported, I knew I’d identified the crucial cloture vote.

Nicely done MAPLight!

While we’re on the subject, by the way, I wanted to pin down exactly what language was removed from HR 6 by the failure of that crucial vote. From a Congressional Research Service report entitled Energy Independence and Security Act of 2007: A Summary of Major Provisions, by Fred Sissine et al., I found this clue:

The White House threatened to veto the House-passed bill mainly because of the presence of provisions for an RPS [renewable portfolio standard] and for the repeal of oil and gas tax subsidies.

On December 7, 2007, a Senate cloture vote on the House-passed version of H.R. 6 — with provisions for an RPS and for the repeal of oil and gas subsidies — failed (52-43). After stripping out the RPS and modifying the tax package, a cloture vote on S.Amdt. 3841 failed (59-40).

To visualize what was subtracted from S.Amdt 3841 to produce S.Admt 3850, I dropped 3841 into Wikipedia, replaced it with 3580, and compared those two versions.

Here’s where the two part ways:

And here, deep in the thickets of 3841, is my beloved pellet stove provision in the left column, matched only by emptiness in the right column:

Now this surely isn’t an appropriate use of Wikipedia, so I’ve withdrawn the page I made to compare these two versions. But it’s odd, don’t you think? Why can we readily analyze the differences between two versions of a fanciful article about the heavy metal umlaut, but not so readily analyze the differences between two versions of a major piece of legislation?

Net-enhanced democracy: Amazing progress, solvable challenges

Last week I spoke with Joshua Tauberer for an upcoming interview about his project, GovTrack.us, which is one of several efforts to make official sources of information about Congressional legislation available in more useful and actionable ways.

On Sunday, my local newspaper ran an article that sent me scrambling to GovTrack, to OpenCongress, to the Washington Post’s Congressional voting database, to the Library of Congress’ THOMAS, and to a flock of other sites. Why? The newspaper article, illustrated by a photo of NH Senator John Sununu standing with the owners of New England Wood Pellet in front of a pile of sawdust that will be pelletized, said:

Sununu supported a tax credit on high-efficiency biomass fuel systems, including wood-pellet stoves. The Senate passed the measure in April, in effect encouraging people to use the appliances, just as those who take on solar or wind power receive financial incentives for doing so.

I’ve invested rather heavily in clean wood-burning technologies — first in a pellet stove, now also in a wood gasification boiler — and I’m hoping these investments will qualify for tax credits. When phone calls to the local offices of my state representatives produced no results, I resolved to research the matter myself.

Here were my questions, along with the answers I found.

Q: Which bill approved by the Senate in April contains these incentives?

A: Bizarrely but typically, it’s not an energy bill. Instead it’s a housing bill, the Foreclosure Prevention Act of 2008, otherwise known as:

Building American Homeownership Act of 2008
Clean Energy Tax Stimulus Act of 2008
FHA Manufactured Housing Loan Modernization Act of 2008
FHA Modernization Act of 2008
Mortgage Disclosure Improvement Act of 2008
REIT Investment Diversification and Empowerment Act of 2008

Q: Which appliances will qualify?

A: The bill says:

(F) a stove which uses the burning of biomass fuel to heat a dwelling unit located in the United States and used as a residence by the taxpayer, or to heat water for use in such a dwelling unit, and which has a thermal efficiency rating of at least 75 percent.

On that basis I believe that both of my appliances should qualify.

Q: What is the status of that legislation?

A: According to GovTrack:

Status:
Introduced Jul 30, 2007
Scheduled for Debate Sep 5, 2007
Amendments (235 proposed) [details]
Passed House [details] Aug 4, 2007
Passed Senate Apr 10, 2008
Differences Resolved
Signed by President
The bill may now proceed to
a conference committee of senators and representatives to work out
differences in the versions of the bill each chamber approved.
The bill then goes to the President before becoming law.
[Last Updated: Jul 20, 2008]

Cool! That’s what I needed to know. It ain’t law yet, but it might be. And thanks to a bevy of Web 2.0 tools and services, it was easy to find that out, right?

Wrong. It took me three hours on Sunday morning to sort this out. Along the way, I uncovered information that challenges the premise of the newspaper story. Entitled “Sununu touts alternative energy”, the story supports NH Senator John Sununu’s own claim to be a champion of long-term energy policy.

Here’s part of the analysis I posted in a comment to that newspaper article:

Sununu (and Gregg) were two of the votes responsible for the failure, by one vote, to include a $22 billion energy tax package in the Energy Independence and Security Act of 2007. Weirdly, that rejected tax package included Sununu’s original pellet stove provision from S1697!

So while the act became Public Law 110-140 in December, $22 billion of renewable energy incentives were jettisoned from it by Sununu. Why? The Congressional Quarterly noted that it “would have been offset in part by stripping about $13 billion in tax breaks from the oil and gas industry,” suggesting that Sununu was influenced by those interests.

Subsequently, I remembered Maplight.org where I found the breakdown for that crucial NO vote:

It’s utterly amazing, and deeply hopeful, that an ordinary citizen with no prior experience with an issue can dive in and do this kind of analysis.

And yet…I’m not really an ordinary citizen. It’s true that I’ve never before attempted this kind of detailed analysis. But I have an unusual degree of skill and perseverence when it comes to searching the Net, finding and using a variety of services, and correlating information from a variety of sources. At one point, trying to pull it all together, I had two dozen browser tabs open. That’s not normal.

It’ll take a few more turns of the crank before this kind of thing becomes routine enough for a truly ordinary citizen who cares about some issue to do the same kind of analysis.

On the first turn of the crank, we’ll need to get to the point where a Net-savvy individual can do what I did in minutes not hours.

On the second turn of the crank, we’ll need to deliver that capability into the hands of folks who aren’t freakishly Net-savvy.

I really believe that’s possible, maybe even inevitable. But challenges remain.

One of the biggest challenges, as Joshua Tauberer says in our upcoming interview and as my experience here confirms, is the analysis of the legislative texts themselves. Even with all of the excellent online services now available, it’s incredibly hard to relate colloquial discussion of legislation to the actual process and to related commentary. Every day, we read statements like this:

The Senate passed the measure in April, in effect encouraging people to use the appliances, just as those who take on solar or wind power receive financial incentives for doing so.

It took me hours to unravel that statement. For starters, it would help if newspapers would cite the legislation they mention. But that’s only a start. In this case, as is typical, the Sununu pellet stove provision began its life in one Senate energy bill that died, migrated into another Senate energy bill that died, and eventually wound up as part of a housing bill that passed. Separately, the provision was included in, then struck from, yet another energy bill that passed.

Crowdsourcing will be part of the answer. But to empower the crowds, we’re going to need power tools that help us visualize how specific bits of legislative language flow through the bewildering thicket of bills, to tie those bits of language to their sponsors, and to relate them to journalistic/blogospheric commentary.

Happily, this is just the sort of problem that should appeal to software engineers. The management of source code, like the management of legislation, is a high-stakes game. And in the software realm, we play that game really well. We take for granted the ability to reliably track changes in a large and evolving corpus of interrelated texts. When that same expertise is applied in the legislative realm, the results could be dramatic.

On a related note, I enjoyed this snippet from a Wired interview with Martin Wattenberg, whom I also interviewed here.

Wired:
Why is a numbers guy like you so interested in large textual data sets?

Martin Wattenberg:
Language is one of the best data-compression mechanisms we have. The information contained in literature, or even email, encodes our identity as human beings. The entire literary canon may be smaller than what comes out of particle accelerators or models of the human brain, but the meaning coded into words can’t be measured in bytes. It’s deeply compressed. Twelve words from Voltaire can hold a lifetime of experience.

Similarly, snippets of legislation can be the tickets to a more participatory democracy — if we can unravel and expose the contexts surrounding those snippets. It should be doable.

John Faughnan’s amazing Outlook hack (and why it matters)

Although I’ve conversed online with John Faughnan since my days at BYTE, we’ve never met, and we had not even spoken on the phone until last week when he joined me on an episode of my Interviews with Innovators podcast. It was a great pleasure to finally connect in realtime with the prolific author of thoughtful analysis and commentary on things in general, on information technology, and on resources for parents of children with cognitive or emotional-behavior disabilities.

John was a country doctor, and he retains his medical license, but he doesn’t see patients nowadays. Instead he directs the development of clinical productivity software, with particular focus on methods of knowledge representation, and on strategies for effective collaboration.

We share a passion for strategies that entail simple but often overlooked uses of common software applications. For example, did you know that it’s possible, in Outlook, to edit the subject of an email message after it’s been received, and is just sitting in your archive? Try it, and you’ll find that you can. Color me amazed. I’m just the sort of personal information management geek who’d have discovered a hack like that, but I never did.

Now, why would you want to do such a thing? It’s a defensive strategy. The message entitled “Re: Next week” probably ought to be entitled something like “Consensus reached among A, B, and C on issue X for project Y.” Which title would you rather scan, in search results, six months later?

(John would like to find, and personally thank, the developer responsible for this feature, so if you know that person, or are that person, speak up!)

You can think of this technique as a kind of enhanced tagging. It’s related to a strategy for enriching email — embodying the journalistic principle of “heads, decks, and leads” — which I described in my book and in this report.

People mainly still think of information architecture as a discipline practiced only by designers and publishers. But what John and I have always thought is that we’re all becoming designers and publishers of streams of information, that those streams can all be navigated and searched in one way or another, and that the value of those streams depends on the ability of ourselves and others to navigate and search them effectively.

We also think that effectiveness requires two things. First, obviously, software that embodies the right principles and enables the right practices. But second, a broad awareness of right principles and practices. Those, we agree in this conversation, are not necessarily intuited by Gen X, Y, or Z just because they’re so-called digital natives. This stuff needs to be articulated, and it needs to be taught.

How to wire up a timer-triggered WPF event handler in IronPython

In the last installment of my little series on turning Internet feeds into TV feeds, I had decided to use IronPython to fetch data from the Internet, but C# to drive the WPF (Windows Presentation Foundation) application whose display my local public access TV station will broadcast. This division of labor between C# and IronPython arose because the XAML that drives the display needs to be refreshed periodically, and I didn’t know how, in IronPython, to properly delegate a timer-based event handler for WPF.

In this comment, Michael Foord, author of IronPython in Action and a major contributor to the IronPython Cookbook, showed me the way. Thanks Michael!

Based on his example, I’ve rewritten the C# program shown here as the IronPython script shown below.

I haven’t yet decided which version to deploy, but I’m leaning toward the IronPython version. Not because it’s more concise. It isn’t, really. Nor because I feel any need to use the same language for both components of the solution — that is, the feed fetcher and the feed displayer. I don’t care about language uniformity for its own sake.

I am, however, thinking that the folks at the TV station may want to modify these programs themselves. They’re pretty simple, and there’s no reason they shouldn’t be able to tinker with them. From that perspective, code that can be modified with nothing more than a text editor will be more accessible than code which requires a compiler.

I’m reminded of my early days as a website operator, when I was always glad to discover that a third-party application was written in Perl, rather than in C. That meant I could, and sometimes did, tweak the application in ways that otherwise would have been difficult or even (lacking C source code) impossible.

The difference here, of course, is that all of the underlying machinery — XAML, WPF, and the entire .NET Framework — is exactly the same1 when approached from a scripting language like IronPython or a compiled language like C#. This ability to use common infrastructure from different langages — and from very different kinds of languages — has always seemed like a big deal to me, and still does.


1 The same, that is, modulo the kind of boundary-crossing issue that stumped me until Michael Foord pointed me to CallTarget0, the wrapper for creating a delegate in IronPython.


import clr

clr.AddReferenceByPartialName("PresentationCore")
clr.AddReferenceByPartialName("PresentationFramework")
clr.AddReferenceByPartialName("WindowsBase")
clr.AddReferenceByPartialName("IronPython")

from System import *
from System.Windows import *
from System.Windows.Markup import *
from System.Windows.Media import *
from System.Windows.Input import *
from System.Windows.Threading import *

from IronPython.Runtime.Calls import CallTarget0

def LoadXaml(filename):
  from System.IO import *
  from System.Windows.Markup import XamlReader
  f = FileStream(filename, FileMode.Open)
  try:
    element = XamlReader.Load(f)
  finally:
	f.Close()
  return element

class Scroller(Application):

  def tickhandler(self,sender,args):
    def update_xaml():
      self.window.Content = LoadXaml(self.xaml)
    self.timer.Dispatcher.Invoke(DispatcherPriority.Normal, 
      CallTarget0(update_xaml))

  def __init__(self):
    Application.__init__(self)
    self.xaml = "scroller.xaml"
    self.window = Window()
    self.window.Content = LoadXaml(self.xaml)
    self.window.WindowStyle = WindowStyle.None       # go fullscreen
    self.window.WindowState = WindowState.Maximized  # 
    self.window.Topmost = True                       #
    self.window.Cursor = Cursors.None                #
    self.window.Background = Brushes.Black           #
    self.window.Foreground = Brushes.White           #
    self.window.Show()
    self.timer = DispatcherTimer()
    self.timer.Interval = TimeSpan(0, 60, 0)         # refresh hourly
    self.timer.Tick += self.tickhandler              #
    self.timer.Start()                               #

Scroller().Run()

Dan Bricklin on becoming a Happy Caster

The Conversations Network is embarking on a new phase in which it will expand its ambition to capture, publish, and curate spoken-word audio from a wide range of sources. One of the challenges will be to help more people effectively capture audio to a reasonable standard of quality. Dan Bricklin, my guest for this week’s ITConversations show, has ascended that learning curve in recent years. In this conversation he explains why he’s become interested in audio recording, and what he has learned about equipment, and techniques, which can be readily transferred to individuals and organizations wanting to make decent recordings of their own events.

When I embarked on my personal audio adventure a few years ago, I naively thought that our fancy new digital technologies would make the whole process very simple. Boy, was I wrong about that. Yes, we’ve made digital photography accessible to the masses, but there was vast demand for enabling the so-called Happy Snapper to point, shoot, and take a decent photo. There’s been comparatively little demand for enabling the Happy Caster to plunk down a microphone, punch record, and capture a decent sound track.

Over the last few years I’ve slowly and painfully assimilated just a fraction of the audio lore possessed by domain experts like the Conversations Network’s founder Doug Kaye, and its senior audio engineer Paul Figgiani. So it was refreshing to hear from Dan Bricklin that it has also been a struggle for him to become competent in this domain.

I guess the demand for point-and-shoot photography will always outstrip, by orders of magnitude, the demand for plunk-and-punch audio recording. But the latter demand is growing, and in this conversation we speculate a bit on what the Happy Caster solution might be.

Mainly, though, Dan focuses on two things. First, the new opportunity to capture spoken-word events that would otherwise be lost, and publish them for audiences that didn’t attend, or couldn’t have attended, in person.

Second, the minimal setup that will enable folks who are not audio experts to accomplish that capture and publication.


PS: A bit of backstory on this recording illustrates some of the challenges of the audio domain. In my FAQ for interviewees, I invite remote interviewees to record themselves locally, then send me the track which I combine with my own locally-recorded track. Why? If you’re sending voice over the network, whether it’s POTS (plain old telephone service) or Skype, there’s a lot that can and often does go wrong. Eliminate the network and you avoid all those problems.

In principle, combining local tracks recorded separately is a great solution. In practice, it has almost never worked out, and this case was no exception.

Usually the problem is that interviewees lack the gear or knowledge required to make a decent local track. Attempts to record directly into a computer always end badly. Most people don’t own standalone digital audio recorders. In one case, a musician who routinely records his music through a mixer nevertheless produced an unusable track because he’s not used to recording his voice and overshot the limits.

In this case, Dan was quite capable of making a good recording, and he did, but things went wrong on my end. What Dan recorded was an MP3 file. What I was expecting was a WAV file, because I was going to edit the combined recording and it’s dicey to uncompress an MP3, edit, and then recompress.

Now, Dan had recorded the MP3 at a bit rate — 192kbps — that he judged would be high enough to survive an edit. But would our discriminating audio engineer Paul Figgiani agree? We weren’t sure, so I sent Paul samples of Dan’s MP3 track and the WAV file I made from the telephone track I’d recorded using the Telos. Paul’s verdict: “I think we can make the 192 kbps mp3 version work. The bit rate is high enough … lets go with it.”

So far, so good. But when I loaded up my local track and Dan’s remote track into Audition, things didn’t line up. My WAV track was slightly longer (or shorter, I can’t remember) than Dan’s MP3 track. The difference was only about 1.5 seconds over an hour-long recording, but still, it had to be dealt with.

Audition has a time-stretch feature that can be used to solve this problem. And I could swear that I’ve used it successfully before in these circumstances. But this time, I couldn’t make it work. Every time I tried to stretch the shorter clip, it snapped back to its original position. I fiddled with every approach I could think of, or could discover by searching, and finally threw up my hands and just used the original recording that had both halves of the conversation in sync. If this Audition behaviour rings a bell with anyone, I’d love to know what went wrong and how to avoid it next time.

The moral, anyway, is that if a reasonably technical guy like me is struggling to keep his head above water in this domain, it’s clear that non-geeky civilians will just drown. I’m quite curious to know when, or perhaps whether, those civilians will constitute a market that technology providers want to serve.

Homophily, anti-recommendation, and Driveway Moments

The folks at National Public Radio love to create driveway moments:

You’re driving along, listening to a story on NPR. Suddenly, you find yourself at your destination, so riveted to a piece that you sit in your idling car to hear it all the way through. That’s a Driveway Moment.

The podcasting counterpart, for me, is the Ashuelot Moment. I’m jogging along the Ashuelot River, and I’m so riveted to a piece that I take a longer route so my run won’t end before the story does.

The Long Now podcasts are my most reliable source of Moments but they’re only on a monthly cycle. TED talks are another good source, though I’ve lost track of how to subscribe to the comprehensive audio-only feed. The Conversations Network, to which I contribute a weekly show, produces occasional Moments, but a lot of the material there is so closely aligned with my own particular interests and inclinations that it doesn’t often surprise or challenge me.

Another good source is Christopher Lydon’s Open Source, which launched in 2005, suffered a setback in 2006, and then recovered in 2007. It took me a while to reconnect after the hiatus, but now I’m finding it to be more stimulating than ever.

Here’s my most recent Moment, from this Open Source show with Ethan Zuckerman and Solana Larsen. Ethan is speaking:

My hope was that with the Internet, suddenly we’re all connected, we hold hands and sing Kumbaya. And it just hasn’t worked out that way.


You loook at a site like Digg, or Reddit, these are sites that promised the future of journalism. We’d all get together and decide what’s important. But, who’s we? Or as per the Lone Ranger, who’s we, white man? Or more to the point, who’s me, white geek?

If you’re getting your news from these sites, you’re getting a very particular, tech-heavy view of politics, a fairly focused view of the world. And you start falling victim to homophily, which is what happens when all of your news and opinions are coming from people who’ve got the same background and the same values as you.

Homophily is the tendency of birds of a feather to flock together. It’s the tendency to walk into a room, find the person most similar to you, and form a bond. It’s a natural human tendency, but it’s probably worth fighting against. Homophily makes you stupid.

Of course I share tribal affiliations with Ethan Zuckerman, so I’d have been likely to find that particular show one way or another. But Global Voices Online, the project that Ethan and Solana discuss on that show, is all about resisting homophily, and enabling us to tune into global perspectives offered by people in circumstances very different from our own.

Just because we can, though, doesn’t mean we will. Homophily is a natural tendency. It’s easy and comfortable to immerse ourselves in the familiar. It’s hard and uncomfortable to seek out the unfamiliar. How do we overcome that?

Recommendation systems don’t help me much. They only suggest things similar to other things I’ve shown interest in. Increasingly that just frustrates me. The most delightful recommendations are those that connect me with things that interest me in unpredictable ways. That happens serendipitously, and I haven’t yet found a reliable way to manufacture the serendipity.

Lately I’ve started to wonder about the notion of anti-recommendation systems. One example of an anti-recommendation system is LibraryThing’s UnSuggester, which find books least likely to coincide with yours. It’s a whimsical feature that honestly hasn’t been useful to me yet, but I think the idea merits exploration and development.

Although it isn’t automated or automatable, I’d argue that the Passion Thursday series on Open Source is a kind of anti-recommender. The series includes shows about birdwatching, the pursuit of truth, poker, the potato, cursive handwriting, and the theremin, an early electronic instrument recently notable in the repertoire of the indie band DeVotchKa. The only common thread is someone’s passionate interest in something.

We’re not inclined to resist homophily and seek out otherness. But passionate storytellers can take us to places we wouldn’t otherwise go, and create Moments there.

Passion is a good way to lubricate the engine of serendipity.

Will people understand and embrace the right identity systems? Maybe yes!

In conversation with English and Welsh friends last week, the subject of Britain’s imminent National Identity Scheme came up. My friends, who are worldly and well-educated but not technical, voiced concerns about the amount of personal information that will be stored. Their understanding was that a lot of this information will be kept on the new ID card. In fact, the proposal says that only a subset will stored on the card, which will be backed by a cloud-based (and decentralized) National Identity Register. But either way, my friends’ concerns are of course valid. If governments or businesses aggregate too much personal information, accidents and abuses will occur.

At the same time, my friends do recognize the need for a strong and secure means of identification. So they’re not opposed to identity cards on principle, they just don’t want those cards to contain, or link to, extensive dossiers.

At this point, channeling Kim Cameron, I launched into an explanation of the laws of identity and the identity metasystem. Well, sort of. I didn’t say anything about cryptography, or digital certificates, or XML web services. But I did paint a picture of a world in which individuals interact with many identity providers and many relying parties, in which all actors trust one another in exactly the ways they already do today, and in which disclosure of personal information is minimal and context-dependent.

Halfway through I thought, well, this will never fly. This whole scheme is based on decentralization and indirection, and I know people don’t take naturally to those concepts.

But…they completely got it! Maybe that’s because the threat of a monolithic system leads people to appreciate the virtues of a decentralized one. Maybe it’s because ongoing experience with the Net makes people more comfortable with the principle of indirection. Maybe it’s both these factors and others as well. In any event, it was a hopeful moment. Identity geeks have struggled, for many years, not only to devise right systems, but also to motivate an understanding of what makes systems right, and why. Now that right systems are coming into existence, it’s good to see that (some) people are ready to appreciate and embrace them.

How the WorldWide Telescope works

On my Perspectives show last week, Curtis Wong and Roy Gould relate the history and educational mission of the WorldWide Telescope. On this week’s show, principal developer Jonathan Fay describes how the underlying technologies enable the WWT’s seamless view of the sky.

There were a bunch of things I wanted to know, including:

How does the WWT project build on, and extend, the SkyServer project to which Jim Gray made fundamental contributions?

What standards and protocols enable the various sky surveys to be woven together?

What’s the relationship between Deep Zoom and the WWT’s own scheme for managing and viewing tiled multi-resolution imagery?

How much of the data is stored on Microsoft servers, how much is stored elsewhere, and in what ways do the supporting data services cooperate?

Jonathan answers all these questions, and he also answers one I didn’t think to ask:

What technique is used to project the stars onto an imaginary sphere at near-infinite distance?

The answer to that last question is that a new kind of spherical projection had to be invented:

Imagine taking a round room, and trying to put a bunch of bathroom tiles on it, and grout it. The tiles seem to come together and have parallel lines for a while, but eventually it stops working well. Maybe you can take one line around the equator, but as you go up you have fewer tiles, and weird-shaped tiles, and nothing lines up.

That’s the problem we have. We’re looking at spherical data, so we had to come up with a new spherical transform that preserves the poles. In previous projects, like Virtual Earth or TerraServer or Google Earth, the poles weren’t important, because nobody lives there and nobody needs map directions for driving around there.

So we had to come up with something called TOAST: tesselated octahedral adaptive subdivision transform. It creates a 360-degree wraparound view that’s either a planet surface or the infinite sphere of the sky, and lets you represent it using a 3D graphics accelerator, very rapidly and efficiently. So we can have an image pyramid the way Deep Zoom does, and TerraServer before it, but we don’t have to give up the poles.

This transform isn’t proprietary, and in fact it’s being applied to the 50-odd full-sky surveys hosted at NASA’s SkyView virtual observatory. The implications are pretty astounding. This imagery is stored in astronomical databases using what’s called tangential projection, which suffers from polar distortion when combined into large mosaics. Now the imagery can be combined into large mosaics — or indeed a complete view of the sky — and seen without distortion. What’s more, multiple surveys can be aligned to that spherical projection. That’s why, in WorldWide Telescope, you can cross-fade between a view of the Milky Way in visible light and views in infrared or ultraviolet light.

What the WorldWide Telescope really is, Jonathan says, is a browser, like a web browser but for an information space defined in astronomical terms. Here’s how he sums up the work that was necessary to make that possible:

The vision of getting everybody access to all this astronomy data required systematic changes at every single level. We built on some things that Jim pioneered with NVO, and worked from there, but it was very systematic. How people process the data. The client to access the data. The protocols over the wire. Educating people, providing the context for it.

We put a lot of things together, but we also created a systematic model for how to do everything end to end, top to bottom, left to right. Now there may be other people who use the pieces that we’ve created, and then change them to use different data sources, different visualizations. Say someone creates a Mac client, or an iPhone client, that’s possible. Or a mobile phone version of it, or a web-based version. Over time we or others can replace various components, but as a reference model for solving all the problems in order to get the data into people’s homes and into their eyeballs — you had to solve for all of those problems, otherwise people are still blocked from being able to really explore.

For Curtis Wong, the WWT is an extension of John Dobson’s sidewalk astromony — a way to bring telescopes to the public and to enable astronomers to share their knowledge of the sky with everybody. For Jonathan Fay, it’s the perfect application of earth and sky visualization technologies he’s been developing throughout his career. Their interests and talents combined, as Jonathan says, like peanut butter and chocolate:

Curtis had been exploring how to create an educational environment with rich tools for exploring space, and he’d been collaborating with Jim Gray on TerraServer, and now he was looking for the technology to make it possible.

Here I had this technology, and was looking for somebody who was enthusiastic about having a purpose for it. So it was the peanut butter and chocolate moment.

Yum.

More ways to turn Internet feeds into TV feeds

Last week I started looking into ways to Internet feeds into TV feeds. Although I did come up with a way to turn a data feed into a video file, that wound up being overkill. It turns out that the local station is willing to broadcast the signal from a computer display. To create that signal, several folks suggested using PowerPoint, but I found that its scrolling credits feature doesn’t accommodate really long lists of credits. So I decided to try XAML, the application markup language that works with Silverlight and the Windows Presentation Foundation (WPF), in concert with IronPython.

The plan was as follows. A long-running IronPython script periodically fetches the feed from a web service, interpolates the text into a XAML template that animates the crawl, and displays the XAML in a fullscreen white-on-black WPF window.

Here’s the XAML template:

<Grid xmlns="http://schemas.microsoft.com/winfx/2006/xaml/presentation"
  xmlns:x="http://schemas.microsoft.com/winfx/2006/xaml"  
  Width="%s" Height="%s">
<TextBlock xml:space="preserve" FontSize="%s" Margin="%s,%s,0,0" 
  FontFamily="Arial">
    <TextBlock.RenderTransform>
      <TranslateTransform x:Name="translate" />
    </TextBlock.RenderTransform>
  <TextBlock.Triggers>
    <EventTrigger RoutedEvent="FrameworkElement.Loaded">
      <BeginStoryboard>
        <Storyboard RepeatBehavior="Forever">
          <DoubleAnimation From="%s" To="-%s" 
            Storyboard.TargetName="translate" 
            Storyboard.TargetProperty="Y"
			Duration="00:%s:%s" />
        </Storyboard>
      </BeginStoryboard>
    </EventTrigger>
   </TextBlock.Triggers>
<Run>
<![CDATA[
%s
]]>
</Run>
</TextBlock>
</Grid>

Some of the values depend on the number of items in the feed, so the script interpolates those values into the template. Then it formats the feed and plugs the formatted text into the template’s CDATA section. The formatted text looks like this:

EVENTS FOR MON JUN 30 2008 FROM THE ELMCITY.INFO CALENDAR

06:00 AM:  lap swim (ymca)

07:00 AM:  AA: On Awakening Group (eventful: Keene Unitarian
      Universalist Church)

After generating the XAML, the IronPython script fires up an Application object, creates a window, loads in the XAML to start the crawl, and sets a timer to refresh the XAML.

I ran into a snag when I tried to set that timer, though. There are a few different timers you might imagine using in this context, including Python’s own timer object and various timers available in the .NET Framework. All but one of these, however, will complain about invalid cross-thread access when you try to update the application’s user interface from a timer event handler.

The right timer to use, it turns out, is .NET’s System.Windows.Threading.DispatcherTimer. But when I tried it, I ran into another snag. In C#, you create a WPF-friendly timer like so:

DispatcherTimer timer = new DispatcherTimer();
timer.Tick += new EventHandler(event_handler);

event_handler is a method, but EventHandler returns a delegate that encapsulates that method. I couldn’t find a straightforward way to create a delegate, and do that encapsulation, in IronPython.

If you know how, I’d love to hear about it. Then again, it really doesn’t matter. Logically this program has two loosely-coupled parts. The engine part reads the feed from a web service and formats it as XAML. It can be a Python script that runs on a scheduled basis to fetch and format the feed.

The user interface part loads, displays, and then periodically refreshes the XAML. It can be a little C# program that runs forever, displays the animation, and refreshes the data, like so:

using System;
using System.Windows;
using System.Windows.Markup;
using System.Windows.Input;
using System.Windows.Media;
using System.IO;
using System.Windows.Threading;

namespace CalendarCrawl
{
  public class CalendarCrawler
  {
    static private Application app = new Application();

    static private StreamReader getXaml()
    {
    StreamReader sr = new StreamReader("WPF.xaml");
    return sr;
    }

    [STAThread]
    static public void Main(string[] args)
    {
        Window win = new Window();
        win.WindowStyle = WindowStyle.None;           // go fullscreen
        win.WindowState = WindowState.Maximized;      // go fullscreen
        win.Topmost = true;                           // go fullscreen
        win.Cursor = Cursors.None;                    // go fullscreen
        win.Content = XamlReader.Load(getXaml().BaseStream); 
        win.Background = Brushes.Black;
        win.Foreground = Brushes.White;
        DispatcherTimer timer = new DispatcherTimer();
        timer.Interval = new TimeSpan(0, 1, 0);       // every hour
        timer.Tick += new EventHandler(eventHandler); // wire up handler
        timer.Start();
        app.Run(win);
    }

    static private void eventHandler(Object sender, EventArgs args) 
    {
        app.Windows[0].Content = XamlReader.Load((getXaml().BaseStream));
    }
  
  }
}

It was odd how reluctantly I came to this division of labor. Evidently I still need to remind myself that in a world of loosely-coupled applications and services, when you need to get something done, There Is More Than One Way To Do It.

Here’s another way. If the engine doesn’t have to talk to the .NET Framework’s WPF machinery, there’s no need to use IronPython. Any flavor of Python makes a handy tool for talking to RESTful web services, wrangling text, and interacting with the file system.

Here’s yet another way: A Silverlight version of the user interface. It’s nice to know that option is available. However, I’m leaning toward the C# version. The target machine is Vista, it already has .NET and WPF, why use a long-running browser instance just to host this tiny little thing?

One final point is worth mentioning. XAML is really just another source language for the .NET runtime and framework, like C# and IronPython and others. You can, for example, create an application window by writing a Window tag in XAML markup, and specifying parameters as attributes. Or you can do it by invoking System.Windows.Window from IronPython or C# or another .NET language, and specifying parameters in code. The boundary between markup and code is very fluid, and you can draw the line for reasons of convenience, maintainability, and taste. It’s a very flexible system, and it becomes even more flexible when you can use a dynamic language like Python to generate the XAML, the code, or both.

From seeing to hearing: A conversation with Susan Gerhart about assistive technologies for the sight-impaired

For many of us, the podcasting revolution has opened up the audio channel as a new option for receiving information that we might otherwise read. But for the sight-impaired, like Susan Gerhart, who joins me for this week’s ITConversations show, the audio channel isn’t optional. Her myopic retinal degeneration has forced her to shift almost entirely into audio mode in order to read, and to work on the computer.

As a lifelong technologist, Susan is a capable user and evaluator of software and computational devices. When she entered the world of assistive technologies — including the NVDA screen reader, the LevelStar Icon, the Kurzweil NFB Reader — she decided to share her experiences on a blog. In our interview she summarizes what she’s learned so far about using these technologies to adapt to her changing vision.

From PowerPoint to IronPython/XAML

As per the comments on yesterday’s item about creating a video crawl for local TV, it turns out there’s no need to produce a video file. Instead it’ll be OK to use a computer display directly. The computer could be running, for example, a PowerPoint slideshow in a loop.

Here’s the apparently standard recommendation for making scrolling credits in PowerPoint. It was written for earlier versions, but seems applicable also to the current 2007 version:

Create movie-style crawling credits in PowerPoint presentations

  1. In a PowerPoint presentation, create a new slide for credits or any other list that you want to scroll from bottom to top.

  2. Type your credits or other text. Don’t worry about text running off the bottom of the slide. In fact, it should run off the bottom if you are going to have enough text to make a crawl effect work well.

  3. Right-click the text, and on the shortcut menu, click Custom Animation.

  4. Select the text that you want to scroll. In the Custom Animation task pane, click Add Effect. Point to Entrance, and click Credits.

  5. Click Play to see how the effect will look on-screen.

  6. Move the text block completely off the top of the slide. When you play your presentation, the text will crawl or scroll from the bottom of the screen and disappear off the top.

But it doesn’t look like you can get more than three screenfuls of data into the crawl. For example, I made a textbox with 200 lines of text numbered accordingly. Then I animated it using several varations on this technique.

First I put the top of the textbox at the top of the slide, like so:

The effect: Line 0 crawls into view from the bottom of an empty slide, and the crawl ends with line 25 at the top and line 50 at the bottom.

Next I put the top of the textbox at line 25, like so:

The effect: Line 0 appears at the top of the slide, the crawl ends with line 50 at the top and line 75 and the bottom.

Is there a way to include more than three screenfuls of data in the crawl? If not, it looks like it’d be necessary to create a series of slides, each with two screenfuls of data. The first slide would need to have its first line of data at its top. But the second and following slides would need to have their middle lines of data at their tops. Gnarly.

I’m sure that could be done, but why bother? Absent a requirement to produce a video file, there a zillion ways to make text crawl up a computer screen. This might be a good opportunity to explore the combination of IronPython and XAML.

Turning Internet feeds into TV feeds

I’ve cobbled together a way to turn an Internet data feed into a video crawl that can run on my local public access cable TV channel. Before explaining how, I need to explain why. Here’s the short answer: As much as I want everyone to use the Internet for all it’s worth, most people don’t yet.

A couple of years ago, I was campaigning in my community to open up the parent portal into PowerSchool, a student information system that was being used internally by teachers and administrators but wasn’t available to parents via the Internet. At one point I made a screencast that addressed the perceived risks, and showed the compelling benefits, of opening up the portal. The screencast was published on the Internet, available to the whole world, and the whole world includes Keene NH, so that ought to be a good way to bring my message to the community, right?

Wrong. Nobody watched it.

A while later, it hit me. There still aren’t many folks here who are inclined to receive a message like that from InfoWorld.com, or from YouTube, or from any other Internet destination I might use. But there are significant numbers who tune into the local public access station. Why not show the screencast there?

So I dubbed it onto a MiniDV tape, took it down to the station, and gave it to the executive producer.

Him: What’s this?

Me: A demo and discussion of the PowerSchool software. Will you run it?

Him: Sure.

And lo, a couple of weeks later, I heard from the assistant superintendant of schools. He thanked me for applying the external pressure that they’d been needing in order to break through an internal logjam, and he invited me into the beta program. Now, two years later, it’s fully deployed and making a big difference.

Meanwhile, I’ve been working on a community information project that’s all about feeds and syndication. But slow learner that I am, I continue to invite people to use Internet feeds and Internet syndication. And people continue to mostly decline the invitation.

For example, I’ve been working on calendar syndication. The syndication flows two ways. First, inward. The service pulls events from various local websites, and I’m working with the proprietors of those sites to clarify why and and to publish true calendar feeds.

Second, it syndicates outward. With a JavaScript call, you can include the day’s events in another website, like CitizenKeene or Cheshire TV.

But this is all still just Internet stuff. And as we’ve seen, the community doesn’t (yet) tune into the Internet for local information. It does tune into public access cable TV.

So why can’t Internet data feeds show up there?

Well, of course, they can. Here’s a prototype video crawl (the link goes to an animated gif, just for convenience) made from yesterday’s combined calendar. We’ll need to work out the details of format and workflow, but I think it’ll work. And it seems like a great way to connect two worlds.

Calendars are just part of the story. Consider, for example, the public library’s RSS feeds announcing new books and DVDs. I’m one of probably a handful of subscribers to those feeds. Now imagine that the feeds showed up as a video crawl on TV. I bet a lot more folks would find out about new books and DVDs. And maybe, just maybe, the reception of that feed via TV would lead to discovery and use of the more convenient and powerful Internet feed.

We’ll see. Meanwhile, below the fold, I describe the method I’ve come up with to do this. The paint isn’t dry, and I’ll be very interested in comments and suggestions.

… the fold …

Our public access TV station, as may be typical (though I dunno), is a mostly Windows-based operation. As is surely typical, there’s little money to spend, either on people to produce these feeds interactively or on software to produce them automatically. So the requirements seem to be:

  1. Windows-based
  2. Cheap or free
  3. Fully automatable

My first idea was to leverage SMIL. I knew it would be easy and free to transform a feed into markup that can be played by Real, or QuickTime, or Windows Media. And I hoped it would also be easy and free to render that markup into a video format. But there I ran aground. If there’s a free, or at least cheap, SMIL renderer that can be scheduled to run automatically, I’d like to know about it, because that’d probably be the ideal solution. But I haven’t found one.

The next idea was to produce the animation frame by frame. And that’s what I’m actually doing for now. It sounded a lot harder than it turned out to be. After installing the Python Imaging Library, it was possible to write this very concise frame generator:

import Image, ImageFont, ImageDraw

s = """
EVENTS FOR WEDS JUNE 30 FROM ELMCITY.INFO (HTTP://ELMCITY.INFO/EVENTS)

06:00 AM lap swim  (ymca) 
07:00 AM Cheshire Walkers: Indoor Walking Program (eventful: Keene Recreation Center) 
...
Trainers Academy - Level II (eventful: Monadnock Humane Society) 
TOR 7pm (swamp bats) 
"""

lines = s.split('\n')

def frame(index,top):
  image = Image.new('RGB',(720,480),(0,0,0))
  draw = ImageDraw.Draw(image)
  font = ImageFont.truetype("arial.ttf", 18)
  for line in lines:
    draw.text((10, top), line, (255,255,255), font=font)
    top += 25
  image.save('cal%0.3d.gif' % index)

top = 450
for index in range(len(lines)*8):
  print index,top
  frame(index,top)
  top -= 4

This yields a sequence like cal000.gif…calnnn.gif.

I wasn’t sure how to make a video directly from that sequence, but I knew that ImageMagick could turn it into an animated GIF, like so:

convert -adjoin cal???.gif animation.gif

So I did that, and went looking for ways to convert that into a video format. ffmpeg will do it, but the results weren’t pretty, and ffmpeg can be a dicey thing to ask people to install. QuickTime, I found, did a better job. You’d need QuickTime Pro for Windows, which isn’t free, but $30 won’t break the bank.

Now the question became: How to automate the QuickTime conversion? I installed the QuickTime SDK, went looking for examples, and found just what the doctor ordered. Thanks, Luc-Eric!

Luc-Eric’s JavaScript solution, which runs on the Windows command line courtesy of the Windows Script Host, turned out to provide a double benefit. In addition to showing how to automate the conversion of a batch of GIFs to an AVI, it showed me that there was, in fact, no need to produce an intermediate animated GIF. You can just point QuickTime at the sequence in the same way that you can point ImageMagick at the sequence. I hadn’t known that! So ImageMagick dropped out of the toolchain, and there was one less component to require the station to install.

So that’s where things stand. I’m pretty sure there’s a better way to meet the requirements, and I’ll be delighted to discover it. But maybe there isn’t, in which case it looks like this will work.

Either way, it’s the end result that will — or maybe won’t — matter. We’ll do the experiment, and we’ll find out.

A conversation with Jean-Claude Bradley about open notebook science and the educational uses of Second Life

On this week’s ITConversations show I finally got to meet Jean-Claude Bradley, the Drexel chemistry professor who coined the phrase open notebook science and who champions the principles behind it.

There were a couple of surprises for me. First, I was intrigued to learn about Jean-Claude’s vision for mechanized research. I’ve always thought of open notebook science as a way to speed up the iterative cycle of research and publication, and to engage more human minds in collaboration. Of course Jean-Claude thinks so too. But he also thinks that when data are published in accessible formats, and exposed to computational processes running in the cloud, we’ll be able to automate certain aspects of research.

It reminds me of George Hripcsak’s effort to mechanize the interpretation of electronic health records. In general, we’re collecting way more data than the collectors can analyze. Crowdsourcing is one solution to this problem. Mechanization is another. We’ll need both.

The other surprise was hearing about Drexel’s fairly aggressive use of Second Life. I’ve been an amused skeptic on that front, but Jean-Claude’s passionate advocacy requires me to rethink that stance.

What didn’t surprise me, but might well surprise tuition-paying parents of Drexel students, was Jean-Claude’s attitude toward the classroom. He mostly doesn’t see a need for it. The content delivery aspect of education, he feels, is best handled in other ways, including screencasts and podcasts as well as traditional texts. There can, and should, be a range of sources, to accommodate the differing inclinations of learners. And teachers need to be competent producers and orchestrators of those sources. But for Jean-Claude, the best way to engage directly with students is to meet with individuals, not with whole classes.

Now admittedly, a chemistry class doesn’t invite and thrive on group discussion in the same way that, for example, a literature class does. And yet Jean-Claude says that a literature class was one of the models for his use of Second Life. When group interaction is central to the educational experience, he thinks that virtual environments — though he doesn’t require their use — may outperform real ones.

I remain skeptical on that point, but I’m always open-minded, so I hope Jean-Claude will take me up on my offer to visit one of his virtual environments and document the interactions that happen there.

The story of the WorldWide Telescope

My guests for this week’s Perspectives are Microsoft researcher Curtis Wong and Harvard-Smithsonian science educator Roy Gould. At Ted 2008 they jointly delivered the first preview of the WorldWide Telescope, an elegant and powerful application for exploring the sky and weaving narratives about it. In this extended interview, you can hear (or read) the whole story behind the WWT.

I’d known that the WWT was based on Jim Gray’s work, and also that it was dedicated to him. I’d also heard several of the talks he’d given about SkyServer, SkyQuery, and the SQL and XML web services technologies powering those projects.

What I hadn’t fully grasped, until I began preparing for the interview with Curtis and Roy, was Jim Gray’s larger vision for that work. In 2002, with Alex Szalay of Johns Hopkins, he published a paper entitled The World-Wide Telescope: An Archetype for Online Science. Here’s the abstract:

Most scientific data will never be directly examined by scientists; rather it will be put into online databases where it will be analyzed and summarized by computer programs. Scientists increasingly see their instruments through online scientific archives and analysis tools, rather than examining the raw data. Today this analysis is primarily driven by scientists asking queries, but scientific archives are becoming active databases that self-organize and recognize interesting and anomalous facts as data arrives.

Although the WWT isn’t an instrument for professional scientists, Roy Gould thinks it will be used by citizen scientists to collaboratively search the fast-growing corpus of sky imagery. That is, of course, a poignant echo of the collaborative search for Jim Gray when his sailboat went missing.

But for Curtis Wong and Roy Gould, who grew up in Los Angeles and New York, respectively, where neither had access to the dark night sky, the WWT is first and foremost a way to reacquaint our society with the night sky, and to teach us about the universe.

Roy Gould says that when his team surveyed high school students around the country, they found that a majority believe that stars reside within the orbit of Pluto. They also believe that galaxies are closer than stars, because “stars are just point sources, no matter what the magnification, so they must be very far away, whereas galaxies, whatever they are, look big, so they must be closer.”

To fulfill its educational mission the WWT delivers seamless navigation of the sky, contextualized in a variety of ways. Objects are described onscreen, and linked to sources on the web. When you find your way to a stellar neighborhood, thumbnails of the objects in that neighborhood invite you to explore images from a variety of catalogs: the Sloan Digital Sky Survey, Hubble, Chandra.

What’s more, the imagery is correlated so you can see the same object in any of the wavelengths of light used to observe it. If you look at the Milky Way in the standard view, and then switch to infrared, a band of incandescent whiteness emerges from the cloud of stars.

You can use the WWT to explore the sky randomly, but most people will enjoy taking one of the guided tours. Curtis Wong’s lifetime of experience as a creator of interactive multimedia is distilled into this feature of the WWT. Tours are slideshows that move from one object in the sky to the next, and may be annotated with text, spoken-word audio, and music. But at any point you can pause the tour — or hop off the bus, as Curtis says — and explore the neighborhood on your own.

The WWT isn’t just a player of tours, it’s also an authoring tool for creating them. You create slides, navigate to objects in the sky, annotate them, and save the results in an XML format that you can reuse and share.

Like images from catalogs, tours are contextually available. So if you happen upon the Ring Nebula while exploring randomly, and if there’s a tour that mentions the Ring Nebula, then that tour will surface.

Curtis envisions a hypermedia web of sky narratives. For him, this storytelling aspect really is the heart of the project. In the interview he reveals for the first time that an early prototype for the WWT, shelved years ago, was to have been called John Dobson’s Universe.

Dobson, a leading amateur astronomer and innovative telescope builder, founded San Francisco Sidewalk Astronomers, a group that encourages telescope owners to take their telescopes out in public and share their knowledge of the sky. The WorldWide Telescope is poised to carry on that great tradition, and take it in some amazing new directions.

A conversation with George Hripcsak about electronic health records and clinical truth

George Hripcsak, professor of biomedical informatics, is one of the recipients of a Microsoft Research grant to support work on the computational challenges of genome-wide association studies. These studies involve scanning complete human genomes, and looking for correlations between certain markers of genetic variation and certain diseases.

Doing that correlation is a computational challenge, but as I learned in my interview with George Hripcsak for Perspectives, that isn’t the challenge his research addresses. Instead he’s tackling a different challenge: mining electronic health records to figure out what they say about the diseases patients may have.

Why? Suppose you’ve sequenced the DNA of thousands of people for a study. If you’re trying to correlate genetic markers with disease, you need to know what diseases those people have. George calls this “collecting the phenotype” — that is, the expression of the genes responsible for diabetes, or a tendency to complications in labor, or whatever.

Traditionally that’s done by interviewing patients, a painstaking process that doesn’t scale. Given electronic health records, how much of this phenotype collection can be done automatically, and to what level of accuracy? That’s a different kind of computational challenge.

There are basically two ways to go. You can try to templatize the process of clinical data collection, so that health records can be harvested more effectively by researchers. Or you can try to understand the language that clinicians actually use when they describe patients.

For a decade now, George Hripcsak and his colleagues have been pursuing the latter approach, using a system for understanding natural language called MedLEE, which was developed at Columbia.

Ultimately I believe, as George Hripcsak does, that we’ll need a hybrid system that makes use of both structured templates and natural language understanding. But given that health records must primarily serve patient care, and can only secondarily serve research, I like how he harmonizes those objectives:

To the degree we make documentation efficient in serving health care, I think it’ll also be more accurate for the sake of research. If you’re filling out a record for the sake of billing, you’ll have an incentive to use diagnosis codes that optimize billing. Does that then reflect clinical accuracy? And would that then be useful for research?

The important thing is to be grounded in the clinical truth. Put health care first, and then use new computational methods to extract accurate information.

Amen.