Excellent debate visualizer at NYTimes.com

Kudos to the New York Times for its remarkable interactive transcript of the Republican debate. The display has two tabs: Video Transcript, and Transcript Analyzer. The Video Transcript is a side-by-side display of the video and the transcript, linked together and randomly accessible from either side, plus a list of topics that you can jump to. These are the kinds of features you’d like to be able to take for granted, but which aren’t always implemented as cleanly as they are here.

But the Transcript Analyzer takes the game to a new level. Or at least, one that I haven’t seen before in a mainstream publication. The entire conversation is chunked and can be visualized in several ways. It’s reminiscent of the Open University’s FlashMeeting technology which I mentioned here.

In the Times’ visualizer, you can see at a glance the length and wordcount of all participants’ contributions — including YouTube participants who are aggregated as “YouTube user”. Selecting a participant highlights their contributions, and when you mouse over the colored bars, that section of the transcript pops up.

Even more wonderful is the ability to search for words, and see at a glance which chunks from which participants contain those words. The found chunks are highlighted and, in a really nice touch, the locations of the found words within the chunks are indicated with small dark bars. Mouse over a found chunk, and the transcript pops up with the found words in bold. Wow! It’s just stunningly well done.

The point of all this, of course, is not to exhibit stunning technical virtuosity, although it does. The point is to be able to type in a word like, say, energy, and instantly discover that only one candidate said anything substantive on the topic. (It was Mitt Romney, by the way.) Somehow, in all of the presidential campaigning, that topic continues to languish. But with tools like this, citizens can begin to focus with laserlike precision not only on what candidates are saying, but also — and in some ways more crucially — on what they are not.

Hats off to the Times’ Shan Carter, Gabriel Dance, Matt Ericson, Tom Jackson, Jonathan Ellis, and Sarah Wheaton for their great work on this amazing conversation visualizer.

16 thoughts on “Excellent debate visualizer at NYTimes.com

  1. Tony

    Jon, thanks for sharing and bringing this to people’s attention. I definitely agree that this is a phenomenal piece of technology that adds tremendous value. I was able to look at the visualization and hone in on keywords that were important for me. Also, I was able to go directly to the questions I was really concerned about.

    Kudos to the folks at the NY Times! Keep up the great work!

    Reply
  2. gabriel dance

    jon,
    thanks for the complimentary write-up. i’m stoked that you find the tool impressive, and even happier that you find it useful. that’s the goal. as far as it not working in firefox, we’re certainly hoping that’s not the case. try refreshing. sometimes i think the streaming server gets a little wonky, but a refresh should set u up right.

    Reply
  3. Pingback: Analizador de debates en el New York Times | Denken Über

  4. martin langhoff

    The engine behind this is great. I wonder how much work is involved in getting the transcription, speakers and timers.

    I work on Moodle (the learning/course management system), and it would be great to be able for the lecturer to upload a video of a lecture, and have a wiki-style page where other can – if they want – transcribe fragments in a simple format. Would also be great for conference videos that are posted on the internet for a long time – the ones that people think are important would get stand a chance of getting transcribed. And with the right payoff — and this kind of browsing is a great payoff — people will be keen on doing it.

    Reply
  5. Pingback: Transcript Analyzer for Republican Debate

  6. Mary Branscombe

    Blinkx uses Autonomy technology to do transcriptions of podcasts and videos but they don’t publish them because the copyright position is uncertain; I long for transcripts of the rich media that’s online so we can mine it like this.

    Reply
  7. Pingback: A conversation with Carl Malamud about access to public information « Jon Udell

  8. Pingback: A conversation with Gabriel Dance and Shan Carter about interactive graphics at the New York Times « Jon Udell

  9. Pingback: A conversation with Gabriel Dance and Shan Carter about interactive graphics at the New York Times | RatZine - Rat stinkin news

  10. Pingback: A conversation with Gabriel Dance and Shan Carter about interactive graphics at the New York Times | RatZine - Rat stinkin news

  11. Pingback: A conversation with Gabriel Dance and Shan Carter about interactive graphics at the New York Times | MyNewsPlus.com

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out / Change )

Twitter picture

You are commenting using your Twitter account. Log Out / Change )

Facebook photo

You are commenting using your Facebook account. Log Out / Change )

Google+ photo

You are commenting using your Google+ account. Log Out / Change )

Connecting to %s