Digital Scholarship and Research

We’ve spent most of this semester discussing the output side of being a digital historian, but this week we turn our attention to how we can also leverage digital technology to be better historians. For most historians, digital technology has simply sped up the process of what they already did with analog tools. For instance, many of my colleagues use a database (like FileMaker) to take notes instead of using index cards as one had typically done in the past. Others have utilized bibliographic software like Endnote, Biblioscape, Reference Manager (formerly ProCite), RefWorks, or now the free Firefox plug-in Zotero to organize their citations (which everyone should do, since it makes documenting your references sooooooo much easier and accurate).

These two examples are simply an extension of the older transition that all academics made when they switched from typewriters to word processing software and don’t necessarily “add” any value to the product of their research. However, this doesn’t mean that there isn’t a great deal of potential in the very near future.

One of my favorite digital historians is Bill Turkel up at the University of Western Onterio. Dr. Turkel represents the cutting edge of historians (in his case and environmental historian) who want to press the envelope of technology to see what it can do for us to further our basic understanding of the past. Dr. Turkel doesn’t just use computers, he tries to think about devices or add-ons for devices that don’t yet exist and then builds them! This is fascinating stuff to say the least!

Another great story of using technology to be better historians comes from the UK last year. Scientists were able to train a computer and an electronic sniffing device to recognize the age of paper without the need to carbon test a fragment (and thereby destroying a part of the original document). When I “tweeted” about this last fall, Bill Turkel wrote to remind me that he had speculated about something similar in a blog post of his in 2008.

These are just a few examples of the potential to use technology as historians to not only present our craft, but to perform our craft more effectively and possibly in entirely new ways. What we need are more people who can “think outside the box” and dare to explore new methods that take advantage of digital technology’s potential and yet remain true to the mission of historians to explore and analyze the past.

So – What’s your big idea going to be?

Digital Tools for a Digital (or sort of digital) World

This week I asked you to reflect on how the digital tools that we have at our hands is transforming the way that historians go about their business (or research). There have been many advances in technology that have allowed historians to explore new geographic areas, new methods of research, and manipulate data in new and inventive ways. The other week, I showed some examples of a new digital tool being developed by IBM’s research lab in Cambridge – ManyEyes. If you didn’t have a chance to check it out after class, I encourage you to still take a look. Here is a good example of several different ways in which ManyEyes can visualize data in a way that had not really been available to average historians without access to very powerful computing power.

This election cycle is also producing some interesting digital tools that might be of interest for present and future historians. One is an adaptation of the visualization that we find at ManyEyes and is a visualization of how people are feeling on election day.  At the New York Times, you can enter in key words about how you are feeling and add it to the data pool being collected by contributions from others around the world. You can visualize the results generically or separate out the McCain and Obama supporters. Over at the Washington Post you can track Twitter, video, and news postings that are referenced by place and time. This visualization is rather interesting because you can move the slidebar and see how coverage in different areas changes over time.

An Archivist at the Stasi Archive Puts Together a Stasi PuzzleOther interesting digital tools that have been created for historians are not yet commercially available, but the technology behind them and the potential uses by historians now and in the future are significant. One of my favorite tech developments has been going on in Germany (surprise!). On the eve of the communist party’s fall from power, the secret police (the Stasi) began shredding documents (amounting to some 16,000 sacks of shredded paper). The people at the Frauenhofer Institut, the same people who brought you the MP3) have been working on a digital solution. Funded by the German federal government and working closely with the archivists at the Stasi Archive, the technicians have been working on software that can match together the shredded and torn pieces of paper to reconstruct the original document. It is estimated that if 30 people were tasked with this job, it would take somewhere between 600 and 800 years! You can read more about this story here and here.

The other project that I am excited about was developed right here at UMass! Working with a set of letters from George Washington, the Center for Intelligent Information Retrieval developed software that can learn how Washington wrote and translate this into data, which in turn can be searched and manipulated. For many of us who have to deal with handwriting manuscripts (luckily most of my sources are typed, but there is a certain degree of marginalia involved as well as occasional handwritten items) being able to scan, store, and manipulate such documents will be a great leap forward. Just take a look at the sample of late 18th century German handwriting (Goethe’s manuscript of Werther) on the left to see how difficult such sources can be to work with.

Speech recognition remains the Holy Grail of the digital historian. Back in 1998 or 1999 (I can’t remember exactly when this was) I attended a German-American young-leaders conference and had been assigned the job of introducing one of our keynote speakers, Walter Isaacson, who was the Editor of Time at the time. We had arranged to have breakfast together so that I could go over a few things with him and I was just making small talk when I asked him what he was currently working on. He responded that he had just gotten back from a week-long visit to Microsoft for a series they were working on about the next generation operating system (what would become XP). He was visiting their voice recognition lab and he noticed a banner hanging on the wall that read – “Wreck a Nice Beach!”  He asked “Bill” what the sign was for – the response was that the unit can takedown the banner when saying “recognize speech” does not translate into “wreck a nice beach” on the screen! Then, and only then, Microsoft would begin to seriously think about integrating this technology into Windows. I don’t use Vista, but I hear that it does have some dictation software built in that is rather capable. The most robust software for PCs out there today is Dragon Naturally Speaking. For Mac users, you need to use MacSpeech Dictate (which licences the technology behind its software from Dragon). IBM had been developing ViaVoice for both platforms, but sold it to ScanSoft (now Nuance) in 2003. Nuance also owns Dragon, OmniPage, and PagePort. Thus, we really only have one speech recognition technology open to the end user at this point. Although each of these performs rather well, they are all designed to capture dictation and not natural speech. Punctuations need to be spoken – “period,” “comma,” “new paragraph,” etc. Even with a claim of 98% accuracy, an 8,000 word entry (this is what a professional transcriptionist can usually type per hour) will still contain some 180 errors that need to be corrected manually, not to mention all of the punctuation marks that I just mentioned.