I wanted to write this month about ChatGPT, and how archives are about as anti-ChatGPT as you can get. Archives can provide learning experiences that ChatGPT can’t fake. First, a simplification that’s useful in thinking about what ChatGPT (and its ilk) can – and can’t – do. ChatGPT is, in technical terms, a “large language model”. That means that the creators fed it massive … [Read more...] about Archives as an Antidote for ChatGPT
Trust in the Truth
Happy February. I thought I’d write about how transcription in the classroom could counter the challenges ChapGPT introduces, but then I ran across this quote from James Baldwin and decided to ruminate on transcribing the difficult parts of history: People are trapped in history and history is trapped in them. Perhaps you're faced with the stark history of injustice -- … [Read more...] about Trust in the Truth
Using Crowdsourced Indexes in Preservica
What can you do with transcripts once your project is done? The Texas Digital Archive transcribed the handwritten index to a 3rd court of appeals and turned it into a look up table by Appellant and Appellee, making their digitized case records much more accessible. The Texas Commission on Libraries and Archives transcribed a handwritten index to their Court of … [Read more...] about Using Crowdsourced Indexes in Preservica
How Do Holidays Affect Crowdsourcing?
How do holidays affect crowdsourcing in cultural heritage? Many people are away from work, spending time with family and friends, but is that true of the retirees that make up the bulk of our volunteers? We thought about this last month, just after Thanksgiving in the USA. A regional holiday seemed like a good opportunity to look for patterns. Over the previous … [Read more...] about How Do Holidays Affect Crowdsourcing?
On Improving Error & Quality in Crowdsourced Transcription
Last June, Ben Brumfield of FromThePage and Austin Mast of Florida State University/iDigBio had a meaty discussion of error and quality in crowdsourced transcription. The discussion, which was moderated by Sara Brumfield, started with the dimensions of quality, compared multi-track (i.e. many transcriptions of the same bits) and arbitration to single track (i.e. a one-person … [Read more...] about On Improving Error & Quality in Crowdsourced Transcription