• Skip to main content
  • Skip to secondary menu
  • Skip to primary sidebar

FromThePage Blog

Crowdsourcing, transcription and indexing for libraries and archives

  • Home
  • Interviews
  • crowdsourcing
  • how-to
  • Back to FromThePage
  • Collections

Is That Transcription Really Human?

October 21, 2025 by FromThePage

Last month, Denyse Allen asked this question on the Genealogy and AI Facebook group:
If volunteers use AI to transcribe documents, is that OK?

I have strong opinions, but want to explain them.

First off, the institutions running big crowdsourcing projects have staff who can automate sending all of their documents to AI engines for transcription. As a software developer, it makes me wince to think of humans cut-and-pasting images into LLM services, then pasting the results back into crowdsourcing platforms, when a computer program could do the same thing. What a waste of time!

That said, there is a much more important reason that libraries and archives ask humans to transcribe text: transcriptions made by humans are fundamentally different from ones made by machines.

Three kinds of transcription

When humans transcribe documents, the resulting transcript usually has these qualities:

  • Verbatim spelling – for most projects, each word will be transcribed letter-for-letter, retaining any irregularities from the original
  • High accuracy – humans are really good at reading handwriting, learning from context and returning to correct readings they got wrong after they master a particular hand or script.
  • Willingness to say “I don’t know” – humans are willing to write things like “[illegible]” or “[unclear: possibly Jones?]” when they aren’t certain about a text.

Traditional OCR/HTR systems like Kraken or Transkribus also produce transcripts with verbatim spelling, but their accuracy can be lower than humans and they usually produce gibberish if they encounter difficult words. But it’s easy to tell that these transcripts were machine-produced, because their errors are so characteristic.

I suspect that the person asking the question was using a Multi-modal Large Language Model like ChatGPT or Google Gemini. Those systems are very exciting, and really easy to use, but the transcripts they produce have very different qualities:

  • Regularized spelling – since LLMs are language models, they tend to prioritize readings that make sense. As a result, many of them silently normalize irregular words.
  • Uneven accuracy – in the case of “filler words” like prepositions and articles, LLMs don’t even “look at” the image, reverting to their “autocomplete on steroids” behavior to just add words that make sense given the context.
  • Overconfidence – if an LLM can’t read something, it will fill in words that make sense to it. This means that you can’t tell when the transcription is accurate, and when it’s just a guess.

So what happens if an archivist sees a transcription that is supposed to be created by a volunteer, but was actually produced by an LLM? Unfortunately, the transcript will make sense – the seductive plausibility of LLM output can easily trick readers–even experienced ones–into believing that the text faithfully reproduces the original. But what if the volunteer checks the AI output before contributing it? Spot-checking LLM text can be fiendishly difficult, and requires special techniques to reduce bias.

This is not going to be an easy challenge for those of us who run crowdsourcing projects. We need to explain the goal to volunteers, persuade them to avoid AI short-cuts, and build responsible (and traceable) AI tools into our software platforms.

Filed Under: ai & crowdsourcing, Uncategorized

Primary Sidebar

What’s Trending on The FromThePage Blog

  • Guide to Digitizing Your Archives
  • How Do I Read Old Handwriting?
  • An Interview with Shanna Raines of the Greenville…
  • ChatGPT, Microsoft, and the Development of Guardrails
  • Learn to Decipher Old Handwriting with Online and…
  • Environmental Impact of AI

Recent Client Interviews

An Interview with Candice Cloud of Stephen F. Austin State University

An Interview with Shanna Raines of the Greenville County Library System

An Interview with Jodi Hoover of Digital Maryland

An Interview with Michael Lapides of the New Bedford Whaling Museum

An Interview with NC State University Libraries

Read More

ai artificial intelligence crowdsourcing features fromthepage projects handwriting history iiif indexing Indianapolis Indianapolis Children's Museum interview Jennifer Noffze machine learning metadata newsletter ocr paleography podcast racism Ryan White spreadsheet transcription transcription transcription software

Copyright © 2026 · Magazine Pro on Genesis Framework · WordPress · Log in

Want more content like this?  We publish a newsletter with interesting thought pieces on transcripion and AI for archives once a month.


By signing up, you agree to our Privacy Policy and Terms of Service. We may send you occasional newsletters and promotional emails about our products and services. You can opt-out at any time.  We never sell your information.