getAllStatuses(); $renderer = get_active_status_renderer(); echo ''; } function topic_icons_css() { echo "\n"; } function topic_icons_label( $label ) { global $topic; if (bb_is_front() || bb_is_forum() || bb_is_view() || bb_is_tag()) { $icon_set_name = topic_icons_get_active_icon_set(); $icon_set_url = ICON_SET_URL_BASE . $icon_set_name; $status = get_active_status_interpreter()->getStatus(bb_get_location(), $topic); $renderer = get_active_status_renderer(); $image = $renderer->renderStatus($status); $tooltip = $renderer->renderStatusTooltip($status); $exists = file_exists(dirname(__FILE__).'/icon-sets/'.$icon_set_name.'/'.$image); if (!$exists) { return sprintf(__('
%s'), get_topic_link($topic->topic_id), ICON_SET_URL_BASE.'/empty.png', ICON_WIDTH, ICON_HEIGHT, $tooltip, $label); } else if (strlen($tooltip) > 0) { return sprintf(__('
%s'), get_topic_link($topic->topic_id), $icon_set_url.'/'.$image, ICON_WIDTH, ICON_HEIGHT, $tooltip, $tooltip, $label); } else { return sprintf(__('
%s'), get_topic_link($topic->topic_id), $icon_set_url.'/'.$image, ICON_WIDTH, ICON_HEIGHT, $tooltip, $label); } } return $label; } function topic_icons_init( ) { remove_filter('bb_topic_labels', 'bb_closed_label', 10); remove_filter('bb_topic_labels', 'bb_sticky_label', 20); add_filter('bb_topic_labels', 'topic_icons_label', 11); add_action('bb_head', 'topic_icons_css'); add_action('bb_admin_menu_generator', 'topic_icons_admin_page_add'); add_action('bb_admin-header.php', 'topic_icons_admin_page_process'); topic_icons_register_status_interpreter('default', new DefaultStatusInterpreter(BUSY_THRESHOLD)); topic_icons_register_status_renderer('default', new DefaultStatusRenderer()); } topic_icons_init(); ?> Digital Humanities Questions & Answers » Topic: Verifying the Accuracy of Double Keying - any studies? Digital Humanities Questions & Answers » Topic: Verifying the Accuracy of Double Keying - any studies? en-US Sun, 24 Mar 2019 23:11:16 +0000 <![CDATA[Search]]> q lit_cht on "Verifying the Accuracy of Double Keying - any studies?" Fri, 06 Jul 2012 15:02:26 +0000 lit_cht 1721@ <p><em>Replying to @<a href=''>lit_cht</a>'s <a href="">post</a>:</em><br /> Dear Colleagues, though it has been some time since this discussion came to a preliminary end, I want to thank you again for all your hints. For those generally concerned with Quality Assurance in large TEI corpora, Deutsches Textarchiv's recently published article might be of interest. English abstract below, text is German, though. All the best<br /> - for the DTA-Team -<br /> Christian Thomas</p> <p>Alexander Geyken, Susanne Haaf, Bryan Jurish, Matthias Schulz, Christian Thomas, Frank Wiegand: "TEI und Textkorpora: Fehlerklassifikation und Qualitätskontrolle vor, während und nach der Texterfassung im Deutschen Textarchiv." In: Jahrbuch für Computerphilologie, <a href="" rel="nofollow"></a> [retrieved 2012-08-09].</p> <p>Abstract:<br /> This paper deals with the issue of quality assurance in very large, XML/TEI-encoded full-text collections. The text corpus edited by the DFG-funded project Deutsches Textarchiv (henceforth: DTA), a large and still growing reference corpus of historical German, is a fine example of such a collection. The following remarks focus on text prepared in a Double-Keying-process, since the major part of the DTA-corpus is com-piled by applying this highly accurate method. An extensive and multi-tiered approach, which is currently applied by the DTA for the analysis and correction of errors in double-keyed text, is introduced. The process of quality assurance is pursued in a formative way in order to prevent as many errors as possible, as well as in a summative way in order to track errors which nevertheless may have occurred in the course of full-text digitization. To facilitate the latter, DTAQ, a web-based, collaborative tool for finding and commenting errors in the corpus, was developed. On the profound basis of practical experience in the past four years, the preliminaries and possible methods of conducting a widespread quality assurance are being discussed. </p> lit_cht on "Verifying the Accuracy of Double Keying - any studies?" Wed, 24 Aug 2011 10:43:22 +0000 lit_cht 1350@ <p>Dear Collegues, thank you very much for your helpful insights, hints and suggestions. We will continue to run a study of our own with the various volumes double-keyed in the past 3 years for Deutsches Textarchiv (<a href="" rel="nofollow"></a>. Presentation in Wuerzburg, tei-mm (<a href="" rel="nofollow"></a>. Hope to see you there, best wishes,<br /> Christian Thomas </p> Wesley on "Verifying the Accuracy of Double Keying - any studies?" Fri, 12 Aug 2011 03:08:16 +0000 Wesley 1339@ <p>In response to your Humanist post, I can offer some insight. I double-keyed the text of a long novel for my dissertation. Though my purpose was not to assess the accuracy of double-keying but rather to achieve the most accurate textual record and to state clearly what my estimate of errors is. It is not truly a controlled test. I assume that one text was single-keyed. I keyed the text myself. I compared the two keying. And I proofread. </p> <p>In a text of approximately 700,000 characters, I believe that around 200 characters were in error after double-keying. (99.999 percent accurate). By a system of planting errors, I was able to estimate that oral proofreading captured 80 percent of errors that remained. </p> <p>When I checked another text (single-keyed, I’m quite certain) it was accurate at a rate of 97.5 percent. </p> <p>The dissertation is posted here. </p> <p><a href="" rel="nofollow"></a></p> <p>See pages 37 to 38 for a discussion of the method for planting and estimating errors. And see page 63 for estimate of accuracy. </p> Dorothea Salo on "Verifying the Accuracy of Double Keying - any studies?" Thu, 11 Aug 2011 02:06:46 +0000 Dorothea Salo 1332@ <p>I should probably clarify at this point (since PFSchaffner is absolutely right to question the "oh, just leave it to the vendor" vibe I was giving off) that at a publishing-services bureau, what's getting keyed is generally recently-printed books or well-edited manuscripts. Many fewer (though not quite zero) interpretation problems!</p> <p>Once you're working with less clean materials, see PFSchaffner's response; the possibilities for error and "error" are indeed vast. </p> PFSchaffner on "Verifying the Accuracy of Double Keying - any studies?" Wed, 10 Aug 2011 23:02:35 +0000 PFSchaffner 1331@ <p><em>Replying to @<a href=''>lit_cht</a>'s <a href="">post</a>:</em><br /> I am not aware of any <strong>studies</strong> of the accuracy of keyed text, but have a lot of experience assessing that accuracy over the past 14 years. Many of the digitization projects undertaken by the University of Michigan library over that period (especially the Middle English Dictionary and Compendium, and the nearly 50,000 books produced so far by the Text Creation Partnership (EEBO-TCP, Evans-TCP, and ECCO-TCP). were generated by outsourced re-keying, sampled and proofread in-house (usually at a 5% sampling rate), and evaluated against a nominal spec of 99.995% accuracy (i.e., less than one error per 20,000 bytes). Having stuck with this approach so long, clearly we find something to recommend in it.</p> <p>A few points and conclusions:</p> <ul> <li> we have never seen any point to specifying a particular *method* (and indeed some of the commercial processes are proprietary secrets): we specify a given accuracy of output, and leave it to the vendors whether that is best achieved by single keying, double blind keying, keying-and-proofing, or the use of their own algorithms, dictionaries, etc.</li> <li> 'Accuracy' and 'accuracy rate' are surprisingly problematic notions for all but the simplest and cleanest texts. Any realistic definition has to take into account the huge problem of ambiguous glyphs and glyph-to-character mappings, even when the source is clean, the typeface modern, and the orthography consistent. When the source is 'dirty' (image quality degraded), the typeface(s) old and obscure (or when handwriting is involved), or the text orthographically odd, it becomes obvious how subjective a concept 'accuracy' really is. Transcription can be every bit as interpretive as markup, and cannot really be separated from it.</li> <li> Our procedure for the TCP texts is to generate (by script) a random sample of roughly 5% of the pages in any incoming text, convert the sample to HTML for ready display, and present it alongside page images of the source for comparison by our proofreaders. Because of the 'noisiness' of many of our books, the proofreaders are instructed to distinguish between 'forced' and 'unforced' errors (the former attributable to some defect or ambiguity in the source); to count only the 'unforced' errors against spec, and to request the resubmission of books that fail to meet spec. The error count is measured against the byte count of the sample, markup excluded. There are exceptions for particular circumstances, editors retain the power to pardon as well as to reject, and small (sub-20KB) books are usually bundled together for sampling and proofing as a batch. </li> <li> Slippery definitions aside, double keying (or equivalent processes) can produce extraordinarily accurate text. When the source is of high quality and consistency, it can approach perfection, with measured accuracy rates that are statistically indistinguishable from 100%. Many of our current TCP texts are at that level; so were the transcriptions done for the e-MED. </li> <li> When keyed text fails our tests, it is difficult to know, from outside the black box where we sit, whether it is because true double keying (or equivalent) was not used, because it was deployed improperly, or whether it was intrinsically not up to the task. And the factors that most obviously affect accuracy are very difficult to quantify: image quality, linguistic difficulty (including the presence of abbreviations and variant orthographies), format and layout, typography, and the expertise and experience of the individual keyer. </li> <li> For there is an irreducible human factor to keying that resists quantification and systematization. At the level of the individual keyer, as the source becomes murkier, the process becomes more one of reading and less one of glyph decipherment--and accurate reading (as a somewhat mysterious gestalt process) varies widely with the talent, linguistic expertise, experience and focus of the individual keyer. We have received transcriptions from some keyers that seem well-nigh miraculous in their ability to extract meaningful text from a blotted, bled-through, or faint source; but also transcriptions that clearly fell too readily for the source's snares and ambiguities, or lacked sufficient knowledge of the language and conventions of the source to interpret it correctly or recognize anomalies. </li> <li> At the corporate level, too, one has to be careful to create incentives that match the character of one's source material: corporations are people too. E.g., if the source is potentially difficult, be sure not to penalize good (or even bad) guesswork, or you will end up with a text full of "illegibile" flags rather than characters (conversion firms tending to prefer safe to sorry.) And I cannot agree with advice not to worry about accuracy or enforce your accuracy specification. Failure to test and verify is itself an incentive, one that will surely be taken advantage of. </li> <li> The obverse to the human factor is this: the very problems that make keying sometimes challenging can make OCR impossible. Keying is better than OCR even when the material is least problematic: as it becomes more so, keying pulls away and leaves OCR in the dust. </li> <li> We have raw proofreading numbers (byte counts, sample sizes, errors found) that we can probably extract for a decade's worth of keying. I'm not sure what you'd do with them, but do let me know if you want me to try to pull them for you. </li> </ul> <p>Good luck! </p> Dorothea Salo on "Verifying the Accuracy of Double Keying - any studies?" Wed, 10 Aug 2011 16:53:10 +0000 Dorothea Salo 1330@ <p>Well, if we're talking about TEI, there are at least two possible meanings of "accuracy" here: accuracy of rekeying of the (presumably) printed text, and accuracy of encoding.</p> <p>99%+ rekeying accuracy via double-keying is almost certainly a reliable estimate for any competent operation; the medical literature (oddly enough!) has done some empirical studies on double-keying. Since encoding is judgment-dependent, though, accuracy there will depend heavily on how clear and detailed your instructions to the keyers are.</p> <p>Back in the day when I worked for a publishing service bureau, we got excellent results from outsourced double-keying, including when markup was involved. Personally, I wouldn't stress too much over rekeying accuracy; focus on quality instructions for the TEI, and you should get decent results back. </p> lit_cht on "Verifying the Accuracy of Double Keying - any studies?" Wed, 10 Aug 2011 12:19:58 +0000 lit_cht 1322@ <p>Dear colleagues,</p> <p>as staff member of the Deutsches Textarchiv (, I would be interested in studies on the accuracy of double keying, esp. in large full text, tei/xml annotated corpora. Service providers advertise accuracy rates of 99.+%, has anyone ever questioned this on an empirical basis? I have found various musings on OCR accuracy and on how to improve results here, but nothing similar on double keying. I would be grateful for any hints.</p> <p>Thanks in advance,<br /> all the best<br /> Christian Thomas </p> <p>(I also mailed this to Humanist, as here and there will not be exactly the same people reading, I hope. If so, please apologize cross-posting...) </p>