Double Data Entry: What Value, What Price?

https://doi.org/10.1016/S0197-2456(97)00096-2Get rights and content

Abstract

We challenge the notion that double data entry is either sufficient or necessary to ensure good-quality data in clinical trials. Although we do not completely reject that notion, we quantify some of the effects that poor quality data have on final study results in terms of estimation, significance testing, and power. By introducing digit errors into simulated blood pressure measurements we demonstrate that simple range checks allow us to detect (and therefore correct) the main errors that impact the final study results and conclusions. The errors that cannot easily be detected by such range checks, although possibly numerous, are shown to be of little importance in drawing the correct conclusions from the statistical analysis of data. Exploratory data analysis cannot identify all errors that a second data entry would detect, but on the other hand, not all errors that are found by exploratory data analysis are detectable by double data entry. Double data entry is concerned solely with ensuring, to a high degree of certainty, that what is recorded on the case record form is transcribed into the database. Exploratory data analysis looks beyond the case record form to challenge the plausibility of the written data. In this sense, the second entering of data has some benefit, but the use of exploratory data analysis methods, either as data entry is ongoing or at the end of data entry and as the first stage in an analysis strategy, should always be mandatory.

Introduction

Few people would dispute the notion that good-quality data are important and certainly preferable to poor-quality data. Although we concur with this notion, we cannot explicitly state what we mean by “good” or “bad” data quality, nor can we specify whether the quality of any specific dataset is adequate for the intended analysis. The better the quality of the data, the more reliable and convincing the inferences will be; thus, on any arbitrary scale that measures quality, all we can really say is that better-quality data lead to sounder and more reliable conclusions. Whether quality is already good or bad, moving along that scale to improve quality seems desirable. We do not have many ideas about how much quality matters despite the fact that quality is often simple to measure.

Many steps are involved in getting data from a patient into a final report [1]. Each may introduce a variety of errors. We believe that building quality into systems is more productive than building checks onto the end.

Despite the many points data pass on their route from patient to final report, data entry may be the only one where, quite routinely, we process everything twice. A few studies have duplicate investigators' assessments; none is likely to have duplicate case record forms (CRFs) to catch transcription errors; the concept of typing a final report twice to check for typographical errors is almost laughable. Rarely, two independent statisticians may analyze the data or the statistician may use more than one software package. So why double data entry but not double everything else? We return to this question later in the discussion.

Although double data entry is useful for identifying typographical errors, we have no objective data on its effectiveness. Some of the differences (or errors) it identifies may be of interpretation; others may be simple typing errors, but we cannot easily determine the number of errors that are not identified by double data entry. What effect do these errors have? Can errors be found in other ways? Does double data entry add value or waste time? “What is of interest is not the error rates themselves but whether the conclusions differ before and after correcting the errors” [2].

Others who have considered some aspects of this question have discussed what is achievable or what types of error rates result from different data entry procedures. Reported achievable rates vary. Neaton et al [3]cites 10 per 10,000 fields; Shenker [4]cites 7 per 10,000 keystrokes (in a project involving 350 million keystrokes); and Jeffreys (see Hosking et al [5]) reported that in a given study, error rates fell over time from 10 to between 4 and 8 per 10,000 fields.

None of these authors, however, addresses whether those errors mattered. Hilner et al [6]describe reviewing dietary data and report critical errors at 0.12%. They define critical errors as “those impacting on nutrient computations.” Such errors might be specifically related to the primary outcome but still not critical in the sense of affecting the conclusions drawn from the study. Other papers have discussed assessments of data quality [7]and objectivity in measuring data quality [8]. An example of objectivity is whether to regard a date as one or three fields. Are errors in the day and month, for example, two errors or simply one? Additional material has been published that directly compares, in an experimental setting, single versus double data entry [9]. Seaman (without specifying single or double data entry) notes that databases have sometimes completely missed some data on the CRF [10].

This article uses an experiment to consider the value of double data entry. We assume that data entry errors occur at random, and that this affects each treatment group similarly.

Section snippets

The model

Consider a hypothetical dataset of blood pressure measurements with two treatment groups, each of size n patients, each patient having systolic blood pressure independently and identically distributed as a log-Normal. Under the null hypothesis, both simulated treatment groups have the same mean, but we also consider alternative hypotheses in which group two has a mean 5 mm, 10 mm, or 15 mm larger than group one. We calculated summary statistics and t-tests for each of N replicate trials and

Methods

We used SAS [16]for all simulations. Random samples Xi, generated as independent observations from a Normal distribution with mean zero and unit variance were converted to a log-Normal distribution with an appropriate mean and variance by the transformation Yi = nearest integer {exp (0.0909 Xi + 4.7)} [4.7 being log (110) and 0.0909 the coefficient of variation, namely σ/μ = 10/110]. Each observation was assigned a random digit Ri uniformly distributed on (0,1), and we introduced errors into

Results

Table 1 shows a typical small dataset illustrating the kinds of errors that may be introduced and the possible corrective action. In treatment group 1, a score of Y6 = 107 mm Hg has been erroneously transcribed as Y6* = 137 mm Hg. Such a value does not appear unreasonable and so would not be queried and corrected. In contrast, the value of Y7 = 109 erroneously transcribed as 189 does look unreasonable. Such a score would be checked and subsequently corrected. The second patient (with a score of

Discussion

We do not advocate or even passively support sloppiness, but we wish to challenge such statements as, “Good-quality data are essential to drawing the right conclusions necessary for successful clinical research” [18]that are often made without any justification or quantification. We do not even advocate that double data entry should not form a part of common work practice. Double data entry has been described as being akin to an insurance policy [19], although we suggest that it may be likened

References (20)

There are more references available in the full text version of this article.

Cited by (74)

  • An automated data verification approach for improving data quality in a clinical registry

    2019, Computer Methods and Programs in Biomedicine
    Citation Excerpt :

    It consists of two steps: an initial entry step and a verification step; each step is performed by a separate data entry clinician. However, it is time-consuming, labor-intensive [8,9] and ineffective for detecting recording errors. Another controversial approach is source data verification (SDV), a verification of conformity of the data recorded in case report forms with source data [10].

  • Using Feedlot Operational Data to Make Valid Conclusions for Improving Health Management

    2015, Veterinary Clinics of North America - Food Animal Practice
    Citation Excerpt :

    Examples in human medicine have shown how computerized data entry forms may be used to reduce medication errors by not having to interpret handwriting, automating dosage calculations, and improving identification of the prescribing physician.3 However, 1 study of handheld computers in human medicine indicated that there were 675 errors per 10,000 entry fields, which is greater than the acceptable rate of 10 errors per 10,000 fields using paper-based double data entry methods.4,5 Computerized health records have become more available in veterinary medicine, yet, data errors are still a potential concern.

View all citing articles on Scopus
View full text