Error in statistical tests of error in statistical tests

Background A recent paper found that terminal digits of statistical values in Nature deviated significantly from an equiprobable distribution, indicating errors or inconsistencies in rounding. This finding, as well as the discovery that a large percentage of p values were inconsistent with reported test statistics, led to a great deal of concern in the popular press and scientific community. The findings ultimately led to new guidelines for all Nature Research Journals. Methods We checked the statistical analysis behind the original paper's tests of equiprobability. Results The original paper tested equiprobability with the Kolmogorov-Smirnov test outside its regime of validity. Correct tests find no statistically significant deviations from equiprobability for the statistical values in Nature. Conclusion Statistical tests should be used correctly.


Background
A recent paper concluded that "statistical practice is generally poor, even in the most renowned scientific journals" [1]. The paper prompted significant attention in the popular press, and serious concern within the scientific community [2][3][4][5][6][7]. It led the editors of Nature Medicine to review their statistical practices, ultimately resulting in new statistical guidelines for all Nature Research Journals [3].
One of the two main results of [1] was that terminal digits of statistical values in Nature deviated significantly from an equiprobable distribution, indicating errors or inconsistencies in rounding. The authors of [1] collected random samples of test statistics and p values published in Nature, and looked at the terminal digits of these numbers. Their raw data is shown in tables 1 and 2. They argued that these terminal digits should be spread evenly among the ten possible digits. Applying the Kolmogorov-Smirnov test with SPSS for Windows, they obtained Z = 2.7, p < 0.0005, for the 610 test statistics, and Z = 1.4, p =

Methods
We ran tests of equiprobability on terminal digits of the test statistics and p values in Nature, using both χ 2 tests, and a modification of the Kolmogorov-Smirnov test for categorical data.

Results and Discussion
The Kolmogorov-Smirnov test is normally used to test whether data follows a specified continuous distribution [8]. A simple calculation from the raw data in [1] shows that the Z and p values obtained there are those based on comparing the data with a distribution uniform on the continuous interval [0,9]. But this distribution is obviously incorrect even before any comparison with the data, since the terminal digit cannot be, for example, 2.68. A check of the documentation for SPSS for Windows confirms that the program runs the Komogorov-Smirnov test for a continuous uniform distribution, rather than a discrete uniform distribution.
Because the terminal digits are naturally discrete, a χ 2 test is appropriate [8]. χ 2 tests yield χ 2 = 6.5, df = 9, p = 0.69, for the 610 test statistics, and χ 2 = 15, df = 9, p = 0.086, for the 181 p values. This changes the results from "significant" to "not significant," and we therefore have insuffi-cient evidence to suggest terminal digit errors in the p values reported in Nature articles.
Because the reader may be suspicious that this is simply a judgment call as to the most natural statistical test, rather than a bona fide mistake in [1], we also rerun the Kolmogorov-Smirnov test. While it is unusual to run this test on discrete data, it is possible (although perhaps poorly motivated in this case), so long as appropriate modifications are made. Instead of incorrectly comparing the data against the distribution P(x) = 1/9 for 0 ≤ x ≤ 9, we use P(x) The authors of [1] also found that 21 of the 181 p values in Nature had problems when compared with the corresponding test statistics. Our analysis does not change this finding, but it is worth remarking on the comparatively minor nature of many of the problems that they found. For example, 3 of the 21 problems come from a single three-row table (table 1 of [9]), in which every entry of a column labelled P reads 0.001. This is indeed somewhat misleading, since the natural implication is that p = 0.001 for all three entries, when in fact the intended meaning of the table was (presumably) that all three results were significant at the 0.1% level (i.e. p < 0.001). But it is hard to imagine that many readers would be badly misled by this table, and in any event, such errors are minor in comparison to the error of using a demonstrably invalid test.

Conclusion
The authors of [1] concluded that statistical tests in papers need to be inspected more closely. However, one of the main findings of their paper is invalidated by incorrect use of a statistical test. It is ironic that despite the great attention that their paper has attracted over the last two years, this error has escaped notice. While their paper still points to the need for greater scrutiny of statistics, that scrutiny would be better directed at the assumptions used in the statistical tests, rather than at the precise p values obtained.

Competing interests
The author(s) declare that they have no competing interests.