Parametric methods outperformed nonparametric methods in comparisons of discrete numerical variables
 Morten W Fagerland^{1}Email author,
 Leiv Sandvik^{1} and
 Petter Mowinckel^{2}
DOI: 10.1186/147122881144
© Fagerland et al; licensee BioMed Central Ltd. 2011
Received: 11 November 2010
Accepted: 13 April 2011
Published: 13 April 2011
Abstract
Background
The number of events per individual is a widely reported variable in medical research papers. Such variables are the most common representation of the general variable type called discrete numerical. There is currently no consensus on how to compare and present such variables, and recommendations are lacking. The objective of this paper is to present recommendations for analysis and presentation of results for discrete numerical variables.
Methods
Two simulation studies were used to investigate the performance of hypothesis tests and confidence interval methods for variables with outcomes {0, 1, 2}, {0, 1, 2, 3}, {0, 1, 2, 3, 4}, and {0, 1, 2, 3, 4, 5}, using the difference between the means as an effect measure.
Results
The Welch U test (the T test with adjustment for unequal variances) and its associated confidence interval performed well for almost all situations considered. The BrunnerMunzel test also performed well, except for small sample sizes (10 in each group). The ordinary T test, the WilcoxonMannWhitney test, the percentile bootstrap interval, and the bootstrapt interval did not perform satisfactorily.
Conclusions
The difference between the means is an appropriate effect measure for comparing two independent discrete numerical variables that has both lower and upper bounds. To analyze this problem, we encourage more frequent use of parametric hypothesis tests and confidence intervals.
Background
Categorical, or discrete, data are characterized by having a finite number of categories or values, whereas continuous data can take on any real value within a given range. For a categorical variable with more than two categories, we distinguish between nominal and ordered variables. Ordered variables have a natural ordering to the categories, for example, degree of pain classified as none, mild, moderate, or severe.
Sometimes, we have data that are essentially categorical, but with numerical properties, or numerical data that can take on only a small number of values. We shall refer to such data as discrete numerical [[1], p.11]. In medical research, discrete numerical data arise mostly in situations where we count the number of events per individual, such as the number of clinical visits, the number of adverse events, or the number of units of blood transfused. As a preliminary assessment of the prevalence of variables reporting the number of events, we considered all randomized, controlled trials (RCTs) published in January and February 2010 in the New England Journal of Medicine, Lancet, Journal of the American Medical Association, and BMJ. Out of a total of 52 papers, 24 (46%) papers reported at least one variable describing the number of events; 16 (31%) papers reported baseline variables, and 15 (29%) papers reported outcome variables.
Discrete numerical data are a blend between categorical and continuous data, and it is not obvious how to analyze such data. Of particular interest is how to compare two independent discrete numerical variables, a common problem in comparisons of two treatment or exposure groups. Should we analyze discrete numerical variables using methods for continuous or for ordered categorical data?
The main problem with using methods for ordered categorical data is information loss. Statistical methods for ordered categorical data do not treat the distance between values or categories as constant. Thus, we may not be getting the most out of the data. Even worse in that regard is to combine the outcomes into two categories and use methods for binary data. By doing so, we may throw away a lot of information. Poor power is often the result, and estimates may be inaccurate [2, 3].
If we intend to analyze discrete numerical data without discarding relevant information, we should consider treating the variables as if they were continuous. Continuous variables with an approximately normal distribution are best analyzed using parametric methods for confidence intervals and hypothesis tests [4, 5]. The usual alternative is a nonparametric test and a nonparametric or bootstrap confidence interval, or a transformation, for example, the logarithmic, prior to parametric methods. To decide if parametric methods are appropriate, the shapes of the underlying distributions are estimated by inspecting histograms, QQplots, and sample moments, or by using prior knowledge about the variable of interest. For discrete numerical variables, however, such tools may not be relevant because of the discrete nature of the underlying distributions.
In the survey of 52 RCTs published in four leading medical journals, 12 (23%) papers used statistical methods to compare discrete numerical variables between groups. All these 12 papers reported pvalues, but only two papers reported effect measures and confidence intervals. Seven papers used nonparametric methods, three papers used parametric methods, one paper used negative binomial regression, one paper stated that the CochranMantelHaenszel test was used, and one paper used either the twosample T test or the WilcoxonMannWhitney test.
The literature on statistical methods for analyzing discrete numerical variables is sparse. Newcombe [6] compares eight confidence interval methods for the mean of a single variable on the scale {0, 1, 2}, but does not consider comparisons of two independent variables. For comparing two continuous variables, on the other hand, a large body of literature exists. A relevant study for the hypothesis tests under investigation in this paper is Fagerland and Sandvik [7]. Confidence intervals for the difference in means of two independent continuous variables are considered in Zhou and Dinh [5] and Wilcox [[8], chapter 5]. Ordered categorical data is the topic of many papers, see for example Ryu and Agresti [9].
Returning to our survey, we counted nine different methods of presenting discrete numerical variables. The most common methods were to tabulate the data using categories such as {0, 13, 4+}, present the group means and standard deviations, or present medians and interquartile ranges (IQRs). Other methods included various combinations of means, medians, ranges, IQRs, and confidence intervals. Only two (8%) of 24 papers reported complete noncategorized data.
There is thus lack of a consensus on how to compare and present discrete numerical variables. In this paper, we shall investigate the performance of standard methods for continuous data applied to discrete numerical variables with outcomes such as {0, 1, 2, 3}. We assume that we are faced with samples from two independent random variables of equal type but with possibly different distributions. We shall further assume that we do not have a composite upper (or lower) limit, such as {0, 1, 2, 3, 4+}, where 4+ indicates outcomes with four or more events. The aim of this paper is to establish strong empirical evidence for recommending a suitable effect measure, methods for hypothesis testing and confidence intervals, and overall manner of presentation.
Methods
Effect measure
When we are dealing with two independent continuous variables, we are usually interested in estimating (and making inference about) the difference between some measure of the central tendencies. For symmetric distributions, most measures of central tendency, such as the arithmetic mean and the median, are equal. However, when distributions are skewed, different measures can vary substantially. The mean can be unduly influenced by outliers and may be a poor representation of the typical value. Choosing an appropriate measure of central tendency can then be quite difficult, particularly because software to analyze the optimal effect measure may not be readily available.
Fortunately, we seldom have the same problem with discrete numerical variables. When the variables have both lower and upper bounds, and when the range of possible values is quite limited, there will be no outliers or extreme values, at least not in the mathematical sense. There is thus no obvious added value of using, for example, the median or a trimmed mean as the measure of central tendency. Moreover, the median of discrete numerical variables often has a small number of possible valuesfive for a threevalued scalewhich makes it an imprecise measure of central tendency and thus unsuitable for demonstrating less than large differences between the groups.
As long as the mean of the variable of interest makes sense for the subject matter, we consider it to be a suitable measure of central tendency and that the difference between the two means is an appropriate effect measure. Note that using the mean is only appropriate for outcome scales without composite limits. If scales such as {0, 1, 2, 3+} is used, where 3+ indicates outcomes with three or more events, the estimated group means may underestimate the true means. The resulting estimate of the difference between the means may then be difficult to interpret. When using methods for continuous data, we strongly recommend against using composite limits.
Another appropriate effect measure for comparing two independent groups is the relative effect, p = Pr(X < Y ), where X and Y are random samples from the two groups. The relative effect is the probability that a random sample from one group is less than a random sample from the other group. If the groups are identically distributed, p = 1/2. Several rankbased methods, such as the WilcoxonMannWhitney test, is based on p or its generalization to tied values, p = Pr(X < Y ) + 0.5. Pr(X = Y ).
As an effect measure, the relative effect has the disadvantage that it is less specific than the difference between the means, and thereby more difficult to interpret. The relative effect can be a good alternative in situations where the mean is a poor estimate of central tendency. For discrete numerical variables with few possible values, the difference between the means is our preferred effect measure.
Simulation study of hypothesis tests
For the main comparison of hypothesis tests, we consider four hypothesis tests: the twosample T test, the modified T test for unequal variances (the Welch U test), the WilcoxonMannWhitney (WMW) test with adjustment for ties, and the BrunnerMunzel generalized WMW test. Details of the test statistics and their distributions can be found in Additional file 1.
Expected values for the six distributions in the simulation study.
Outcome scale  

Distribution  {0, 1, 2}  {0, 1, 2, 3}  {0, 1, 2, 3, 4}  {0, 1, 2, 3, 4, 5} 
Uniform  1.0  1.5  2.0  2.5 
Normal  1.0  1.5  2.0  2.5 
Ushaped  1.0  1.5  2.0  2.5 
Linear trend  0.63  1.0  1.25  1.625 
Step  0.8 & 0.6*  0.9  1.1  1.6 
Skewed  *  0.6  0.75  0.88 
Summary of the simulation setup (hypothesis tests).
Hypothesis tests  T: twosample T test 
U: Welch U test (T test for unequal variances)  
WMW: WilcoxonMannWhitney test  
BM: BrunnerMunzel test  
Outcome scales  {0, 1, 2}, {0, 1, 2, 3}, {0, 1, 2, 3, 4}, {0, 1, 2, 3, 4, 5} 
Distributions  Uniform, normal, ushaped, linear trend, step, skewed 
Sample sizes (m, n)  (10,10), (25,25), (50,50), (100,100), (25,10), (50,10), (100,50), (100,25), (100,10) 
Nominal sig.level  5% 
Replications  100 000 
Programming language  Matlab [10] with the Statistics Toolbox 
It has been suggested that a permutation test based on the BrunnerMunzel test statistic is appropriate for comparing smallsample discrete data [11]. We assessed this testusing 10 000 random permutations for each calculated testin a small separate simulation study. Only the sample sizes m = n = 10, the nine combinations of distributions with equal expected values, and 10 000 replications were used.
Simulation study of confidence intervals
The variance estimates used in the T and the Welch U tests are frequently used for the computation of confidence intervals. We refer to those intervals as the T confidence interval and the Welch U confidence interval. These are reported in most general purpose statistical software packages. It is clear from the results of the simulation study of hypothesis tests (see Results section) that the variance estimate for the T test is inaccurate for most situations where the sample sizes are unequal. As such, we include the Welch U confidence interval, but not the T confidence interval, in our investigation. In like manner, we do not consider nonparametric confidence intervals based on the WMW statistic because the WMW test performed poorly in the simulation study of hypothesis tests.
As alternatives to the Welch U confidence interval, we consider two simple bootstrap intervals: the percentile bootstrap and the bootstrapt [12], both with 2000 samples.
Summary of the simulation setup (confidence intervals).
Confidence intervals  U: the Welch U (T adjusted for unequal variances) 
PB: percentile bootstrap (2000 samples)  
Bt: bootstrapt (2000 samples)  
Outcome scales  {0, 1, 2}, {0, 1, 2, 3,}, {0, 1, 2, 3, 4}, {0, 1, 2, 3, 4, 5} 
Distributions  Uniform, normal, ushaped, linear trend, step, skewed 
Sample sizes (m, n)  (10, 10), (50, 50), (25, 10), (100, 25) 
Nominal confidence level  95% 
Replications  10 000 
Programming language  Matlab [10] with the Statistics Toolbox 
Results
Hypothesis tests
For each combination of outcome scale, sample sizes, and distributions, the rejection rates of the tests were recorded. When the expected values of the two distributions were equal, the rejection rates estimated the true significance level of the tests for the hypothesis of equal means. For distributions with unequal expected values, the rejection rates estimated the power of the tests to detect departures from equality of means.
For the assessment of true significance levels, we defined robustness criteria. If, for a given setting, the estimated true significance level of one of the tests deviated less than 10% from the nominal level, the test was defined as 10% robust. Similarly, if the estimated true significance level deviated less than 20% from the nominal level, the test was defined as 20% robust. A test with true significance levels that deviated more than 20% from the nominal level was defined as nonrobust. For a nominal significance level of 5%, the three robustness categories were

10% robust: 4.5 ≤ p ≤ 5.5

20% robust: 4.0 ≤ p ≤ 5.0

Nonrobust: p < 4.0 or p > 6.0
where p denotes the estimated true significance level. These robustness criteria have been used previously [7, 13]. We refer to Bradley [14] for a general discussion of robustness criteria.
We present the full results of the main simulation study in Additional file 2: Web Tables 338. Table cells are colored green, yellow, and red to indicate 10% robustness, 20% robustness, and nonrobustness, respectively.
Simulation results (hypothesis tests) for the outcome scale {0, 1, 2}.
Mean deviation from 5%  Relative power (%)  

Sample size  T  U  WMW  BM  T  U  WMW  BM 
10, 10  0.19  0.30  0.60  0.75  90.2  88.4  86.7  100.0 
25, 25  0.10  0.08  0.33  0.23  96.2  96.1  99.1  100.0 
50, 50  0.04  0.05  0.23  0.11  97.5  97.4  100.0  99.7 
100, 100  0.07  0.07  0.27  0.09  98.3  98.3  100.0  99.7 
25, 10  1.27  0.29  1.24  0.42  97.9  94.1  96.7  100.0 
50, 10  1.85  0.35  1.90  0.50  98.2  91.2  100.0  98.3 
100, 50  0.87  0.06  0.89  0.08  98.0  97.4  100.0  99.2 
100, 25  1.63  0.11  1.64  0.17  97.1  93.7  100.0  97.3 
100, 10  2.25  0.36  2.27  0.51  97.1  86.9  100.0  94.3 
Simulation results (hypothesis tests) for the outcome scale {0, 1, 2, 3}.
Mean deviation from 5%  Relative power (%)  

Sample size  T  U  WMW  BM  T  U  WMW  BM 
10, 10  0.24  0.28  0.55  0.59  100.0  96.8  89.9  98.4 
25, 25  0.06  0.06  0.35  0.19  100.0  99.8  94.8  95.6 
50, 50  0.06  0.06  0.33  0.11  100.0  99.9  96.1  95.8 
100, 100  0.07  0.07  0.33  0.10  100.0  100.0  97.7  97.3 
25, 10  1.35  0.27  1.34  0.44  100.0  95.4  92.7  94.6 
50, 10  2.05  0.39  2.00  0.51  100.0  91.5  93.4  91.3 
100, 50  1.00  0.04  1.03  0.06  100.0  99.7  96.7  96.1 
100, 25  1.78  0.09  1.77  0.12  100.0  97.6  95.6  94.1 
100, 10  2.55  0.41  2.42  0.50  100.0  88.4  93.6  89.0 
For the permutation test, the rejection rates for the outcome scale {0, 1, 2} ranged from 6.5% to 9.8%. The mean deviation from 5% was 2.89. When the other outcome scales were used, the rejection rates decreased but were still quite high and greater than those of the other tests.
Columns 69 display the relative power of the four main tests. The relative power is calculated by adding the per cent rejection rates for all combinations of distributions with unequal expected values and using the largest sum as the reference value. The greatest power is marked with bold type and the lowest power is marked with italic type.
The proportions of true significance levels that were nonrobustaveraged over all outcome scales, sample sizes, and distribution combinationswere 31% (T), 6.8% (U), 34% (WMW), and 1.8% (BM).
Confidence intervals
For each calculated confidence interval, we note three items: (i) does the interval contain the true difference between the means? (ii) the length of the interval; (iii) does the confidence limits extend beyond the maximum possible difference for the scale? For example, when using the outcome scale {0, 1, 2}, the maximum possible difference between the means is ±2.
The first item is used to estimate the coverage probability of the confidence intervals. The coverage probability should be close the nominal confidence level of 95%. If two or more confidence intervals have similar coverage probabilities, we can compare the intervals' lengths. Note that an interval with a low coverage probability can be expected to be shorter than an interval with a coverage probability close to the nominal level. From the third item, we compute the overshoot ratethe rate at which the intervals give nonsensical results.
The full results of the simulation study are given in Additional file 2: Web Tables 3954. As before, green table cells indicate 10% robustness (94.5 ≤ c ≤ 95.5), yellow table cells indicate 20% robustness (94.0 ≤ c ≤ 96.0), and red cells indicate nonrobustness (c < 94.0 or c > 96.0), where c denotes the estimated coverage probability.
Simulation results (confidence intervals).
Mean coverage prob.  Mean interval length  

Sample size  U  PB  Bt  U  PB  Bt 
Outcome scale {0, 1, 2}  
10, 10  95.2  93.2  95.8  1.45  1.27  1.49 
50, 50  94.8  94.7  94.8  0.61  0.60  0.61 
25, 10  94.3  92.5  94.1  1.21  1.07  1.19 
100, 25  95.0  94.3  94.6  0.70  0.66  0.69 
Outcome scale {0, 1, 2, 3}  
10, 10  95.0  93.0  95.8  1.95  1.71  2.01 
50, 50  94.8  94.6  94.8  0.82  0.80  0.82 
25, 10  94.5  92.5  94.3  1.63  1.44  1.61 
100, 25  94.8  94.0  94.4  0.94  0.89  0.92 
Outcome scale {0, 1, 2, 3, 4}  
10, 10  94.9  92.7  95.8  2.36  2.05  2.43 
50, 50  94.8  94.5  94.8  0.99  0.97  0.99 
25, 10  94.5  92.5  94.5  1.96  1.73  1.95 
100, 25  94.7  94.1  94.4  1.13  1.07  1.11 
Outcome scale {0, 1, 2, 3, 4, 5}  
10, 10  94.9  92.5  95.8  2.84  2.47  2.94 
50, 50  94.8  94.5  94.8  1.20  1.17  1.20 
25, 10  94.5  92.5  94.5  2.37  2.09  2.35 
100, 25  94.7  94.0  94.4  1.35  1.29  1.34 
The proportions of confidence intervals that were nonrobustaveraged over all outcome scales, sample sizes, and distribution combinationswere 6.6% (U), 60% (PB), and 18% (Bt).
The overshoot rate was zero for all intervals for all settings.
Recommendations: hypothesis test
For the outcome scale {0, 1, 2}, both the Welch U test and the BrunnerMunzel test had true significance levels that were close to the nominal level, although the BrunnerMunzel test did not perform well for the smallest sample size combination (m = n = 10). Among the two tests, the BrunnerMunzel test had superior power. The WMW test had true significance levels close to the nominal level when both samples were drawn from identical distributions. However, it was severely nonrobust for unequal distributions, particularly when the sample sizes were unequal. In these cases, the true significance level of the WMW test was sometimes above the nominal level and sometimes below the nominal level. The WMW test is thus not a reliable test of equality of distributions as it would have poor power in many situations. The ordinary T test performed similarly to the WMW test and neither test can be recommended. Nor can we recommend the NeubertBrunner permutation test, which performed poorly for m = n = 10. Instead, we recommend the BrunnerMunzel test, expect for small sample sizes where the Welch U test is our first choice.
The results from using the outcome scales {0, 1, 2, 3}, {0, 1, 2, 3, 4}, and {0, 1, 2, 3, 4, 5} were similar and are considered together. The T and WMW tests were usually robust when distributions were equal, but for unequal distributions, the nonrobustness of both tests increased with increasing number of outcome values. Of all tests considered, the T test had the greatest power. The Welch U test had superior robustness properties and performed well both for equal and unequal sample sizes, except for some cases where the sample size difference was large. Its power was often quite close to that of the T test. The BrunnerMunzel test performed generally well, but were usually slightly inferior to the Welch U test. We recommend the Welch U test and note that the BrunnerMunzel test can be a useful alternative. The T test and the WMW test are not recommended.
Recommendations: confidence intervals
The results for all three confidence interval methods were consistent over all the outcome scales. The coverage probability for the percentile bootstrap interval was considerably below the nominal level for most situations, and as such, we cannot recommend its use. The bootstrapt interval performed well when both sample sizes were 50. However, for the other sample size combinations, the coverage probability often deviated markedly, and in both directions, from the nominal level. In general, the Welch U interval had coverage probabilities close to the nominal level, although some distribution combinations produced coverage probabilities in the range 9394% when the sample sizes were unequal. The interval lengths of the Welch U and the bootstrapt intervals were similar.
Overall, the Welch U confidence interval performed better than the two bootstrap intervals and we recommend its use.
Recommendations: presentation of results
Reporting guidelines recommendand many journals now requirethat the principal analyses of a study are presented with the three key statistical items: point estimate, confidence interval, and pvalue [15, 16]. No exception should be made for discrete numerical data. In addition, given the discrete nature of the data, a 2 × g table (where g is the number of outcome values) representing the entire body of data can easily be presented, at least when the number of outcome values is small. This will show the distribution of data across the possible outcome values and allow readers to perform alternative analyses. Unfortunately, such reporting is rarely done in practice. The usual method of presentation is to report the group means or medians with either the standard deviations or the interquartile ranges. In the next section, we consider data from two clinical trials and illustrate how discrete numerical data can be analyzed and presented.
Clinical example: postcard intervention to reduce repetition of deliberate self poisoning
In a randomized controlled trial of patients hospitalized for deliberate self poisoning, Carter et al. [17] randomized 378 patients to an intervention group and 394 patients to a control group. All patients received standard treatment. In addition, the patients in the intervention group received eight postcards over 12 months. The main outcome measures were the proportion of patients with one or more repeat episodes of deliberate self poisoning and the number of episodes of deliberate self poisoning per patient during 12 months. The latter outcome measure is discrete numerical. As the maximum number of observed episodes per patient was four, the outcome scale was {0, 1, 2, 3, 4}.
Number of repeat episodes of deliberate self poisoning in men.
Number of repetitions  

0  1  2  3  4  mean (std)  
Control group (n = 102)  86  13  2  0  1  0.21 (0.57) 
Postcard group (n = 145)  125  13  10  2  0  0.20 (0.56) 
The authors of this trial used negative binomial regression to compare the risk of repeat episodes in the two groups. They found that the incidence risk ratio for the postcard group was 0.97 (95% confidence interval 0.48 to 1.98, p = 0.94)a similar result to the one above. The negative binomial distribution is based on a sequence of Bernoulli trials where the probability of an event (an episode of self poisoning) is constant. We are not convinced that it is appropriate to treat the number of repeat episodes of deliberate self poisoning as a sequence of event/nonevent trials. Furthermore, it does not seem likely that the probability of an episode of self poisoning is independent of the number of previous episodes.
Clinical example: intensive versus standard asthma education program
Does an intensive asthma education program reduce the number of visits to the emergency department and the number of hospitalizations for asthmatic children? Ng et al. [18] examined this issue by randomizing 100 children with an acute attack of asthma to either an intensive asthma education program (n = 55) or a standard asthma education program (n = 45).
Number of visits to the emergency department.
Number of visits  

0  1  2  3  4  mean (std)  
Standard program (n = 45)  19  10  7  3  6  1.27 (1.42) 
Intensive program (n = 55)  39  8  8  0  0  0.44 (0.74) 
Finally, we note a small discrepancy between our calculations and the results reported in Ng et al. [18]. No effect measure and confidence interval were presented in that paper, but the pvalue was given as 0.004 with either the T test or the WilcoxonMannWhitney test. We get p = 0.0003 with the ordinary T test, p = 0.0007 with the Welch U test, and p = 0.001 with the WilcoxonMannWhitney test.
Discusssion
We have considered how to compare two independent discrete numerical variables, a problem for which the difference between the two means is a suitable effect measure. Through two simulation studies, we find that the Welch U test and confidence interval can be recommended for statistical inference. The BrunnerMunzel test can also be recommendedexcept for small sample sizeshowever, if it is used in conjunction with the Welch U confidence interval, consistency between the test and confidence interval is not guaranteed. We prefer a unified approach, where test and confidence interval is based on similar principles. We further recommend that a table summarizing all the data is presented, at least for primary outcome variables.
We are not aware of any other paper that explicitly deal with the problem of comparing two independent discrete numerical variables by using statistical methods for continuous data. Our small survey illustrates that comparisons and presentations of such variables are performed in various fashions in the medical research literature. Few reported effect measures and confidence intervals, and few presented complete data. Nonparametric methodswhich were outperformed by their parametric counterparts in our studywere the most commonly used statistical methods. As such, this paper provides a necessary justification for using standard parametric methods for continuous data when comparing discrete numerical variables.
If we compare the recommendations in this paper with the results from studies of two continuous variables, there is some, but not complete, agreement. For approximately normal distributed variables, the twosample T test and confidence interval are well known to be the optimal methods for comparing the means. It is for nonnormal data, and especially skewed data, that alternative methods might be preferable. In a study of hypothesis tests, Fagerland and Sandvik [7] found that no test can be recommended for all situations, although the Welch U test performed best overall. They recommend that the selection of test is based on a thorough investigation of distribution properties. Zhou and Dinh [5] compared the ordinary T interval, the bootstrapt interval, the biascorrected and accelerated interval, and three intervals based on transformation of the tstatistic. They found that the bootstrapt interval gave consistent and best coverage, and that two of the transformation intervals were better than the ordinary T interval. The most noticeable difference between these two studies and the present one is the performance of the bootstrapt interval. In our study, the bootstrapt interval performed poorly for small sample sizes and for unequal sample sizes compared with the Welch U interval. This difference in performance may be due to the fact that Zhou and Dinh used continuous distributions, whereas we used discrete distributions. Unfortunately, Zhou and Dinh [5] did not include the Welch U interval in their simulation study, thus a comparison of the Welch U and the bootstrapt intervals for continuous distributions is not available. Regarding the poor performance of the WMW test, Lehmann [[19], Section 1.4 and p.60] notes some concerns both for the exact and the asymptotic WMW test in the presence of many ties.
Sometimes, it might be of interest to compare other aspects of the variables besides the means. In the example of intensive versus standard asthma education program, for example, not only the means but also the standard deviations are quite different (Table 8). The KolmogorovSmirnov test is commonly used to test the hypothesis that two variables have identical distributions. In the presence of many tiesas is the case with discrete numerical dataNeuhäuser [20] suggests a permutation test based on the BaumgartnerWeißSchindler statistic, and shows that this test is superior to the KolmogorovSmirnov test and five other tests for the hypothesis of equal distributions.
Our two simulation studies were limited to the outcome scales {0, 1, 2}, {0, 1, 2, 3}, {0, 1, 2, 3, 4}, and {0, 1, 2, 3, 4, 5}. Nevertheless, we extend the recommendations to wider discrete numerical outcome scalesprovided they have both upper and lower limitswith some confidence. This is due to two reasons: (i) the results from the simulation studies were quite similar for the outcome scales {0, 1, 2, 3}, {0, 1, 2, 3, 4}, and {0, 1, 2, 3, 4, 5}; (ii) as long as the variables have lower and upper bounds that are not too far apart, the mean will be an appropriate measure of central tendency. We thereby expect that the methods under investigation in this study will perform similarly on variables with outcomes such as {0, 1,...., 10}.
One benefit of being able to use simple parametric tests and confidence intervals for the comparison of two samples is that there is a natural way of generalizing the approach to situations with more than two samples and to the regression setting. It would be useful to perform a study to assess the performance of linear regression models with discrete numerical dependent variables. Based on the results from this study, we are optimistic about the prospects from such an investigation.
Conclusions
In the medical research literature, discrete numerical variablesusually reporting the number of events per individualare common. Until now, no studies has assessed the performance of parametric methods for comparing such variables. In our study, the Welch U test and confidence interval outperformed the WilcoxonMannWhitney test and two simple bootstrap intervals. We encourage more frequent use of parametric methods for comparing discrete numerical variables.
Declarations
Authors’ Affiliations
References
 Altman DG: Practical Statistics For Medical Research. 1991, Boca Raton, FL: Chapman & Hall/CRCGoogle Scholar
 Strömberg U: Collapsing ordered outcome categories: a note of concern. Am J Epidemiol. 1996, 144: 421424.View ArticlePubMedGoogle Scholar
 Sankey SS, Weissfeld LA: A study of the effect of dichotomizing ordinal data upon modeling. Commun StatSimul C. 1998, 27: 871887. 10.1080/03610919808813515.View ArticleGoogle Scholar
 Skovlund E, Fenstad GU: Should we always choose a nonparametric test when comparing two apparently nonnormal distributions?. J Clin Epidemiol. 2001, 54: 8692. 10.1016/S08954356(00)00264X.View ArticlePubMedGoogle Scholar
 Zhou XH, Dinh P: Nonparametric confidence intervals for the one and twosample problems. Biostatistics. 2005, 6: 187200. 10.1093/biostatistics/kxi002.View ArticlePubMedGoogle Scholar
 Newcombe RG: Confidence intervals for the mean of a variable taking the values 0, 1 and 2. Stat Med. 2003, 22: 27372750. 10.1002/sim.1479.View ArticlePubMedGoogle Scholar
 Fagerland MW, Sandvik L: Performance of five twosample location tests for skewed distributions with unequal variances. Contemp Clin Trials. 2009, 30: 490496. 10.1016/j.cct.2009.06.007.View ArticlePubMedGoogle Scholar
 Wilcox RR: Introduction To Robust Estimation And Hypothesis Testing. 2005, San Diego, CA: Academic Press, 2Google Scholar
 Ryu E, Agresti A: Modeling and inference for an ordinal effect size measure. Stat Med. 2008, 27: 17031717. 10.1002/sim.3079.View ArticlePubMedGoogle Scholar
 Matlab 7. 2005, The MathWorks, Inc., Natick, MA
 Neubert K, Brunner E: A studentized permutation test for the nonparametric BehrensFisher problem. Comput Stat Data An. 2007, 51: 51925204. 10.1016/j.csda.2006.05.024.View ArticleGoogle Scholar
 Davison AC, Hinkley DV: Bootstrap Methods And Their Application. 1997, Cambridge, UK: Cambridge University PressView ArticleGoogle Scholar
 Fagerland MW, Sandvik L: The WilcoxonMannWhitney test under scrutiny. Stat Med. 2009, 28: 14871497. 10.1002/sim.3561.View ArticlePubMedGoogle Scholar
 Bradley JV: Robustness?. Brit J Math Stat Psy. 1978, 31: 14452.View ArticleGoogle Scholar
 Altman DG, Schulz KF, Moher D, et al: The revised CONSORT statement for reporting randomized trials: explanation and elaboration. Ann Intern Med. 2001, 134: 663694.View ArticlePubMedGoogle Scholar
 Vandenbroucke JP, von Elm E, Altman DG, et al: Strengthening the reporting of observational studies in epidemiology (STROBE): explanation and elaboration. PLoS Med. 2007, 4: 16281654. 10.1371/journal.pmed.0040297.View ArticleGoogle Scholar
 Carter GL, Clover K, Whyte IM, Dawson AH, D'Este C: Postcards from the EDge project: randomised controlled trial of an intervention using postcards to reduce repetition of hospital treated deliberate self poisoning. BMJ. 2005, 331: 80510.1136/bmj.38579.455266.E0.View ArticlePubMedPubMed CentralGoogle Scholar
 Ng DKK, Chow PY, Lai WP, Chan KC, Chang BL, So HY: Effect of a structured asthma education program on hospitalized asthmatic children: a randomized controlled trial. Pediatr Int. 2006, 48: 158162. 10.1111/j.1442200X.2006.02185.x.View ArticlePubMedGoogle Scholar
 Lehmann EL: Nonparametrics. Statistical Methods Based On Ranks. 1975, Upper Saddle River, NJ: PrenticeHall, IncGoogle Scholar
 Neuhäuser M: A note on the exact test based on the BaumgartnerWeißSchindler statistic in the presence of ties. Comput Stat Data An. 2003, 42: 561568. 10.1016/S01679473(02)001214.View ArticleGoogle Scholar
 The prepublication history for this paper can be accessed here:http://www.biomedcentral.com/14712288/11/44/prepub
Prepublication history
Copyright
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.