Frequent Links
Mann–Whitney U test
In statistics, the Mann–Whitney U test (also called the Mann–Whitney–Wilcoxon (MWW), Wilcoxon ranksum test (WRS), or Wilcoxon–Mann–Whitney test) is a nonparametric test of the null hypothesis that two samples come from the same population against an alternative hypothesis, especially that a particular population tends to have larger values than the other.
It has greater efficiency than the ttest on nonnormal distributions, such as a mixture of normal distributions, and it is nearly as efficient as the ttest on normal distributions.
The Wilcoxon ranksum test is not the same as the Wilcoxon signedrank test, although both are nonparametric and involve summation of ranks.
Contents
Assumptions and formal statement of hypotheses
Although Mann and Whitney^{[1]} developed the MWW test under the assumption of continuous responses with the alternative hypothesis being that one distribution is stochastically greater than the other, there are many other ways to formulate the null and alternative hypotheses such that the MWW test will give a valid test.^{[2]}
A very general formulation is to assume that:
 All the observations from both groups are independent of each other,
 The responses are ordinal (i.e. one can at least say, of any two observations, which is the greater),
 The distributions of both groups are equal under the null hypothesis, so that the probability of an observation from one population (X) exceeding an observation from the second population (Y) equals the probability of an observation from Y exceeding an observation from X. That is, there is a symmetry between populations with respect to probability of random drawing of a larger observation.
 Under the alternative hypothesis, the probability of an observation from one population (X) exceeding an observation from the second population (Y) (after exclusion of ties) is not equal to 0.5. The alternative may also be stated in terms of a onesided test, for example: P(X > Y) + 0.5 P(X = Y) > 0.5.
Under more strict assumptions than those above, e.g., if the responses are assumed to be continuous and the alternative is restricted to a shift in location (i.e. F_{1}(x) = F_{2}(x + δ)), we can interpret a significant MWW test as showing a difference in medians. Under this location shift assumption, we can also interpret the MWW as assessing whether the Hodges–Lehmann estimate of the difference in central tendency between the two populations differs from zero. The Hodges–Lehmann estimate for this twosample problem is the median of all possible differences between an observation in the first sample and an observation in the second sample.
Calculations
The test involves the calculation of a statistic, usually called U, whose distribution under the null hypothesis is known. In the case of small samples, the distribution is tabulated, but for sample sizes above ~20 approximation using the normal distribution is fairly good. Some books tabulate statistics equivalent to U, such as the sum of ranks in one of the samples, rather than U itself.
The U test is included in most modern statistical packages. It is also easily calculated by hand, especially for small samples. There are two ways of doing this.
Method one:
For comparing two small sets of observations, a direct method is quick, and gives insight into the meaning of the U statistic, which corresponds to the number of wins out of all pairwise contests (see the tortoise and hare example under Examples below). For each observation in one set, count the number of times it wins over any observations in the other set (the other value loses if it is larger). Count 0.5 for any ties. The sum of wins and ties is U for the first set. U for the other set is the converse.
Method two:
For larger samples:
 Assign numeric ranks to all the observations, beginning with 1 for the smallest value. Where there are groups of tied values, assign a rank equal to the midpoint of unadjusted rankings [e.g., the ranks of (3, 5, 5, 9) are (1, 2.5, 2.5, 4)].
 Now, add up the ranks for the observations which came from sample 1. The sum of ranks in sample 2 is now determinate, since the sum of all the ranks equals N(N + 1)/2 where N is the total number of observations.
 U is then given by:^{[3]}
 <math>U_1=R_1  {n_1(n_1+1) \over 2} \,\!</math>
 where n_{1} is the sample size for sample 1, and R_{1} is the sum of the ranks in sample 1.
 Note that it doesn't matter which of the two samples is considered sample 1. An equally valid formula for U is
 <math>U_2= R_2  {n_2(n_2+1) \over 2} \,\!</math>
 The smaller value of U_{1} and U_{2} is the one used when consulting significance tables. The sum of the two values is given by
 <math>U_1 + U_2 = R_1  {n_1(n_1+1) \over 2} + R_2  {n_2(n_2+1) \over 2}. \,\!</math>
 The smaller value of U_{1} and U_{2} is the one used when consulting significance tables. The sum of the two values is given by
 Knowing that R_{1} + R_{2} = N(N + 1)/2 and N = n_{1} + n_{2} , and doing some algebra, we find that the sum is
 <math>U_1 + U_2 = n_1n_2. \,\!</math>
 Knowing that R_{1} + R_{2} = N(N + 1)/2 and N = n_{1} + n_{2} , and doing some algebra, we find that the sum is
Properties
The maximum value of U is the product of the sample sizes for the two samples. In such a case, the "other" U would be 0.
Examples
Illustration of calculation methods
Suppose that Aesop is dissatisfied with his classic experiment in which one tortoise was found to beat one hare in a race, and decides to carry out a significance test to discover whether the results could be extended to tortoises and hares in general. He collects a sample of 6 tortoises and 6 hares, and makes them all run his race at once. The order in which they reach the finishing post (their rank order, from first to last crossing the finish line) is as follows, writing T for a tortoise and H for a hare:
 T H H H H H T T T T T H
What is the value of U?
 Using the direct method, we take each tortoise in turn, and count the number of hares it beats, getting 6, 1, 1, 1, 1, 1, which means that U = 11. Alternatively, we could take each hare in turn, and count the number of tortoises it beats. In this case, we get 5, 5, 5, 5, 5, 0, so U = 25. Note that the sum of these two values for U is 36, which is 6 × 6.
 Using the indirect method:
 rank the animals by the time they take to complete the course, so give the first animal home rank 12, the second rank 11, and so forth.
 the sum of the ranks achieved by the tortoises is 12 + 6 + 5 + 4 + 3 + 2 = 32.
 Therefore U = 32 − (6×7)/2 = 32 − 21 = 11 (same as method one).
 the sum of the ranks achieved by the hares is 11 + 10 + 9 + 8 + 7 + 1 = 46, leading to U = 46 − 21 = 25.
Illustration of object of test
A second example race illustrates the point that the Mann–Whitney does not test for inequality of medians, but rather for difference of distributions. Consider another hare and tortoise race, with 19 participants of each species, in which the outcomes are as follows, from first to last past the finishing post:
 H H H H H H H H H T T T T T T T T T T H H H H H H H H H H T T T T T T T T T
If we simply compared medians, we would conclude that the median time for tortoises is less than the median time for hares, because the median tortoise here comes in at position 19, and thus actually beats the median hare, which comes in at position 20. However, the value of U is 100 (using the quick method of calculation described above, we see that each of 10 tortoises beats each of 10 hares, so U = 10 × 10). Consulting tables, or using the approximation below, we find that this U value gives significant evidence that hares tend to have lower completion times than tortoises (p < 0.05, twotailed). Obviously these are extreme distributions that would be spotted easily, but in larger samples something similar could happen without it being so apparent. Notice that the problem here is not that the two distributions of ranks have different variances; they are mirror images of each other, so their variances are the same, but they have very different skewness.
Normal approximation
For large samples, U is approximately normally distributed. In that case, the standardized value
 <math>z = \frac{ U  m_U }{ \sigma_U }, \, </math>
where m_{U} and σ_{U} are the mean and standard deviation of U, is approximately a standard normal deviate whose significance can be checked in tables of the normal distribution. m_{U} and σ_{U} are given by
 <math>m_U = \frac{n_1 n_2}{2}, \, </math> and
 <math>\sigma_U=\sqrt{n_1 n_2 (n_1 + n_2+1) \over 12}. \, </math>
The formula for the standard deviation is more complicated in the presence of tied ranks; the full formula is given in the text books referenced below.^{[citation needed]} However, if the number of ties is small (and especially if there are no large tie bands) ties can be ignored when doing calculations by hand. The computer statistical packages will use the correctly adjusted formula as a matter of routine.
Note that since U_{1} + U_{2} = n_{1} n_{2}, the mean n_{1} n_{2}/2 used in the normal approximation is the mean of the two values of U. Therefore, the absolute value of the z statistic calculated will be same whichever value of U is used.
Effect sizes
It is a widely recommended practice for scientists to report an effect size for an inferential test.^{[4]}^{[5]}
Common language effect size
One method of reporting the effect size for the MannWhitney U test is with the common language effect size.^{[6]} As a sample statistic, the common language effect size is computed by forming all possible pairs between the two groups, then finding the proportion of pairs that support a hypothesis.^{[7]} To illustrate, in a study with a sample of ten hares and ten tortoises, the total number of pairs is ten times ten or 100 pairs of hares and tortoises. Suppose the results show that the hare ran faster than the tortoise in 90 of the 100 sample pairs; in that case, the sample common language effect size is 90%. This sample value is an unbiased estimator of the population value, so the sample suggests that the best estimate of the common language effect size in the population is 90%.^{[8]}
Rankbiserial correlation
A second method of reporting the effect size for the MannWhitney U test is with the rankbiserial correlation. Edward Cureton introduced and named the measure.^{[9]} Like other correlational measures, the rankbiserial correlation can range from minus one to plus one, with a value of zero indicating no relationship. Dave Kerby ^{[6]} introduced the simple difference formula to compute the rankbiserial correlation from the common language effect size: the correlation is the difference between the proportion of pairs that support the hypothesis minus the proportion that do not. Stated another way, the correlation is the difference between the common language effect size and its complement. For example, consider the example where hares run faster than tortoises in 90 of 100 pairs. The common language effect size is 90%, so the rankbiserial correlation is 90% minus 10%, and the rankbiserial r = .80.
Hans Wendt ^{[10]} described a formula to compute the rankbiserial from the MannWhitney U and the sample sizes of each group: r = 1 – (2U)/ (n1 * n2). This formula is useful when the data are not available, but when there is a published report, because U and the sample sizes are routinely reported. Using the example above with 90 pairs that favor the hares and 10 pairs that favor the tortoise, U is the smaller of the two, so U = 10. The Wendt formula is then r = 1  (2*10) / (10 * 10) = .80, which of course is the same result as with the Kerby simple difference formula.
Relation to other tests
Comparison to Student's ttest
The U test is more widely applicable than independent samples Student's ttest, and the question arises of which should be preferred.
 Ordinal data
 U remains the logical choice when the data are ordinal but not interval scaled, so that the spacing between adjacent values cannot be assumed to be constant.
 Robustness
 As it compares the sums of ranks,^{[11]} the Mann–Whitney test is less likely than the ttest to spuriously indicate significance because of the presence of outliers – i.e. Mann–Whitney is more robust.^{[clarification needed]}^{[citation needed]}
 Efficiency
 When normality holds, MWW has an (asymptotic) efficiency of <math>3/\pi</math> or about 0.95 when compared to the t test.^{[12]} For distributions sufficiently far from normal and for sufficiently large sample sizes, the MWW is considerably more efficient than the t.^{[13]}
Overall, the robustness makes the MWW more widely applicable than the t test, and for large samples from the normal distribution, the efficiency loss compared to the t test is only 5%, so one can recommend MWW as the default test for comparing interval or ordinal measurements with similar distributions.^{[citation needed]}
The relation between efficiency and power in concrete situations isn't trivial though. For small sample sizes one should investigate the power of the MWW vs t.
MWW will give very similar results to performing an ordinary parametric twosample t test on the rankings of the data.^{[14]}
Areaundercurve (AUC) statistic for ROC curves
The U statistic is equivalent to the area under the receiver operating characteristic curve that can be readily calculated.^{[15]}^{[16]}
 <math>AUC_1 = {U_1 \over n_1n_2}</math>
Because of its probabilistic form, the U statistic can be generalised to a measure of a classifier's separation power for more than two classes:^{[17]}
 <math>M = {1 \over c(c1)} \sum AUC_{k,l}</math>
Where c is the number of classes, and the <math>R_{k,l}</math> term of <math>AUC_{k,l}</math> considers only the ranking of the items belonging to classes k and l (i.e., items belonging to all other classes are ignored) according to the classifier's estimates of the probability of those items belonging to class k. <math>AUC_{k,k}</math> will always be zero but, unlike in the twoclass case, generally <math>AUC_{k,l} \ne AUC_{l,k}</math>, which is why the <math>M</math> measure sums over all (k, l) pairs, in effect using the average of <math>AUC_{k,l}</math> and <math>AUC_{l,k}</math>.
Different distributions
If one is only interested in stochastic ordering of the two populations (i.e., the concordance probability P(Y > X)), the U test can be used even if the shapes of the distributions are different. The concordance probability is exactly equal to the area under the receiver operating characteristic curve (ROC) that is often used in the context.^{[citation needed]}
Alternatives
If one desires a simple shift interpretation, the U test should not be used when the distributions of the two samples are very different, as it can give erroneously significant results.^{[citation needed]} In that situation, the unequal variances version of the t test is likely to give more reliable results, but only if normality holds.^{[citation needed]}
Alternatively, some authors (e.g. Conover^{[full citation needed]}) suggest transforming the data to ranks (if they are not already ranks) and then performing the t test on the transformed data, the version of the t test used depending on whether or not the population variances are suspected to be different. Rank transformations do not preserve variances, but variances are recomputed from samples after rank transformations.
The Brown–Forsythe test has been suggested as an appropriate nonparametric equivalent to the F test for equal variances.^{[citation needed]}
History
The statistic appeared in a 1914 article ^{[18]} by the German Gustav Deuchler (with a missing term in the variance).
As a onesample statistic, the signed rank was proposed by Frank Wilcoxon in 1945,^{[19]} with some discussion of a twosample variant for equal sample sizes, in a test of significance with a point nullhypothesis against its complementary alternative (that is, equal versus not equal).
A thorough analysis of the statistic, which included a recurrence allowing the computation of tail probabilities for arbitrary sample sizes and tables for sample sizes of eight or less appeared in the article by Henry Mann and his student Donald Ransom Whitney in 1947.^{[1]} This article discussed alternative hypotheses, including a stochastic ordering (where the cumulative distribution functions satisfied the pointwise inequality <math> F_X(t) < F_Y(t) </math>). This paper also computed the first four moments and established the limiting normality of the statistic under the null hypothesis, so establishing that it is asymptotically distributionfree.
Related test statistics
Kendall's τ
The U test is related to a number of other nonparametric statistical procedures. For example, it is equivalent to Kendall's τ correlation coefficient if one of the variables is binary (that is, it can only take two values).^{[citation needed]}
ρ statistic
A statistic called ρ that is linearly related to U and widely used in studies of categorization (discrimination learning involving concepts),^{[citation needed]} and elsewhere,^{[20]} is calculated by dividing U by its maximum value for the given sample sizes, which is simply n_{1} × n_{2}. ρ is thus a nonparametric measure of the overlap between two distributions; it can take values between 0 and 1, and it is an estimate of P(Y > X) + 0.5 P(Y = X), where X and Y are randomly chosen observations from the two distributions. Both extreme values represent complete separation of the distributions, while a ρ of 0.5 represents complete overlap. The usefulness of the ρ statistic can be seen in the case of the odd example used above, where two distributions that were significantly different on a Utest nonetheless had nearly identical medians: the ρ value in this case is approximately 0.723 in favour of the hares, correctly reflecting the fact that even though the median tortoise beat the median hare, the hares collectively did better than the tortoises collectively.^{[citation needed]}
Example statement of results
In reporting the results of a Mann–Whitney test, it is important to state:
 A measure of the central tendencies of the two groups (means or medians; since the Mann–Whitney is an ordinal test, medians are usually recommended)
 The value of U
 The sample sizes
 The significance level.
In practice some of this information may already have been supplied and common sense should be used in deciding whether to repeat it. A typical report might run,
 "Median latencies in groups E and C were 153 and 247 ms; the distributions in the two groups differed significantly (Mann–Whitney U = 10.5, n_{1} = n_{2} = 8, P < 0.05 twotailed)."
A statement that does full justice to the statistical status of the test might run,
 "Outcomes of the two treatments were compared using the Wilcoxon–Mann–Whitney twosample ranksum test. The treatment effect (difference between treatments) was quantified using the Hodges–Lehmann (HL) estimator, which is consistent with the Wilcoxon test.^{[21]} This estimator (HLΔ) is the median of all possible differences in outcomes between a subject in group B and a subject in group A. A nonparametric 0.95 confidence interval for HLΔ accompanies these estimates as does ρ, an estimate of the probability that a randomly chosen subject from population B has a higher weight than a randomly chosen subject from population A. The median [quartiles] weight for subjects on treatment A and B respectively are 147 [121, 177] and 151 [130, 180] kg. Treatment A decreased weight by HLΔ = 5 kg (0.95 CL [2, 9] kg, 2P = 0.02, ρ = 0.58)."
However it would be rare to find so extended a report in a document whose major topic was not statistical inference.
Implementations
In many software packages, the Mann–Whitney test (of the hypothesis of equal distributions against appropriate alternatives) has been poorly documented. Some packages incorrectly treat ties or fail to document asymptotic techniques (e.g., correction for continuity). A 2000 review discussed versions of the following packages:^{[22]}
 Matlab has ranksum within its Statistics Toolbox.
 R's statistics basepackage implements the test
wilcox.test
and in its COIN package.  SAS implements the test in its PROC NPAR1WAY procedure.
 Stata implements the test in its ranksum command.
 Python (programming language) has an implementation of this test provided by SciPy: [1]
See also
Notes
 ^ ^{a} ^{b} Mann, Henry B.; Whitney, Donald R. (1947). "On a Test of Whether one of Two Random Variables is Stochastically Larger than the Other". Annals of Mathematical Statistics 18 (1): 50–60. MR 22058. Zbl 0041.26103. doi:10.1214/aoms/1177730491.
 ^ Fay, Michael P.; Proschan, Michael A. (2010). "Wilcoxon–Mann–Whitney or ttest? On assumptions for hypothesis tests and multiple interpretations of decision rules". Statistics Surveys 4: 1–39. MR 2595125. PMC 2857732. PMID 20414472. doi:10.1214/09SS051.
 ^ Zar, Jerrold H. (1998). Biostatistical Analysis. New Jersey: Prentice Hall International, INC. p. 147. ISBN 0130823902.
 ^ Wilkinson, Leland; APA Task Force on Statistical Inference (1999). "Statistical methods in psychology journals: Guidelines and explanations". American Psychologist 54 (8): 594–604. doi:10.1037/0003066X.54.8.594.
 ^ Nakagawa, Shinichi; Cuthill, Innes C (2007). "Effect size, confidence interval and statistical significance: a practical guide for biologists". Biological Reviews Cambridge Philosophical Society 82 (4): 591–605. PMID 17944619. doi:10.1111/j.1469185X.2007.00027.x.
 ^ ^{a} ^{b} Kerby, D. S. (2014). The simple difference formula: An approach to teaching nonparametric correlation. Innovative Teaching, volume 3, article 1. doi:10.2466/11.IT.3.1. link to pdf
 ^ McGraw, K.O., & Wong, S.P. (1992). A common language effect size statistic. Psychological Bulletin, volume 111(2), pages 361365. doi:10.1037/00332909.111.2.361
 ^ Grissom RJ (1994). "Statistical analysis of ordinal categorical status after therapies". Journal of Consulting and Clinical Psychology 62 (2): 281–284. doi:10.1037/0022006X.62.2.281.
 ^ Cureton, E.E. (1956). Rankbiserial correlation. Psychometrika, volume 21(3), pages 287290. doi:10.1007/BF02289138
 ^ Wendt, H. W. (1972). Dealing with a common problem in social science: A simplified rankbiserial coefficient of correlation based on the U statistic. European Journal of Social Psychology, 2(4), 463–465. doi:10.1002/ejsp.2420020412
 ^ Motulsky, Harvey J.; Statistics Guide, San Diego, CA: GraphPad Software, 2007, p. 123
 ^ Lehamnn, Erich L.; Elements of Large Sample Theory, Springer, 1999, p. 176
 ^ Conover, William J.; Practical Nonparametric Statistics, John Wiley & Sons, 1980 (2nd Edition), pp. 225–226
 ^ Conover, William J.; Iman, Ronald L. (1981). "Rank Transformations as a Bridge Between Parametric and Nonparametric Statistics". The American Statistician 35 (3): 124–129. JSTOR 2683975. doi:10.2307/2683975.
 ^ Hanley, James A.; McNeil, Barbara J. (1982). "The Meaning and Use of the Area under a Receiver Operating (ROC) Curve Characteristic". Radiology 143 (1): 29–36. PMID 7063747.
 ^ Mason, Simon J.; Graham, Nicholas E. (2002). "Areas beneath the relative operating characteristics (ROC) and relative operating levels (ROL) curves: Statistical significance and interpretation" (PDF). Quarterly Journal of the Royal Meteorological Society (128): 2145–2166.
 ^ Hand, David J.; Till, Robert J. (2001). "A Simple Generalisation of the Area Under the ROC Curve for Multiple Class Classification Problems" (PDF). Machine Learning 45 (2): 171–186. doi:10.1023/A:1010920819831.
 ^ Kruskal, William H. (September 1957). "Historical Notes on the Wilcoxon Unpaired TwoSample Test". Journal of the American Statistical Association 52 (279): 356–360. doi:10.2307/2280906.
 ^ Wilcoxon, Frank (1945). "Individual comparisons by ranking methods". Biometrics Bulletin 1 (6): 80–83. JSTOR 3001968. doi:10.2307/3001968.
 ^ Herrnstein, Richard J.; Loveland, Donald H.; Cable, Cynthia (1976). "Natural Concepts in Pigeons". Journal of Experimental Psychology: Animal Behavior Processes 2: 285–302. doi:10.1037/00977403.2.4.285.
 ^ Myles Hollander and Douglas A. Wolfe (1999). Nonparametric Statistical Methods (2 ed.). WileyInterscience. ISBN 9780471190455.
 ^ Bergmann, Reinhard; Ludbrook, Will P. J. M.; Spooren (2000). "Different Outcomes of the WilcoxonMannWhitney Test from Different Statistics Packages". The American Statistician 54 (1): 72–77. JSTOR 2685616. doi:10.1080/00031305.2000.10474513.
References
 Hettmansperger, T. P.; McKean, J. W. (1998). Robust nonparametric statistical methods. Kendall's Library of Statistics 5 (First ed., rather than Taylor and Francis (2010) second ed.). London; New York: Edward Arnold; John Wiley and Sons, Inc. pp. xiv+467. ISBN 0340549378. MR 1604954.
 Corder, G. W.; Foreman, D. I. (2014). Nonparametric Statistics: A StepbyStep Approach. Wiley. ISBN 9781118840313.
 Hodges, J. L.; Lehmann, E. L. (1963). "Estimation of location based on ranks". Annals of Mathematical Statistics 34 (2): 598–611. JSTOR 2238406. MR 152070. Zbl 0203.21105. doi:10.1214/aoms/1177704172. PE euclid.aoms/1177704172.
 Lehmann, Erich L. (2006). Nonparametrics: Statistical methods based on ranks. With the special assistance of H. J. M. D'Abrera (Reprinting of 1988 revision of 1975 HoldenDay ed.). New York: Springer. pp. xvi+463. ISBN 9780387352121. MR 395032.
 Oja, Hannu (2010). Multivariate nonparametric methods with R: An approach based on spatial signs and ranks. Lecture Notes in Statistics 199. New York: Springer. pp. xiv+232. ISBN 9781441904676. MR 2598854. doi:10.1007/9781441904683.
 Sen, Pranab Kumar (December 1963). "On the estimation of relative potency in dilution(direct) assays by distributionfree methods". Biometrics 19 (4): 532–552. JSTOR 2527532. Zbl 0119.15604. doi:10.2307/2527532.
External links
 Table of critical values of U (pdf)
 Interactive calculator for U and its significance
