Non-parametric test

In statistics, the term non-parametric statistics refers to statistics that do not assume the data or population have any characteristic structure or parameters. For example, non-parametric statistics are suitable for examining the order in which runners complete a race, while parametric statistics would be more appropriate for looking at the actual race times (which may possess parameters such as a mean and standard deviation). In other words, the order (or "rank") of the values is used rather than the actual values themselves.


In statistics, the term non-parametric statistics has at least two different meanings:

  1. The first meaning of non-parametric covers techniques that do not rely on data belonging to any particular distribution. These include, among others:
    The following discussion is taken from Kendall's.[1]
    Statistical hypotheses concern the behavior of observable random variables.... For example, the hypothesis (a) that a normal distribution has a specified mean and variance is statistical; so is the hypothesis (b) that it has a given mean but unspecified variance; so is the hypothesis (c) that a distribution is of normal form with both mean and variance unspecified; finally, so is the hypothesis (d) that two unspecified continuous distributions are identical.

    It will have been noticed that in the examples (a) and (b) the distribution underlying the observations was taken to be of a certain form (the normal) and the hypothesis was concerned entirely with the value of one or both of its parameters. Such a hypothesis, for obvious reasons, is called parametric.

    Hypothesis (c) was of a different nature, as no parameter values are specified in the statement of the hypothesis; we might reasonable call such a hypothesis non-parametric. Hypothesis (d) is also non-parametric but, in addition, it does not even specify the underlying form of the distribution and may now be reasonably termed distribution-free. Notwithstanding these distinctions, the statistical literature now commonly applies the label "non-parametric" to test procedures that we have just termed "distribution-free", thereby losing a useful classification.

  2. The second meaning of non-parametric covers techniques that do not assume that the structure of a model is fixed. Typically, the model grows in size to accommodate the complexity of the data. In these techniques, individual variables are typically assumed to belong to parametric distributions, and assumptions about the types of connections among variables are also made. These techniques include, among others:
    • non-parametric regression, which refers to modeling where the structure of the relationship between variables is treated non-parametrically, but where nevertheless there may be parametric assumptions about the distribution of model residuals.
    • non-parametric hierarchical Bayesian models, such as models based on the Dirichlet process, which allow the number of latent variables to grow as necessary to fit the data, but where individual variables still follow parametric distributions and even the process controlling the rate of growth of latent variables follows a parametric distribution.

Applications and purpose

Non-parametric methods are widely used for studying populations that take on a ranked order (such as movie reviews receiving one to four stars). The use of non-parametric methods may be necessary when data have a ranking but no clear numerical interpretation, such as when assessing preferences. In terms of levels of measurement, non-parametric methods result in "ordinal" data.

As non-parametric methods make fewer assumptions, their applicability is much wider than the corresponding parametric methods. In particular, they may be applied in situations where less is known about the application in question. Also, due to the reliance on fewer assumptions, non-parametric methods are more robust.

Another justification for the use of non-parametric methods is simplicity. In certain cases, even when the use of parametric methods is justified, non-parametric methods may be easier to use. Due both to this simplicity and to their greater robustness, non-parametric methods are seen by some statisticians as leaving less room for improper use and misunderstanding.

The wider applicability and increased robustness of non-parametric tests comes at a cost: in cases where a parametric test would be appropriate, non-parametric tests have less power. In other words, a larger sample size can be required to draw conclusions with the same degree of confidence.

Non-parametric models

Non-parametric models differ from parametric models in that the model structure is not specified a priori but is instead determined from data. The term non-parametric is not meant to imply that such models completely lack parameters but that the number and nature of the parameters are flexible and not fixed in advance.


Non-parametric (or distribution-free) inferential statistical methods are mathematical procedures for statistical hypothesis testing which, unlike parametric statistics, make no assumptions about the probability distributions of the variables being assessed. The most frequently used tests include

  • Anderson–Darling test: tests whether a sample is drawn from a given distribution
  • Statistical Bootstrap Methods: estimates the accuracy/sampling distribution of a statistic
  • Cochran's Q: tests whether k treatments in randomized block designs with 0/1 outcomes have identical effects
  • Cohen's kappa: measures inter-rater agreement for categorical items
  • Friedman two-way analysis of variance by ranks: tests whether k treatments in randomized block designs have identical effects
  • Kaplan–Meier: estimates the survival function from lifetime data, modeling censoring
  • Kendall's tau: measures statistical dependence between two variables
  • Kendall's W: a measure between 0 and 1 of inter-rater agreement
  • Kolmogorov–Smirnov test: tests whether a sample is drawn from a given distribution, or whether two samples are drawn from the same distribution
  • Kruskal-Wallis one-way analysis of variance by ranks: tests whether >2 independent samples are drawn from the same distribution
  • Kuiper's test: tests whether a sample is drawn from a given distribution, sensitive to cyclic variations such as day of the week
  • Logrank Test: compares survival distributions of two right-skewed, censored samples
  • Mann–Whitney U or Wilcoxon rank sum test: tests whether two samples are drawn from the same distribution, as compared to a given alternative hypothesis
  • McNemar's test: tests whether, in 2 × 2 contingency tables with a dichotomous trait and matched pairs of subjects, row and column marginal frequencies are equal
  • median test: tests whether two samples are drawn from distributions with equal medians
  • Pitman's permutation test: a statistical significance test that yields exact p values by examining all possible rearrangements of labels
  • Rank products: detects differentially expressed genes in replicated microarray experiments
  • Siegel–Tukey test: tests for differences in scale between two groups
  • sign test: tests whether matched pair samples are drawn from distributions with equal medians
  • Spearman's rank correlation coefficient: measures statistical dependence between two variables using a monotonic function
  • Squared ranks test: tests equality of variances in two or more samples
  • Wald–Wolfowitz runs test: tests whether the elements of a sequence are mutually independent/random
  • Wilcoxon signed-rank test: tests whether matched pair samples are drawn from populations with different mean ranks

See also


General references

  • Bagdonavicius, V., Kruopis, J., Nikulin, M.S. (2011). "Non-parametric tests for complete data", ISTE&WILEY: London&Hoboken. ISBN 978-1-84821-269-5
  • Corder, G.W. & Foreman, D.I. (2009) Nonparametric Statistics for Non-Statisticians: A Step-by-Step Approach, Wiley ISBN 978-0-470-45461-9
  • Gibbons, Jean Dickinson and Chakraborti, Subhabrata (2003) Nonparametric Statistical Inference, 4th Ed. CRC ISBN 0-8247-4052-1
  • Wasserman, Larry (2007) All of nonparametric statistics, Springer. ISBN 0-387-25145-6
This article was sourced from Creative Commons Attribution-ShareAlike License; additional terms may apply. World Heritage Encyclopedia content is assembled from numerous content providers, Open Access Publishing, and in compliance with The Fair Access to Science and Technology Research Act (FASTR), Wikimedia Foundation, Inc., Public Library of Science, The Encyclopedia of Life, Open Book Publishers (OBP), PubMed, U.S. National Library of Medicine, National Center for Biotechnology Information, U.S. National Library of Medicine, National Institutes of Health (NIH), U.S. Department of Health & Human Services, and, which sources content from all federal, state, local, tribal, and territorial government publication portals (.gov, .mil, .edu). Funding for and content contributors is made possible from the U.S. Congress, E-Government Act of 2002.
Crowd sourced content that is contributed to World Heritage Encyclopedia is peer reviewed and edited by our editorial staff to ensure quality scholarly research articles.
By using this site, you agree to the Terms of Use and Privacy Policy. World Heritage Encyclopedia™ is a registered trademark of the World Public Library Association, a non-profit organization.