In probability theory and statistics, the cumulants κ_{n} of a probability distribution are a set of quantities that provide an alternative to the moments of the distribution. The moments determine the cumulants in the sense that any two probability distributions whose moments are identical will have identical cumulants as well, and similarly the cumulants determine the moments. In some cases theoretical treatments of problems in terms of cumulants are simpler than those using moments.
Just as for moments, where joint moments are used for collections of random variables, it is possible to define joint cumulants.
Contents

Definition 1

Alternative definition of the cumulant generating function 1.1

Uses in statistics 2

Cumulants of some discrete probability distributions 3

Cumulants of some continuous probability distributions 4

Some properties of the cumulant generating function 5

Some properties of cumulants 6

Invariance and equivariance 6.1

Homogeneity 6.2

Additivity 6.3

A negative result 6.4

Cumulants and moments 6.5

Cumulants and setpartitions 6.6

Cumulants and combinatorics 6.7

Joint cumulants 7

Conditional cumulants and the law of total cumulance 7.1

Relation to statistical physics 8

History 9

Cumulants in generalized settings 10

Formal cumulants 10.1

Bell numbers 10.2

Cumulants of a polynomial sequence of binomial type 10.3

Free cumulants 10.4

See also 11

References 12

External links 13
Definition
The cumulants κ_{n} of a random variable X are defined via the cumulantgenerating function K(t), which is the natural logarithm of the momentgenerating function:

K(t)=\log\mathbb{E}\!\left[e^{tX}\right].
The cumulants κ_{n} are obtained from a power series expansion of the cumulant generating function:

K(t)=\sum_{n=1}^\infty \kappa_{n} \frac{t^{n}}{n!}.
This expansion is a Maclaurin series, so that nth cumulant can be obtained by differentiating the above expansion n times and evaluating the result at zero:^{[1]}

\kappa_{n} = K^{(n)}(0).
If the momentgenerating function does not exist, the cumulants can be defined in terms of the relationship between cumulants and moments discussed later.
Alternative definition of the cumulant generating function
Some writers^{[2]}^{[3]} prefer to define the cumulantgenerating function as the natural logarithm of the characteristic function, which is sometimes also called the second characteristic function,^{[4]}^{[5]}

H(t)=\log\mathbb{E}\!\left[e^{i t X}\right]=\sum_{n=1}^\infty \kappa_n \frac{(it)^n}{n!}=\mu it  \sigma^2 \frac{ t^2}{2} + \cdots
An advantage of H(t)—in some sense the function K(t) evaluated for purely imaginary arguments—is that E(e^{itX}) is well defined for all real values of t even when E(e^{tX}) is not well defined for all real values of t, such as can occur when there is "too much" probability that X has a large magnitude. Although the function H(t) will be well defined, it nonetheless may mimic K(t) by not having a Maclaurin series beyond (or, rarely, even to) linear order in the argument t. Thus, many cumulants may still not be well defined. Nevertheless, even when H(t) does not have a long Maclaurin series, it can be used directly in analyzing and, particularly, adding random variables. Both the Cauchy distribution (also called the Lorentzian) and stable distribution (related to the Lévy distribution) are examples of distributions for which the powerseries expansions of the generating functions have only finitely many welldefined terms.
Uses in statistics
Working with cumulants can have an advantage over using moments because for statistically independent random variables X and Y,

\begin{align} K_{X+Y}(t) & =\log\mathbb{E}\!\left[e^{t(X+Y)}\right] \\ &= \log\left(\mathbb{E}\left[e^{tX}\right]\mathbb{E}\left[e^{tY}\right]\right) \\ & = \log\mathbb{E}\left[e^{tX}\right] + \log\mathbb{E}\left[e^{tY}\right] \\ &= K_X(t) + K_Y(t), \end{align}
so that each cumulant of a sum of independent random variables is the sum of the corresponding cumulants of the addends. A related result is that a cumulantgenerating function for a product of independent random variables (defined as a random sum of independent realizations) is the composition of the respective cumulantgenerating functions:^{[6]}

\begin{align} K_{XY}(t) &= \log \mathbb{E} \left[ e^{t \sum_{i=1}^{X} Y_i} \right] \\ &= K_X(K_Y(t)). \end{align}
A distribution with given cumulants κ_{n} can be approximated through an Edgeworth series.
Cumulants of some discrete probability distributions

The constant random variables X = μ. The cumulant generating function is K(t) =μt. The first cumulant is κ_{1} = K '(0) = μ and the other cumulants are zero, κ_{2} = κ_{3} = κ_{4} = ... = 0.

The Bernoulli distributions, (number of successes in one trial with probability p of success). The cumulant generating function is K(t) = log(1−p+pe^{t}). The first cumulants are κ_{1} = K '(0) = p and κ_{2} = K′′(0) = p·(1 − p). The cumulants satisfy a recursion formula


\kappa_{n+1}=p (1p) \frac{d\kappa_n}{dp}.

The geometric distributions, (number of failures before one success with probability p of success on each trial). The cumulant generating function is K(t) = log(p / (1 + (p−1)e^{t})). The first cumulants are κ_{1} = K′(0) = p^{−1} − 1, and κ_{2} = K′′(0) = κ_{1}p^{−1}. Substituting p = (μ + 1)^{−1} gives K(t) = −log(1 + μ(1−e^{t})) and κ_{1} = μ.

The Poisson distributions. The cumulant generating function is K(t) = μ(e^{t}−1). All cumulants are equal to the parameter: κ_{1} = κ_{2} = κ_{3} = ... = μ.

The binomial distributions, (number of successes in n independent trials with probability p of success on each trial). The special case n = 1 is a Bernoulli distribution. Every cumulant is just n times the corresponding cumulant of the corresponding Bernoulli distribution. The cumulant generating function is K(t) = n log(1−p+pe^{t}). The first cumulants are κ_{1} = K′(0) = np and κ_{2} = K′′(0) = κ_{1}(1 − p). Substituting p = μ·n^{−1} gives K '(t) = ((μ^{−1} − n^{−1})·e^{−t} + n^{−1})^{−1} and κ_{1} = μ. The limiting case n^{−1} = 0 is a Poisson distribution.

The negative binomial distributions, (number of failures before n successes with probability p of success on each trial). The special case n = 1 is a geometric distribution. Every cumulant is just n times the corresponding cumulant of the corresponding geometric distribution. The derivative of the cumulant generating function is K '(t) = n·((1−p)^{−1}·e^{−t}−1)^{−1}. The first cumulants are κ_{1} = K '(0) = n·(p^{−1}−1), and κ_{2} = K ' '(0) = κ_{1}·p^{−1}. Substituting p = (μ·n^{−1}+1)^{−1} gives K′(t) = ((μ^{−1} + n^{−1})e^{−t} − n^{−1})^{−1} and κ_{1} = μ. Comparing these formulas to those of the binomial distributions explains the name 'negative binomial distribution'. The limiting case n^{−1} = 0 is a Poisson distribution.
Introducing the variancetomean ratio

\varepsilon=\mu^{1}\sigma^2=\kappa_1^{1}\kappa_2,
the above probability distributions get a unified formula for the derivative of the cumulant generating function:

K'(t)=\mu\cdot(1+\varepsilon\cdot (e^{t}1))^{1}.
The second derivative is

K''(t)=g'(t)\cdot(1+e^t\cdot (\varepsilon^{1}1))^{1}
confirming that the first cumulant is κ_{1} = K′(0) = μ and the second cumulant is κ_{2} = K′′(0) = με. The constant random variables X = μ have ε = 0. The binomial distributions have ε = 1 − p so that 0 < ε < 1. The Poisson distributions have ε = 1. The negative binomial distributions have ε = p^{−1} so that ε > 1. Note the analogy to the classification of conic sections by eccentricity: circles ε = 0, ellipses 0 < ε < 1, parabolas ε = 1, hyperbolas ε > 1.
Cumulants of some continuous probability distributions

For the normal distribution with expected value μ and variance σ^{2}, the cumulant generating function is K(t) = μt + σ^{2}t^{2}/2. The first and second derivatives of the cumulant generating function are K '(t) = μ + σ^{2}·t and K"(t) = σ^{2}. The cumulants are κ_{1} = μ, κ_{2} = σ^{2}, and κ_{3} = κ_{4} = ... = 0. The special case σ^{2} = 0 is a constant random variable X = μ.
Some properties of the cumulant generating function
The cumulant generating function K(t), if it exists, is infinitely differentiable and convex, and passes through the origin. Its first derivative ranges monotonically in the open interval from the infimum to the supremum of the support of the probability distribution, and its second derivative is strictly positive everywhere it is defined, except for the degenerate distribution of a single point mass. The cumulantgenerating function exists if and only if the tails of the distribution are majorized by an exponential decay, that is, (see Big O notation,)

\exists c>0, F(x)=O(e^{cx}), x\to\infty; and

\exists d>0, 1F(x)=O(e^{dx}),x\to+\infty;
where F is the cumulative distribution function. The cumulantgenerating function will have vertical asymptote(s) at the infimum of such c, if such an infimum exists, and at the supremum of such d, if such a supremum exists, otherwise it will be defined for all real numbers.
If the support of a random variable X has finite upper or lower bounds, then its cumulantgenerating function y=K(t), if it exists, approaches asymptote(s) whose slope is equal to the supremum and/or infimum of the support,

y=(t+1)\inf \mathrm{supp}X\mu(X), and

y=(t1)\sup\mathrm{supp}X+\mu(X),
respectively, lying above both these lines everywhere. (The integrals

\int_{\infty}^0 \left[t\inf \mathrm{supp}XK'(t)\right]dt, \qquad \int_{\infty}^0 \left[t\inf \mathrm{supp}XK'(t)\right]dt
yield the yintercepts of these asymptotes, since K(0)=0.)
For a shift of the distribution by c, K_{X+c}(t)=K_X(t)+ct. For a degenerate point mass at c, the cgf is the straight line K_c(t)=ct, and more generally, K_{X+Y}=K_X+K_Y if and only if X and Y are independent and their cgfs exist; (subindependence and the existence of second moments sufficing to imply independence.^{[7]})
The natural exponential family of a distribution may be realized by shifting or translating K(t), and adjusting it vertically so that it always passes through the origin: if f is the pdf with cgf K(t)=\log M(t), and f\theta is its natural exponential family, then f(x\theta)=\frac1{M(\theta)}e^{\theta x} f(x), and K(t\theta)=K(t+\theta)K(\theta).
If K(t) is finite for a range t_{1} < Re(t) < t_{2} then if t_{1} < 0 < t_{2} then K(t) is analytic and infinitely differentiable for t_{1} < Re(t) < t_{2}. Moreover for t real and t_{1} < t < t_{2} K(t) is strictly convex, and K'(t) is strictly increasing.
Some properties of cumulants
Invariance and equivariance
The first cumulant is shiftequivariant; all of the others are shiftinvariant. This means that, if we denote by κ_{n}(X) the nth cumulant of the probability distribution of the random variable X, then for any constant c:

\kappa_1(X + c) = \kappa_1(X) + c ~ \text{ and}

\kappa_n(X + c) = \kappa_n(X) ~ \text{ for } ~ n \ge 2.
In other words, shifting a random variable (adding c) shifts the first cumulant (the mean) and doesn't affect any of the others.
Homogeneity
The nth cumulant is homogeneous of degree n, i.e. if c is any constant, then

\kappa_n(cX)=c^n\kappa_n(X). \,
Additivity
If X and Y are independent random variables then κ_{n}(X + Y) = κ_{n}(X) + κ_{n}(Y).
A negative result
Given the results for the cumulants of the normal distribution, it might be hoped to find families of distributions for which κ_{m} = κ_{m+1} = ... = 0 for some m > 3, with the lowerorder cumulants (orders 3 to m − 1) being nonzero. There are no such distributions.^{[8]} The underlying result here is that the cumulant generating function cannot be a finiteorder polynomial of degree greater than 2.
Cumulants and moments
The moment generating function is given by:

M(t) = 1+\sum_{n=1}^\infty \frac{\mu'_n t^n}{n!}=\exp\left(\sum_{n=1}^\infty \frac{\kappa_n t^n}{n!}\right) = \exp(K(t)).
So the cumulant generating function is the logarithm of the moment generating function

K(t) = \log M(t).
The first cumulant is the expected value; the second and third cumulants are respectively the second and third central moments (the second central moment is the variance); but the higher cumulants are neither moments nor central moments, but rather more complicated polynomial functions of the moments.
The moments can be recovered in terms of cumulants by evaluating the nth derivative of \exp(K(t)) at t = 0,

\mu'_n = M^{(n)}(0) = \frac{\mathrm{d}^n \exp (K(t))}{\mathrm{d}t^n}\Big_{t=0}.
Likewise, the cumulants can be recovered in terms of moments by evaluating the nth derivative of \log M(t) at t=0,

\kappa_n = K^{(n)}(0) = \frac{\mathrm{d}^n \log M(t)}{\mathrm{d}t^n}\Big_{t=0}.
The explicit expression for the nth moment in terms of the first n cumulants, and vice versa, can be obtained by using Faà di Bruno's formula for higher derivatives of composite functions. In general, we have

\mu'_n = \sum_{k=1}^n B_{n,k}(\kappa_1,\ldots,\kappa_{nk+1})

\kappa_n = \sum_{k=1}^n (1)^{k1} (k1)! B_{n,k}(\mu'_1,\ldots,\mu'_{nk+1}),
where B_{n,k} are incomplete (or partial) Bell polynomials.
In the like manner, if the mean is given by \mu, the central moment generating function is given by

C(t) = \mathbb{E}[e^{t(x\mu)}] = e^{\mu t} M(t) = \exp(K(t)  \mu t),
and the nth central moment is obtained in terms of cumulants as

\mu_n = C^{(n)}(0) = \frac{\mathrm{d}^n}{\mathrm{d}t^n} \exp (K(t)  \mu t)\Big_{t=0} = \sum_{k=1}^n B_{n,k}(0,\kappa_2,\ldots,\kappa_{nk+1}).
Also, for n>1, the nth cumulant in terms of the central moments is,

\kappa_n = K^{(n)}(0) = \frac{\mathrm{d}^n}{\mathrm{d}t^n} (\log C(t) + \mu t) \Big_{t=0} = \sum_{k=1}^n (1)^{k1} (k1)! B_{n,k}(0,\mu_2,\ldots,\mu_{nk+1}).
The nth moment μ′_{n} is an nthdegree polynomial in the first n cumulants. The first few expressions are:

\mu'_1=\kappa_1\,

\mu'_2=\kappa_2+\kappa_1^2\,

\mu'_3=\kappa_3+3\kappa_2\kappa_1+\kappa_1^3\,

\mu'_4=\kappa_4+4\kappa_3\kappa_1+3\kappa_2^2+6\kappa_2\kappa_1^2+\kappa_1^4\,

\mu'_5=\kappa_5+5\kappa_4\kappa_1+10\kappa_3\kappa_2 +10\kappa_3\kappa_1^2+15\kappa_2^2\kappa_1 +10\kappa_2\kappa_1^3+\kappa_1^5\,

\mu'_6=\kappa_6+6\kappa_5\kappa_1+15\kappa_4\kappa_2+15\kappa_4\kappa_1^2 +10\kappa_3^2+60\kappa_3\kappa_2\kappa_1+20\kappa_3\kappa_1^3+15\kappa_2^3 +45\kappa_2^2\kappa_1^2+15\kappa_2\kappa_1^4+\kappa_1^6.\,
The "prime" distinguishes the moments μ′_{n} from the central moments μ_{n}. To express the central moments as functions of the cumulants, just drop from these polynomials all terms in which κ_{1} appears as a factor:

\mu_1=0\,

\mu_2=\kappa_2\,

\mu_3=\kappa_3\,

\mu_4=\kappa_4+3\kappa_2^2\,

\mu_5=\kappa_5+10\kappa_3\kappa_2\,

\mu_6=\kappa_6+15\kappa_4\kappa_2+10\kappa_3^2+15\kappa_2^3.\,
Similarly, the nth cumulant κ_{n} is an nthdegree polynomial in the first n noncentral moments. The first few expressions are:

\kappa_1=\mu'_1\,

\kappa_2=\mu'_2{\mu'_1}^2\,

\kappa_3=\mu'_33\mu'_2\mu'_1+2{\mu'_1}^3\,

\kappa_4=\mu'_44\mu'_3\mu'_13{\mu'_2}^2+12\mu'_2{\mu'_1}^26{\mu'_1}^4\,

\kappa_5=\mu'_55\mu'_4\mu'_110\mu'_3\mu'_2+20\mu'_3{\mu'_1}^2+30{\mu'_2}^2\mu'_160\mu'_2{\mu'_1}^3+24{\mu'_1}^5\,

\kappa_6=\mu'_66\mu'_5\mu'_115\mu'_4\mu'_2+30\mu'_4{\mu'_1}^210{\mu'_3}^2+120\mu'_3\mu'_2\mu'_1120\mu'_3{\mu'_1}^3+30{\mu'_2}^3270{\mu'_2}^2{\mu'_1}^2+360\mu'_2{\mu'_1}^4120{\mu'_1}^6\,.
To express the cumulants κ_{n} for n > 1 as functions of the central moments, drop from these polynomials all terms in which μ'_{1} appears as a factor:

\kappa_2=\mu_2\,

\kappa_3=\mu_3\,

\kappa_4=\mu_43{\mu_2}^2\,

\kappa_5=\mu_510\mu_3\mu_2\,

\kappa_6=\mu_615\mu_4\mu_210{\mu_3}^2+30{\mu_2}^3\,.
To express the cumulants κ_{n} for n > 2 as functions of the standardized central moments, also set μ'_{2}=1 in the polynomials:

\kappa_3=\mu_3\,

\kappa_4=\mu_43\,

\kappa_5=\mu_510\mu_3\,

\kappa_6=\mu_615\mu_410{\mu_3}^2+30\,.
The cumulants are also related to the moments by the following recursion formula:

\kappa_n=\mu'_n\sum_{m=1}^{n1}{n1 \choose m1}\kappa_m \mu_{nm}'.
Cumulants and setpartitions
These polynomials have a remarkable combinatorial interpretation: the coefficients count certain partitions of sets. A general form of these polynomials is

\mu'_n=\sum_\Pi \prod_{B\in\pi}\kappa_{\leftB\right}
where

π runs through the list of all partitions of a set of size n;

"B ∈ π" means B is one of the "blocks" into which the set is partitioned; and

B is the size of the set B.
Thus each monomial is a constant times a product of cumulants in which the sum of the indices is n (e.g., in the term κ_{3} κ_{2}^{2} κ_{1}, the sum of the indices is 3 + 2 + 2 + 1 = 8; this appears in the polynomial that expresses the 8th moment as a function of the first eight cumulants). A partition of the integer n corresponds to each term. The coefficient in each term is the number of partitions of a set of n members that collapse to that partition of the integer n when the members of the set become indistinguishable.
Cumulants and combinatorics
Further connection between cumulants and combinatorics can be found in the work of GianCarlo Rota and Jianhong (Jackie) Shen, where links to invariant theory, symmetric functions, and binomial sequences are studied via umbral calculus.^{[9]}
Joint cumulants
The joint cumulant of several random variables X_{1}, ..., X_{n} is defined by a similar cumulant generating function

K(t_1,t_2,\dots,t_n)=\log E(\mathrm e^{\sum_{j=1}^n t_j X_j}).
A consequence is that

\kappa(X_1,\dots,X_n) =\sum_\pi (\pi1)!(1)^{\pi1}\prod_{B\in\pi}E\left(\prod_{i\in B}X_i\right)
where π runs through the list of all partitions of { 1, ..., n }, B runs through the list of all blocks of the partition π, and π is the number of parts in the partition. For example,

\kappa(X,Y,Z)=E(XYZ)E(XY)E(Z)E(XZ)E(Y)E(YZ)E(X)+2E(X)E(Y)E(Z).\,
If any of these random variables are identical, e.g. if X = Y, then the same formulae apply, e.g.

\kappa(X,X,Z)=E(X^2Z)2E(XZ)E(X)E(X^2)E(Z)+2E(X)^2E(Z),\,
although for such repeated variables there are more concise formulae. For zeromean random vectors,

\kappa(X,Y,Z)=E(XYZ).\,

\kappa(X,Y,Z,W) = E(XYZW)E(XY)E(ZW)E(XZ)E(YW)E(XW)E(YZ).\,
The joint cumulant of just one random variable is its expected value, and that of two random variables is their covariance. If some of the random variables are independent of all of the others, then any cumulant involving two (or more) independent random variables is zero. If all n random variables are the same, then the joint cumulant is the nth ordinary cumulant.
The combinatorial meaning of the expression of moments in terms of cumulants is easier to understand than that of cumulants in terms of moments:

E(X_1\cdots X_n)=\sum_\pi\prod_{B\in\pi}\kappa(X_i : i \in B).
For example:

E(XYZ)=\kappa(X,Y,Z)+\kappa(X,Y)\kappa(Z)+\kappa(X,Z)\kappa(Y) +\kappa(Y,Z)\kappa(X)+\kappa(X)\kappa(Y)\kappa(Z).\,
Another important property of joint cumulants is multilinearity:

\kappa(X+Y,Z_1,Z_2,\dots)=\kappa(X,Z_1,Z_2,\dots)+\kappa(Y,Z_1,Z_2,\dots).\,
Just as the second cumulant is the variance, the joint cumulant of just two random variables is the covariance. The familiar identity

\operatorname{var}(X+Y)=\operatorname{var}(X)+2\operatorname{cov}(X,Y)+\operatorname{var}(Y)\,
generalizes to cumulants:

\kappa_n(X+Y)=\sum_{j=0}^n {n \choose j} \kappa(\,\underbrace{X,\dots,X}_j,\underbrace{Y,\dots,Y}_{nj}\,).\,
Conditional cumulants and the law of total cumulance
The law of total expectation and the law of total variance generalize naturally to conditional cumulants. The case n = 3, expressed in the language of (central) moments rather than that of cumulants, says

\mu_3(X)=E(\mu_3(X\mid Y))+\mu_3(E(X\mid Y)) +3\,\operatorname{cov}(E(X\mid Y),\operatorname{var}(X\mid Y)).
In general,^{[10]}

\kappa(X_1,\dots,X_n)=\sum_\pi \kappa(\kappa(X_{\pi_1}\mid Y),\dots,\kappa(X_{\pi_b}\mid Y))
where

the sum is over all partitions π of the set { 1, ..., n } of indices, and

π_{1}, ..., π_{b} are all of the "blocks" of the partition π; the expression κ(X_{πm}) indicates that the joint cumulant of the random variables whose indices are in that block of the partition.
Relation to statistical physics
In statistical physics many extensive quantities – that is quantities that are proportional to the volume or size of a given system – are related to cumulants of random variables. The deep connection is that in a large system an extensive quantity like the energy or number of particles can be thought of as the sum of (say) the energy associated with a number of nearly independent regions. The fact that the cumulants of these nearly independent random variables will (nearly) add make it reasonable that extensive quantities should be expected to be related to cumulants.
A system in equilibrium with a thermal bath at temperature T can occupy states of energy E. The energy E can be considered a random variable, having the probability density. The partition function of the system is

Z(\beta) = \langle\exp(\beta E)\rangle,\,
where β = 1/(kT) and k is Boltzmann's constant and the notation \langle A \rangle has been used rather than \mathbb{E}\!\left[A\right] for the expectation value to avoid confusion with the energy, E. The Helmholtz free energy is then

F(\beta) = \beta^{1}\log Z \,
and is clearly very closely related to the cumulant generating function for the energy. The free energy gives access to all of the thermodynamics properties of the system via its first second and higher order derivatives, such as its internal energy, entropy, and specific heat. Because of the relationship between the free energy and the cumulant generating function, all these quantities are related to cumulants e.g. the energy and specific heat are given by

E = \langle E \rangle_c

C= dE/dT = k \beta^2\langle E^2 \rangle_c = k \beta^2(\langle E^2\rangle  \langle E\rangle ^2)
and \langle E^2\rangle_c symbolizes the second cumulant of the energy. Other free energy is often also a function of other variables such as the magnetic field or chemical potential \mu, e.g.

\Omega=\beta^{1}\log(\langle \exp(\beta E \beta\mu N) \rangle),\,
where N is the number of particles and \Omega is the grand potential. Again the close relationship between the definition of the free energy and the cumulant generating function implies that various derivatives of this free energy can be written in terms of joint cumulants of E and N.
History
The history of cumulants is discussed by Anders Hald.^{[11]}^{[12]}
Cumulants were first introduced by Thorvald N. Thiele, in 1889, who called them semiinvariants.^{[13]} They were first called cumulants in a 1932 paper^{[14]} by Ronald Fisher and John Wishart. Fisher was publicly reminded of Thiele's work by Neyman, who also notes previous published citations of Thiele brought to Fisher's attention.^{[15]} Stephen Stigler has said that the name cumulant was suggested to Fisher in a letter from Harold Hotelling. In a paper published in 1929,^{[16]} Fisher had called them cumulative moment functions. The partition function in statistical physics was introduced by Josiah Willard Gibbs in 1901. The free energy is often called Gibbs free energy. In statistical mechanics, cumulants are also known as Ursell functions relating to a publication in 1927.
Cumulants in generalized settings
Formal cumulants
More generally, the cumulants of a sequence { m_{n} : n = 1, 2, 3, ... }, not necessarily the moments of any probability distribution, are, by definition,

1+\sum_{n=1}^\infty m_n t^n/n!=\exp\left(\sum_{n=1}^\infty\kappa_n t^n/n!\right) ,
where the values of κ_{n} for n = 1, 2, 3, ... are found formally, i.e., by algebra alone, in disregard of questions of whether any series converges. All of the difficulties of the "problem of cumulants" are absent when one works formally. The simplest example is that the second cumulant of a probability distribution must always be nonnegative, and is zero only if all of the higher cumulants are zero. Formal cumulants are subject to no such constraints.
Bell numbers
In combinatorics, the nth Bell number is the number of partitions of a set of size n. All of the cumulants of the sequence of Bell numbers are equal to 1. The Bell numbers are the moments of the Poisson distribution with expected value 1.
Cumulants of a polynomial sequence of binomial type
For any sequence { κ_{n} : n = 1, 2, 3, ... } of scalars in a field of characteristic zero, being considered formal cumulants, there is a corresponding sequence { μ ′ : n = 1, 2, 3, ... } of formal moments, given by the polynomials above. For those polynomials, construct a polynomial sequence in the following way. Out of the polynomial

\mu'_6 = \kappa_6+6\kappa_5\kappa_1+15\kappa_4\kappa_2+15\kappa_4\kappa_1^2 +10\kappa_3^2+60\kappa_3\kappa_2\kappa_1 + 20\kappa_3\kappa_1^3+15\kappa_2^3 +45\kappa_2^2\kappa_1^2+15\kappa_2\kappa_1^4+\kappa_1^6
make a new polynomial in these plus one additional variable x:

p_6(x) = \kappa_6 \,x + (6\kappa_5\kappa_1 + 15\kappa_4\kappa_2 + 10\kappa_3^2)\,x^2m+(15\kappa_4\kappa_1^2+60\kappa_3\kappa_2\kappa_1+15\kappa_2^3)\,x^3 +(45\kappa_2^2\kappa_1^2)\,x^4+(15\kappa_2\kappa_1^4)\,x^5 +(\kappa_1^6)\,x^6,
and then generalize the pattern. The pattern is that the numbers of blocks in the aforementioned partitions are the exponents on x. Each coefficient is a polynomial in the cumulants; these are the Bell polynomials, named after Eric Temple Bell.
This sequence of polynomials is of binomial type. In fact, no other sequences of binomial type exist; every polynomial sequence of binomial type is completely determined by its sequence of formal cumulants.
Free cumulants
In the identity

E(X_1\cdots X_n)=\sum_\pi\prod_{B\in\pi}\kappa(X_i : i\in B)
one sums over all partitions of the set { 1, ..., n }. If instead, one sums only over the noncrossing partitions, then one gets "free cumulants" rather than conventional cumulants treated above. These play a central role in free probability theory.^{[17]} In that theory, rather than considering independence of random variables, defined in terms of Cartesian products of algebras of random variables, one considers instead "freeness" of random variables, defined in terms of free products of algebras rather than Cartesian products of algebras.
The ordinary cumulants of degree higher than 2 of the normal distribution are zero. The free cumulants of degree higher than 2 of the Wigner semicircle distribution are zero.^{[17]} This is one respect in which the role of the Wigner distribution in free probability theory is analogous to that of the normal distribution in conventional probability theory.
See also
References

^ Weisstein, Eric W. "Cumulant". From MathWorld – A Wolfram Web Resource. http://mathworld.wolfram.com/Cumulant.html

^ Kendall, M. G., Stuart, A. (1969) The Advanced Theory of Statistics, Volume 1 (3rd Edition). Griffin, London. (Section 3.12)

^ Lukacs, E. (1970) Characteristic Functions (2nd Edition). Griffin, London. (Page 27)

^ Lukacs, E. (1970) Characteristic Functions (2nd Edition). Griffin, London. (Section 2.4)

^ Aapo Hyvarinen, Juha Karhunen, and Erkki Oja (2001) Independent Component Analysis, John Wiley & Sons. (Section 2.7.2)

^ Reluga, T. (2009). "Branching Processes and Noncommuting Random Variables in Population Biology". Canadian Applied Mathematics Quarterly 17 (2): 387–408.

^ Hamedani, G. G.; Volkmer, Hans; Behboodian, J. (20120301). "A note on subindependent random variables and a class of bivariate mixtures". Studia Scientiarum Mathematicarum Hungarica 49 (1): 19–25.

^ Lukacs, E. (1970) Characteristic Functions (2nd Edition), Griffin, London. (Theorem 7.3.5)

^ Rota, G.C.; Shen, J. (2000). "On the Combinatorics of Cumulants". Journal of Combinatorial Theory. Series A 91 (1–2): 283–304.

^ Brillinger, D.R. (1969). "The Calculation of Cumulants via Conditioning". Annals of the Institute of Statistical Mathematics 21: 215–218.

^ )

^

^ H. Cramér (1946) Mathematical Methods of Statistics, Princeton University Press, Section 15.10, p. 186.

^ Fisher, R.A. , John Wishart, J.. (1932) The derivation of the pattern formulae of twoway partitions from those of simpler patterns, Proceedings of the London Mathematical Society, Series 2, v. 33, pp. 195–208 doi:10.1112/plms/s233.1.195

^ Neyman, J. (1956): ‘Note on an Article by Sir Ronald Fisher,’ Journal of the Royal Statistical Society, Series B (Methodological), 18, pp. 288–94.

^ Fisher, R. A. (1929). "Moments and Product Moments of Sampling Distributions". Proceedings of the London Mathematical Society 30: 199238.

^ ^{a} ^{b} Novak, Jonathan; Śniady, Piotr (2011). "What Is a Free Cumulant?".
External links
This article was sourced from Creative Commons AttributionShareAlike License; additional terms may apply. World Heritage Encyclopedia content is assembled from numerous content providers, Open Access Publishing, and in compliance with The Fair Access to Science and Technology Research Act (FASTR), Wikimedia Foundation, Inc., Public Library of Science, The Encyclopedia of Life, Open Book Publishers (OBP), PubMed, U.S. National Library of Medicine, National Center for Biotechnology Information, U.S. National Library of Medicine, National Institutes of Health (NIH), U.S. Department of Health & Human Services, and USA.gov, which sources content from all federal, state, local, tribal, and territorial government publication portals (.gov, .mil, .edu). Funding for USA.gov and content contributors is made possible from the U.S. Congress, EGovernment Act of 2002.
Crowd sourced content that is contributed to World Heritage Encyclopedia is peer reviewed and edited by our editorial staff to ensure quality scholarly research articles.
By using this site, you agree to the Terms of Use and Privacy Policy. World Heritage Encyclopedia™ is a registered trademark of the World Public Library Association, a nonprofit organization.