Additive white Gaussian noise (AWGN) is a basic noise model used in Information theory to mimic the effect of many random processes that occur in nature. The modifiers denote specific characteristics:

Additive because it is added to any noise that might be intrinsic to the information system.

White refers to the idea that it has uniform power across the frequency band for the information system. It is an analogy to the color white which has uniform emissions at all frequencies in the visible spectrum.

Gaussian because it has a normal distribution in the time domain with an average time domain value of zero.
Wideband noise comes from many natural sources, such as the thermal vibrations of atoms in conductors (referred to as thermal noise or JohnsonNyquist noise), shot noise, black body radiation from the earth and other warm objects, and from celestial sources such as the Sun. The central limit theorem of probability theory indicates that the summation of many random processes will tend to have distribution called Gaussian or Normal.
AWGN is often used as a channel model in which the only impairment to communication is a linear addition of wideband or white noise with a constant spectral density (expressed as watts per hertz of bandwidth) and a Gaussian distribution of amplitude. The model does not account for fading, frequency selectivity, interference, nonlinearity or dispersion. However, it produces simple and tractable mathematical models which are useful for gaining insight into the underlying behavior of a system before these other phenomena are considered.
The AWGN channel is a good model for many satellite and deep space communication links. It is not a good model for most terrestrial links because of multipath, terrain blocking, interference, etc. However, for terrestrial path modeling, AWGN is commonly used to simulate background noise of the channel under study, in addition to multipath, terrain blocking, interference, ground clutter and self interference that modern radio systems encounter in terrestrial operation.
Contents

Channel capacity 1

Channel capacity and sphere packing 1.1

Achievability 1.2

Coding theorem converse 1.3

Effects in time domain 2

Effects in phasor domain 3

See also 4

References 5
Channel capacity
The AWGN channel is represented by a series of outputs Y_i at discrete time event index i. Y_i is the sum of the input X_i and noise, Z_i, where Z_i is independent and identically distributed and drawn from a zeromean normal distribution with variance N (the noise). The Z_i are further assumed to not be correlated with the X_i.

Z_i \sim \mathcal{N}(0, N) \,\!

Y_i = X_i + Z_i\sim \mathcal{N}(X_i, N). \,\!
The capacity of the channel is infinite unless the noise n is nonzero, and the X_i are sufficiently constrained. The most common constraint on the input is the socalled "power" constraint, requiring that for a codeword (x_1, x_2, \dots , x_k) transmitted through the channel, we have:

\frac{1}{k}\sum_{i=1}^k x_i^2 \leq P,
where P represents the maximum channel power. Therefore, the channel capacity for the powerconstrained channel is given by:

C = \max_{f(x) \text{ s.t. }E \left( X^2 \right) \leq P} I(X;Y) \,\!
Where f(x) is the distribution of X. Expand I(X;Y), writing it in terms of the differential entropy:

\begin{align} I(X;Y) = h(Y)  h(YX) &= h(Y)h(X+ZX) &= h(Y)h(ZX) \end{align} \,\!
But X and Z are independent, therefore:

I(X;Y) = h(Y)  h(Z) \,\!
Evaluating the differential entropy of a Gaussian gives:

h(Z) = \frac{1}{2} \log(2 \pi e N) \,\!
Because X and Z are independent and their sum gives Y:

E(Y^2) = E(X+Z)^2 = E(X^2) + 2E(X)E(Z)+E(Z^2) = P + N \,\!
From this bound, we infer from a property of the differential entropy that

h(Y) \leq \frac{1}{2} \log(2 \pi e(P+N)) \,\!
Therefore the channel capacity is given by the highest achievable bound on the mutual information:

I(X;Y) \leq \frac{1}{2}\log(2 \pi e (P+N))  \frac {1}{2}\log(2 \pi e N) \,\!
Where I(X;Y) is maximized when:

X \sim \mathcal{N}(0, P) \,\!
Thus the channel capacity C for the AWGN channel is given by:

C = \frac {1}{2} \log\left(1+\frac{P}{N}\right) \,\!
Channel capacity and sphere packing
Suppose that we are sending messages through the channel with index ranging from 1 to M, the number of distinct possible messages. If we encode the M messages to n bits, then we define the rate R as:

R = \frac {\log M}{n} \,\!
A rate is said to be achievable if there is a sequence of codes so that the maximum probability of error tends to zero as n approaches infinity. The capacity C is the highest achievable rate.
Consider a codeword of length n sent through the AWGN channel with noise level N. When received, the codeword vector variance is now N, and its mean is the codeword sent. The vector is very likely to be contained in a sphere of radius \sqrt{n(N+\epsilon)} around the codeword sent. If we decode by mapping every message received onto the codeword at the center of this sphere, then an error occurs only when the received vector is outside of this sphere, which is very unlikely.
Each codeword vector has an associated sphere of received codeword vectors which are decoded to it and each such sphere must map uniquely onto a codeword. Because these spheres therefore must not intersect, we are faced with the problem of sphere packing. How many distinct codewords can we pack into our nbit codeword vector? The received vectors have a maximum energy of n(P+N) and therefore must occupy a sphere of radius \sqrt{n(P+N)}. Each codeword sphere has radius \sqrt{nN}. The volume of an ndimensional sphere is directly proportional to r^n, so the maximum number of uniquely decodeable spheres that can be packed into our sphere with transmission power P is:

\frac{(n(P+N))^\frac{n}{2}}{(nN)^\frac{n}{2}} = 2^{\frac{n}{2}\log(1+P/N)} \,\!
By this argument, the rate R can be no more than \frac{1}{2}\log(1+P/N).
Achievability
In this section, we show achievability of the upper bound on the rate from the last section.
A codebook, known to both encoder and decoder, is generated by selecting codewords of length n, i.i.d. Gaussian with variance P\epsilon and mean zero. For large n, the empirical variance of the codebook will be very close to the variance of its distribution, thereby avoiding violation of the power constraint probabilistically.
Received messages are decoded to a message in the codebook which is uniquely jointly typical. If there is no such message or if the power constraint is violated, a decoding error is declared.
Let X^n(i) denote the codeword for message i, while Y^n is, as before the received vector. Define the following three events:

Event U:the power of the received message is larger than P.

Event V: the transmitted and received codewords are not jointly typical.

Event E_j: (X^n(j), Y^n) is in A_\epsilon^{(n)}, the typical set where i \neq j, which is to say that the incorrect codeword is jointly typical with the received vector.
An error therefore occurs if U, V or any of the E_i occur. By the law of large numbers, P(U) goes to zero as n approaches infinity, and by the joint Asymptotic Equipartition Property the same applies to P(V). Therefore, for a sufficiently large n, both P(U) and P(V) are each less than \epsilon. Since X^n(i) and X^n(j) are independent for i \neq j, we have that X^n(i) and Y^n are also independent. Therefore, by the joint AEP, P(E_j) = 2^{n(I(X;Y)3\epsilon)}. This allows us to calculate P^{(n)}_e, the probability of error as follows:

\begin{align} P^{(n)}_e & \leq P(U) + P(V) + \sum_{j \neq i} P(E_j) \\ & \leq \epsilon + \epsilon + \sum_{j \neq i} 2^{n(I(X;Y)3\epsilon)} \\ & \leq 2\epsilon + (2^{nR}1)2^{n(I(X;Y)3\epsilon)} \\ & \leq 2\epsilon + (2^{3n\epsilon})2^{n(I(X;Y)R)} \\ & \leq 3\epsilon \end{align}
Therefore, as n approaches infinity, P^{(n)}_e goes to zero and R < I(X;Y)  3\epsilon. Therefore there is a code of rate R arbitrarily close to the capacity derived earlier.
Coding theorem converse
Here we show that rates above the capacity C = \frac {1}{2} \log(1+\frac{P}{N}) are not achievable.
Suppose that the power constraint is satisfied for a codebook, and further suppose that the messages follow a uniform distribution. Let W be the input messages and \hat{W} the output messages. Thus the information flows as:
W \longrightarrow X^{(n)}(W) \longrightarrow Y^{(n)} \longrightarrow \hat{W}
Making use of Fano's inequality gives:
H(W\hat{W}) \leq 1+nRP^{(n)}_e = n \epsilon_n where \epsilon_n \rightarrow 0 as P^{(n)}_e \rightarrow 0
Let X_i be the encoded message of codeword index i. Then:

\begin{align} nR & = H(W) \\ & =I(W;\hat{W}) + H(W\hat{W}) \\ & \leq I(W;\hat{W}) + n\epsilon_n \\ & \leq I(X^{(n)}; Y^{(n)}) + n\epsilon_n \\ & = h(Y^{(n)})  h(Y^{(n)}X^{(n)}) + n\epsilon_n \\ & = h(Y^{(n)})  h(Z^{(n)}) + n\epsilon_n \\ & \leq \sum_{i=1}^{n} Y_i h(Z^{(n)}) + n\epsilon_n \\ & \leq \sum_{i=1}^{n} I(X_i; Y_i) + n\epsilon_n \end{align}
Let P_i be the average power of the codeword of index i:

P_i = \frac{1}{2^{nR}}\sum_{w}x^2_i(w) \,\!
Where the sum is over all input messages w. X_i and Z_i are independent, thus the expectation of the power of Y_i is, for noise level N:

E(Y_i^2) = P_i+N \,\!
And, if Y_i is normally distributed, we have that

h(Y_i) \leq \frac{1}{2}\log{2 \pi e} (P_i +N) \,\!
Therefore,

\begin{align} nR & \leq \sum(h(Y_i)h(Z_i)) + n \epsilon_n \\ & \leq \sum \left( \frac{1}{2} \log(2 \pi e (P_i + N))  \frac{1}{2}\log(2 \pi e N)\right) + n \epsilon_n \\ & = \sum \frac{1}{2} \log (1 + \frac{P_i}{N}) + n \epsilon_n \end{align}
We may apply Jensen's equality to \log(1+x), a concave (downward) function of x, to get:

\frac{1}{n} \sum_{i=1}^{n} \frac{1}{2}\log\left(1+\frac{P_i}{N}\right) \leq \frac{1}{2}\log\left(1+\frac{1}{n}\sum_{i=1}^{n}\frac{P_i}{N}\right) \,\!
Because each codeword individually satisfies the power constraint, the average also satisfies the power constraint. Therefore

\frac{1}{n}\sum_{i=1}^{n} \frac{P_i}{N} \,\!
Which we may apply to simplify the inequality above and get:

\frac{1}{2}\log\left(1+\frac{1}{n}\sum_{i=1}^{n}\frac{P_i}{N}\right) \leq \frac{1}{2}\log\left(1+\frac{P}{N}\right) \,\!
Therefore, it must be that R \leq \frac{1}{2}\log \left(1+ \frac{P}{N}\right) + \epsilon_n. Therefore, R must be less than a value arbitrarily close to the capacity derived earlier, as \epsilon_n \rightarrow 0.
Effects in time domain
ZeroCrossings of a Noisy Cosine
In serial data communications, the AWGN mathematical model is used to model the timing error caused by random jitter (RJ).
The graph to the right shows an example of timing errors associated with AWGN. The variable Δt represents the uncertainty in the zero crossing. As the amplitude of the AWGN is increased, the signaltonoise ratio decreases. This results in increased uncertainty Δt.^{[1]}
When affected by AWGN, The average number of either positive going or negative going zerocrossings per second at the output of a narrow bandpass filter when the input is a sine wave is:

\frac{\mathrm{positive\ zero\ crossings}}{\mathrm{second}} = \frac{\mathrm{negative \ zero\ crossings}}{\mathrm{second}}

= f_0 \sqrt{\frac{\mathrm{SNR} + 1 + \frac{B^2}{12f_0^2}}{\mathrm{SNR} + 1}}
Where

f_{0} = the center frequency of the filter

B = the filter bandwidth

SNR = the signaltonoise power ratio in linear terms
Effects in phasor domain
AWGN Contributions in the Phasor Domain
In modern communication systems, bandlimited AWGN cannot be ignored. When modeling bandlimited AWGN in the phasor domain, statistical analysis reveals that the amplitudes of the real and imaginary contributions are independent variables which follow the Gaussian distribution model. When combined, the resultant phasor's magnitude is a Rayleigh distributed random variable while the phase is uniformly distributed from 0 to 2π.
The graph to the right shows an example of how bandlimited AWGN can affect a coherent carrier signal. The instantaneous response of the Noise Vector cannot be precisely predicted, however its timeaveraged response can be statistically predicted. As shown in the graph, we confidently predict that the noise phasor will reside inside the 1σ circle about 38% of the time; the noise phasor will reside inside the 2σ circle about 86% of the time; and the noise phasor will reside inside the 3σ circle about 98% of the time.^{[1]}
See also
References

^ ^{a} ^{b} McClaning, Kevin, Radio Receiver Design, Noble Publishing Corporation
Noise (physics and telecommunications)


General



Noise in...



Class of noise



Engineering
terms



Ratios



Related topics



This article was sourced from Creative Commons AttributionShareAlike License; additional terms may apply. World Heritage Encyclopedia content is assembled from numerous content providers, Open Access Publishing, and in compliance with The Fair Access to Science and Technology Research Act (FASTR), Wikimedia Foundation, Inc., Public Library of Science, The Encyclopedia of Life, Open Book Publishers (OBP), PubMed, U.S. National Library of Medicine, National Center for Biotechnology Information, U.S. National Library of Medicine, National Institutes of Health (NIH), U.S. Department of Health & Human Services, and USA.gov, which sources content from all federal, state, local, tribal, and territorial government publication portals (.gov, .mil, .edu). Funding for USA.gov and content contributors is made possible from the U.S. Congress, EGovernment Act of 2002.
Crowd sourced content that is contributed to World Heritage Encyclopedia is peer reviewed and edited by our editorial staff to ensure quality scholarly research articles.
By using this site, you agree to the Terms of Use and Privacy Policy. World Heritage Encyclopedia™ is a registered trademark of the World Public Library Association, a nonprofit organization.