- Additive white Gaussian noise
=Explanation=

In communications, the

**additive white Gaussian noise**(**AWGN**) channel model is one in which the only impairment is the linear addition ofwideband orwhite noise with a constantspectral density (expressed aswatt s perhertz of bandwidth) and aGaussian distribution of amplitude. The model does not account for the phenomena offading ,frequency selectivity,interference ,nonlinearity or dispersion. However, it produces simple, tractable mathematical models which are useful for gaining insight into the underlying behavior of a system before these other phenomena are considered.Wideband Gaussian noise comes from many natural sources, such as the thermal vibrations of atoms in antennas (referred to as

thermal noise orJohnson-Nyquist noise ),shot noise , black body radiation from the earth and other warm objects, and from celestial sources such as theSun .The AWGN channel is a good model for many

satellite and deep space communication links. It is not a good model for most terrestrial links because of multipath, terrain blocking, interference, etc. However for terrestrial path modeling, AWGN is commonly used to simulate background noise of the channel under study, in addition to multipath, terrain blocking, interference, ground clutter and self interference that modern radio systems encounter in terrestrial operation.**Channel Capacity**The AWGN channel is represented by a series of outputs $Y\_i$ at discrete time event index $i$. $Y\_i$ is the sum of the input $X\_i$ and noise, $Z\_i$, where $Z\_i$ is i.i.d. and drawn from a zero mean normal distribution with variance $n$ (the noise). The $Z\_i$ are further assumed to not be correlated with the $X\_i$.:$Z\_i\; sim\; N(0,\; n),!$:$Y\_i\; =\; X\_i\; +\; Z\_i.,!$

The capacity of the channel is infinite unless the noise n is nonzero, and the $X\_i$ are sufficiently constrained. The most common constraint on the input is the so-called "power" constraint, requiring that for a codeword $(x\_1,\; x\_2,\; ...\; ,\; x\_n)$ transmitted through the channel, we have:

:$frac\{1\}\{n\}sum\_\{i=1\}^n\; x\_i^2\; leq\; P$

Therefore the

channel capacity for the power-constrained channel is given by::$C\; =\; max\_\{f(x)\; s.t.\; E\; left(\; X^2\; ight)\; leq\; P\}\; I(X;Y),!$

Where $f(x)$ is the distribution of $X$. Expand $I(X;Y)$, writing it in terms of the

differential entropy ::$egin\{align\}I(X;Y)\; =\; h(Y)\; -\; h(Y|X)=\; h(Y)-h(X+Z|X)=\; h(Y)-h(Z|X)end\{align\},!$But $X$ and $Z$ are independent, therefore::$I(X;Y)\; =\; h(Y)\; -\; h(Z),!$

Evaluating the

differential entropy of a Gaussian gives::$h(Z)\; =\; frac\{1\}\{2\}\; log(2\; pi\; e\; n),!$Because $X$ and $Z$ are independent and their sum gives $Y$:

:$E(Y^2)\; =\; E(X+Z)^2\; =\; E(X^2)\; +\; 2E(X)E(Z)+E(Z^2)\; =\; P\; +\; n,!$

From this bound, we infer from a property of the differential entropy that

:$h(Y)\; leq\; frac\{1\}\{2\}\; log(2\; pi\; e(P+n)),!$

Therefore the channel capacity is given by the highest achievable bound on the mutual information::$I(X;Y)\; leq\; frac\{1\}\{2\}log(2\; pi\; e\; (P+n))\; -\; frac\; \{1\}\{2\}log(2\; pi\; e\; n),!$

Where $I(X;Y)$ is maximized when:

:$X\; sim\; N(0,\; P),!$

Thus the channel capacity $C$ for the AWGN channel is given by::$C\; =\; frac\; \{1\}\{2\}\; log(1+frac\{P\}\{n\}),!$

**Channel Capacity and Sphere Packing**Suppose that we are sending messages through the channel with index ranging from $1$ to $M$, the number of distinct possible messages. If we encode the $M$ messages to $n$ bits, then we define the rate $R$ as:

:$R\; =\; frac\; \{log\; M\}\{n\},!$

A rate is said to be achievable if there is a sequence of codes so that the maximum probability of error tends to zero as $n$ approaches infinity. The capacity $C$ is the highest achievable rate.

Consider a codeword of length of length $n$ sent through the AWGN channel with noise level $N$. When received, the codeword vector variance is now $N$, and its mean is the codeword sent. The vector is very likely to be contained in a sphere of radius $sqrt\{n(N+epsilon)\}$ around the codeword sent. If we decode by mapping every message received onto the codeword at the center of this sphere, then an error occurs only when the received vector is outside of this sphere, which is very unlikely.

Each codeword vector has an associated sphere of received codeword vectors which are decoded to it and each such sphere must map uniquely onto a codeword. Because these spheres therefore must not intersect, we are faced with the problem of

sphere packing . How many distinct codewords can we pack into our $n$-bit codeword vector? The received vectors have a maximum energy of $n(P+N)$ and therefore must occupy a sphere of radius $sqrt\{n(P+N)\}$. Each codeword sphere has radius $sqrt\{nN\}$. The volume of an n-dimensional sphere is directly proportional to $r^n$, so the maximum number of uniquely decodeable spheres that can be packed into our sphere with transmission power P is::$frac\{(n(P+N))^frac\{n\}\{2\{(nN)^frac\{n\}\{2\; =\; 2^frac\{n\}\{2\}log(1+P/N),!$

By this argument, the rate R can be no more than $frac\{1\}\{2\}log(1+P/N)$.

**Achievability**In this section, we show achievability of the upper bound on the rate from the last section.

A codebook, known to both encoder and decoder, is generated by selecting codewords of length n, i.i.d. Gaussian with variance $P-epsilon$ and mean zero. For large n, the empirical variance of the codebook will be very close to the variance of its distribution, thereby avoiding violation of the power constraint probabilistically.

Received messages are decoded to a message in the codebook which is uniquely jointly typical. If there is no such message or if the power constraint is violated, a decoding error is declared.

Let $X^n(i)$ denote the codeword for message $i$, while $Y^n$ is, as before the received vector. Define the following three events:

# Event $U$:the power of the received message is larger than $P$.

# Event $V$: the transmitted and received codewords are not jointly typical.

# Event $E\_j$: $(X^n(j),\; Y^n)$ is in $A\_epsilon^\{(n)\}$, thetypical set where $i\; eq\; j$, which is to say that the incorrect codeword is jointly typical with the received vector.An error therefore occurs if $U$, $V$ or any of the $E\_i$ occur. By the law of large numbers, $P(U)$ goes to zero as n approaches infinity, and by the joint

Asymptotic Equipartition Property the same applies to $P(V)$. Therefore, for a sufficiently large $n$, both $P(U)$ and $P(V)$ are each less than $epsilon$. Since $X^n(i)$ and $X^n(j)$ are independent for $i\; eq\; j$, we have that $X^n(i)$ and $Y^n$ are also independent. Therefore, by the joint AEP, $P(E\_j)\; =\; 2^\{-n(I(X;Y)-3epsilon)\}$. This allows us to calculate $P^\{(n)\}\_e$, the probability of error as follows:We may apply Jensen's equality to $log(1+x)$, a concave (downward) function of x, to get::$frac\{1\}\{n\}\; sum\_\{i=1\}^\{n\}\; frac\{1\}\{2\}logleft(1+frac\{P\_i\}\{N\}\; ight)\; leqfrac\{1\}\{2\}logleft(1+frac\{1\}\{n\}sum\_\{i=1\}^\{n\}frac\{P\_i\}\{N\}\; ight),!$

Because each codeword individually satisfies the power constraint, the average also satisfies the power constraint. Therefore

:$frac\{1\}\{n\}sum\_\{i=1\}^\{n\}\; frac\{P\_i\}\{N\},!$

Which we may apply to simplify the inequality above and get::$frac\{1\}\{2\}logleft(1+frac\{1\}\{n\}sum\_\{i=1\}^\{n\}frac\{P\_i\}\{N\}\; ight)\; leqfrac\{1\}\{2\}logleft(1+frac\{P\}\{N\}\; ight),!$

Therefore, it must be that $R\; leq\; frac\{1\}\{2\}log\; left(1+\; frac\{P\}\{N\}\; ight)\; +\; epsilon\_n$. Therefore, R must be less than a value arbitrarily close to the capacity derived earlier, as $epsilon\_n\; ightarrow\; 0$.

**Effects in Time Domain**In serial data communications, the AWGN mathematical model is used to model the timing error caused by random

jitter (RJ).The graph to the right shows an example of timing errors associated with AWGN. The variable Δt represents the uncertainty in the zero crossing. As the amplitude of the AWGN is increased, the

Signal-to-noise ratio decreases. This results in increased uncertainty Δt. [1]When affected by AWGN, The average number of either positive going or negative going zero-crossings per second at the output of a narrow bandpass filter when the input is a sine wave is:

Where

* f

_{0}= the center frequency of the filter

* B = the filter bandwidth

* SNR = the signal-to-noise power ratio in linear terms**Effects in Phasor Domain**In modern communication systems, bandlimited AWGN cannot be ignored. When modeling bandlimited AWGN in the phasor domain, statistical analysis reveals that the amplitudes of the real and imaginary contributions are independent variables which follow the

Gaussian distribution model. When combined, the resultant phasor's magnitude is a Rayleigh distributed random variable (seeRayleigh distribution ) while the phase is uniformly distributed from 0 to 2π.The graph to the right shows an example of how bandlimited AWGN can affect a coherent carrier signal. The instantaneous response of the Noise Vector cannot be precisely predicted, however its time-averaged response can be statistically predicted. As shown in the graph, we confidently predict that the noise phasor will reside inside the 1σ circle about 38% of the time; the noise phasor will reside inside the 2σ circle about 86% of the time; and the noise phasor will reside inside the 3σ circle about 98% of the time. [

*citation*]

title = Radio Receiver Design

first = Kevin | last = McClaning

publisher = Noble Publishing Corporation**See also***

Gaussian noise

*White noise

*ground bounce

*thermal noise /Johnson-Nyquist noise

*shot noise **References**

*Wikimedia Foundation.
2010.*

### Look at other dictionaries:

**Gaussian noise**— is noise that has a probability density function (abbreviated pdf) of the normal distribution (also known as Gaussian distribution). In other words, the values that the noise can take on are Gaussian distributed. It is most commonly used as… … Wikipedia**White noise**— is a random signal (or process) with a flat power spectral density. In other words, the signal contains equal power within a fixed bandwidth at any center frequency. White noise draws its name from white light in which the power spectral density… … Wikipedia**Noise (audio)**— Noise in audio, recording, and broadcast systems refers to the residual low level sound (usually hiss and hum) that is heard in quiet periods of a programme. In audio engineering, it can refer either to the acoustic noise from loudspeakers, or to … Wikipedia**Noise shaping**— is a technique typically used in digital audio, image, and video processing, usually in combination with dithering, as part of the process of quantization or bit depth reduction of a digital signal. Its purpose is to increase the apparent signal… … Wikipedia**Noise measurement**— is carried out in various fields. In acoustics, it can be for the purpose of measuring environmental noise, or part of a test procedure using white noise, or some other specialised form of test signal. In electronics it relates to the sensitivity … Wikipedia**Noise figure**— (NF) is a measure of degradation of the signal to noise ratio (SNR), caused by components in a radio frequency (RF) signal chain. The noise figure is defined as the ratio of the output noise power of a device to the portion thereof attributable… … Wikipedia**Noise (electronics)**— Electronic noise [1] is a random fluctuation in an electrical signal, a characteristic of all electronic circuits. Noise generated by electronic devices varies greatly, as it can be produced by several different effects. Thermal noise is… … Wikipedia**Noise reduction**— For sound proofing, see soundproofing. For scientific aspects of noise reduction of machinery and products, see noise control. Noise reduction is the process of removing noise from a signal. All recording devices, both analogue or digital, have… … Wikipedia**Noise temperature**— In electronics, noise temperature is one way of expressing the level of available noise power introduced by a component or source. The power spectral density of the noise is expressed in terms of the temperature (in kelvins) that would produce… … Wikipedia**Noise (video)**— For other uses, see Noise (disambiguation). On an empty channel, an analog TV receiver shows a snowy noise pattern. A similar pattern is shown if on a digital channel when not using a DTV converter box … Wikipedia