Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Additive white Gaussian noise

From Wikipedia, the free encyclopedia
Basic noise model used in information theory
icon
This articlerelies largely or entirely on asingle source. Relevant discussion may be found on thetalk page. Please helpimprove this article byintroducing citations to additional sources.
Find sources: "Additive white Gaussian noise" – news ·newspapers ·books ·scholar ·JSTOR
(February 2020)

Additive white Gaussian noise (AWGN) is a basic noise model used ininformation theory to mimic the effect of many random processes that occur in nature. The modifiers denote specific characteristics:

  • Additive because it is added to any noise that might be intrinsic to the information system.
  • White refers to the idea that it has uniformpower spectral density across the frequency band for the information system. It is an analogy to thecolor white which may be realized by uniform emissions at all frequencies in thevisible spectrum.
  • Gaussian because it has anormal distribution in the time domain with an average time domain value of zero (Gaussian process).

Wideband noise comes from many natural noise sources, such as the thermal vibrations of atoms in conductors (referred to as thermal noise orJohnson–Nyquist noise),shot noise,black-body radiation from the earth and other warm objects, and from celestial sources such as the Sun. Thecentral limit theorem ofprobability theory indicates that the summation of many random processes will tend to have distribution called Gaussian or Normal.

AWGN is often used as achannel model in which the only impairment to communication is a linear addition ofwideband orwhite noise with a constantspectral density (expressed aswatts perhertz ofbandwidth) and aGaussian distribution of amplitude. The model does not account forfading,frequency selectivity,interference,nonlinearity ordispersion. However, it produces simple and tractable mathematical models which are useful for gaining insight into the underlying behavior of a system before these other phenomena are considered.

The AWGN channel is a good model for manysatellite and deep space communication links. It is not a good model for most terrestrial links because of multipath, terrain blocking, interference, etc. However, for terrestrial path modeling, AWGN is commonly used to simulate background noise of the channel under study, in addition to multipath, terrain blocking, interference, ground clutter and self interference that modern radio systems encounter in terrestrial operation.

Channel capacity

[edit]

The AWGN channel is represented by a series of outputsYi{\displaystyle Y_{i}} at discrete-time event indexi{\displaystyle i}.Yi{\displaystyle Y_{i}} is the sum of the inputXi{\displaystyle X_{i}} and noise,Zi{\displaystyle Z_{i}}, whereZi{\displaystyle Z_{i}} isindependent and identically distributed and drawn from a zero-meannormal distribution withvarianceN{\displaystyle N} (the noise). TheZi{\displaystyle Z_{i}} are further assumed to not be correlated with theXi{\displaystyle X_{i}}.

ZiN(0,N){\displaystyle Z_{i}\sim {\mathcal {N}}(0,N)\,\!}
Yi=Xi+Zi.{\displaystyle Y_{i}=X_{i}+Z_{i}.\,\!}

The capacity of the channel is infinite unless the noiseN{\displaystyle N} is nonzero, and theXi{\displaystyle X_{i}} are sufficiently constrained. The most common constraint on the input is the so-called "power" constraint, requiring that for a codeword(x1,x2,,xk){\displaystyle (x_{1},x_{2},\dots ,x_{k})} transmitted through the channel, we have:

1ki=1kxi2P,{\displaystyle {\frac {1}{k}}\sum _{i=1}^{k}x_{i}^{2}\leq P,}

whereP{\displaystyle P} represents the maximum channel power.Therefore, thechannel capacity for the power-constrained channel is given by:[clarification needed]

C=max{I(X;Y):f s.t. E(X2)P}{\displaystyle C=\max \left\{I(X;Y):f{\text{ s.t. }}E\left(X^{2}\right)\leq P\right\}\,\!}

wheref{\displaystyle f} is the distribution ofX{\displaystyle X}. ExpandI(X;Y){\displaystyle I(X;Y)}, writing it in terms of thedifferential entropy:

I(X;Y)=h(Y)h(YX)=h(Y)h(X+ZX)=h(Y)h(ZX){\displaystyle {\begin{aligned}&I(X;Y)=h(Y)-h(Y\mid X)\\[5pt]={}&h(Y)-h(X+Z\mid X)\\[5pt]={}&h(Y)-h(Z\mid X)\end{aligned}}\,\!}

ButX{\displaystyle X} andZ{\displaystyle Z} are independent, therefore:

I(X;Y)=h(Y)h(Z){\displaystyle I(X;Y)=h(Y)-h(Z)\,\!}

Evaluating thedifferential entropy of a Gaussian gives:

h(Z)=12log(2πeN){\displaystyle h(Z)={\frac {1}{2}}\log(2\pi eN)\,\!}

BecauseX{\displaystyle X} andZ{\displaystyle Z} are independent and their sum givesY{\displaystyle Y}:

E(Y2)=E((X+Z)2)=E(X2)+2E(X)E(Z)+E(Z2)P+N{\displaystyle E(Y^{2})=E((X+Z)^{2})=E(X^{2})+2E(X)E(Z)+E(Z^{2})\leq P+N\,\!}

From this bound, we infer from a property of the differential entropy that

h(Y)12log(2πe(P+N)){\displaystyle h(Y)\leq {\frac {1}{2}}\log(2\pi e(P+N))\,\!}

Therefore, the channel capacity is given by the highest achievable bound on themutual information:

I(X;Y)12log(2πe(P+N))12log(2πeN){\displaystyle I(X;Y)\leq {\frac {1}{2}}\log(2\pi e(P+N))-{\frac {1}{2}}\log(2\pi eN)\,\!}

WhereI(X;Y){\displaystyle I(X;Y)} is maximized when:

XN(0,P){\displaystyle X\sim {\mathcal {N}}(0,P)\,\!}

Thus the channel capacityC{\displaystyle C} for the AWGN channel is given by:

C=12log(1+PN){\displaystyle C={\frac {1}{2}}\log \left(1+{\frac {P}{N}}\right)\,\!}

Channel capacity and sphere packing

[edit]

Suppose that we are sending messages through the channel with index ranging from1{\displaystyle 1} toM{\displaystyle M}, the number of distinct possible messages. If we encode theM{\displaystyle M} messages ton{\displaystyle n} bits, then we define the rateR{\displaystyle R} as:

R=logMn{\displaystyle R={\frac {\log M}{n}}\,\!}

A rate is said to be achievable if there is a sequence of codes so that the maximum probability of error tends to zero asn{\displaystyle n} approaches infinity. The capacityC{\displaystyle C} is the highest achievable rate.

Consider a codeword of lengthn{\displaystyle n} sent through the AWGN channel with noise levelN{\displaystyle N}. When received, the codeword vector variance is nowN{\displaystyle N}, and its mean is the codeword sent. The vector is very likely to be contained in a sphere of radiusn(N+ε){\textstyle {\sqrt {n(N+\varepsilon )}}} around the codeword sent. If we decode by mapping every message received onto the codeword at the center of this sphere, then an error occurs only when the received vector is outside of this sphere, which is very unlikely.

Each codeword vector has an associated sphere of received codeword vectors which are decoded to it and each such sphere must map uniquely onto a codeword. Because these spheres therefore must not intersect, we are faced with the problem ofsphere packing. How many distinct codewords can we pack into ourn{\displaystyle n}-bit codeword vector? The received vectors have a maximum energy ofn(P+N){\displaystyle n(P+N)} and therefore must occupy a sphere of radiusn(P+N){\textstyle {\sqrt {n(P+N)}}}. Each codeword sphere has radiusnN{\displaystyle {\sqrt {nN}}}. The volume of ann-dimensional sphere is directly proportional torn{\displaystyle r^{n}}, so the maximum number of uniquely decodeable spheres that can be packed into our sphere with transmission powerP is:

(n(P+N))n/2(nN)n/2=2(n/2)log(1+P/N){\displaystyle {\frac {(n(P+N))^{n/2}}{(nN)^{n/2}}}=2^{(n/2)\log \left(1+P/N\right)}\,\!}

By this argument, the rateR can be no more than12log(1+PN){\displaystyle {\frac {1}{2}}\log \left(1+{\frac {P}{N}}\right)}.

Achievability

[edit]

In this section, we show achievability of the upper bound on the rate from the last section.

A codebook, known to both encoder and decoder, is generated by selecting codewords of lengthn, i.i.d. Gaussian with variancePε{\displaystyle P-\varepsilon } and mean zero. For large n, the empirical variance of the codebook will be very close to the variance of its distribution, thereby avoiding violation of the power constraint probabilistically.

Received messages are decoded to a message in the codebook which is uniquely jointly typical. If there is no such message or if the power constraint is violated, a decoding error is declared.

LetXn(i){\displaystyle X^{n}(i)} denote the codeword for messagei{\displaystyle i}, whileYn{\displaystyle Y^{n}} is, as before the received vector. Define the following three events:

  1. EventU{\displaystyle U}:the power of the received message is larger thanP{\displaystyle P}.
  2. EventV{\displaystyle V}: the transmitted and received codewords are not jointly typical.
  3. EventEj{\displaystyle E_{j}}:(Xn(j),Yn){\displaystyle (X^{n}(j),Y^{n})} is inAε(n){\displaystyle A_{\varepsilon }^{(n)}}, thetypical set whereij{\displaystyle i\neq j}, which is to say that the incorrect codeword is jointly typical with the received vector.

An error therefore occurs ifU{\displaystyle U},V{\displaystyle V} or any of theEi{\displaystyle E_{i}} occur. By the law of large numbers,P(U){\displaystyle P(U)} goes to zero as n approaches infinity, and by the jointAsymptotic Equipartition Property the same applies toP(V){\displaystyle P(V)}. Therefore, for a sufficiently largen{\displaystyle n}, bothP(U){\displaystyle P(U)} andP(V){\displaystyle P(V)} are each less thanε{\displaystyle \varepsilon }. SinceXn(i){\displaystyle X^{n}(i)} andXn(j){\displaystyle X^{n}(j)} are independent forij{\displaystyle i\neq j}, we have thatXn(i){\displaystyle X^{n}(i)} andYn{\displaystyle Y^{n}} are also independent. Therefore, by the joint AEP,P(Ej)=2n(I(X;Y)3ε){\displaystyle P(E_{j})=2^{-n(I(X;Y)-3\varepsilon )}}. This allows us to calculatePe(n){\displaystyle P_{e}^{(n)}}, the probability of error as follows:

Pe(n)P(U)+P(V)+jiP(Ej)ε+ε+ji2n(I(X;Y)3ε)2ε+(2nR1)2n(I(X;Y)3ε)2ε+(23nε)2n(I(X;Y)R)3ε{\displaystyle {\begin{aligned}P_{e}^{(n)}&\leq P(U)+P(V)+\sum _{j\neq i}P(E_{j})\\&\leq \varepsilon +\varepsilon +\sum _{j\neq i}2^{-n(I(X;Y)-3\varepsilon )}\\&\leq 2\varepsilon +(2^{nR}-1)2^{-n(I(X;Y)-3\varepsilon )}\\&\leq 2\varepsilon +(2^{3n\varepsilon })2^{-n(I(X;Y)-R)}\\&\leq 3\varepsilon \end{aligned}}}

Therefore, asn approaches infinity,Pe(n){\displaystyle P_{e}^{(n)}} goes to zero andR<I(X;Y)3ε{\displaystyle R<I(X;Y)-3\varepsilon }. Therefore, there is a code of rate R arbitrarily close to the capacity derived earlier.

Coding theorem converse

[edit]

Here we show that rates above the capacityC=12log(1+PN){\displaystyle C={\frac {1}{2}}\log \left(1+{\frac {P}{N}}\right)} are not achievable.

Suppose that the power constraint is satisfied for a codebook, and further suppose that the messages follow a uniform distribution. LetW{\displaystyle W} be the input messages andW^{\displaystyle {\hat {W}}} the output messages. Thus the information flows as:

WX(n)(W)Y(n)W^{\displaystyle W\longrightarrow X^{(n)}(W)\longrightarrow Y^{(n)}\longrightarrow {\hat {W}}}

Making use ofFano's inequality gives:

H(WW^)1+nRPe(n)=nεn{\displaystyle H(W\mid {\hat {W}})\leq 1+nRP_{e}^{(n)}=n\varepsilon _{n}} whereεn0{\displaystyle \varepsilon _{n}\rightarrow 0} asPe(n)0{\displaystyle P_{e}^{(n)}\rightarrow 0}

LetXi{\displaystyle X_{i}} be the encoded message of codeword indexi. Then:

nR=H(W)=I(W;W^)+H(WW^)I(W;W^)+nεnI(X(n);Y(n))+nεn=h(Y(n))h(Y(n)X(n))+nεn=h(Y(n))h(Z(n))+nεni=1nh(Yi)h(Z(n))+nεni=1nI(Xi;Yi)+nεn{\displaystyle {\begin{aligned}nR&=H(W)\\&=I(W;{\hat {W}})+H(W\mid {\hat {W}})\\&\leq I(W;{\hat {W}})+n\varepsilon _{n}\\&\leq I(X^{(n)};Y^{(n)})+n\varepsilon _{n}\\&=h(Y^{(n)})-h(Y^{(n)}\mid X^{(n)})+n\varepsilon _{n}\\&=h(Y^{(n)})-h(Z^{(n)})+n\varepsilon _{n}\\&\leq \sum _{i=1}^{n}h(Y_{i})-h(Z^{(n)})+n\varepsilon _{n}\\&\leq \sum _{i=1}^{n}I(X_{i};Y_{i})+n\varepsilon _{n}\end{aligned}}}

LetPi{\displaystyle P_{i}} be the average power of the codeword of index i:

Pi=12nRwxi2(w){\displaystyle P_{i}={\frac {1}{2^{nR}}}\sum _{w}x_{i}^{2}(w)\,\!}

where the sum is over all input messagesw{\displaystyle w}.Xi{\displaystyle X_{i}} andZi{\displaystyle Z_{i}} are independent, thus the expectation of the power ofYi{\displaystyle Y_{i}} is, for noise levelN{\displaystyle N}:

E(Yi2)=Pi+N{\displaystyle E(Y_{i}^{2})=P_{i}+N\,\!}

And, ifYi{\displaystyle Y_{i}} is normally distributed, we have that

h(Yi)12log2πe(Pi+N){\displaystyle h(Y_{i})\leq {\frac {1}{2}}\log {2\pi e}(P_{i}+N)\,\!}

Therefore,

nR(h(Yi)h(Zi))+nεn(12log(2πe(Pi+N))12log(2πeN))+nεn=12log(1+PiN)+nεn{\displaystyle {\begin{aligned}nR&\leq \sum (h(Y_{i})-h(Z_{i}))+n\varepsilon _{n}\\&\leq \sum \left({\frac {1}{2}}\log(2\pi e(P_{i}+N))-{\frac {1}{2}}\log(2\pi eN)\right)+n\varepsilon _{n}\\&=\sum {\frac {1}{2}}\log \left(1+{\frac {P_{i}}{N}}\right)+n\varepsilon _{n}\end{aligned}}}

We may apply Jensen's equality tolog(1+x){\displaystyle \log(1+x)}, a concave (downward) function ofx, to get:

1ni=1n12log(1+PiN)12log(1+1ni=1nPiN){\displaystyle {\frac {1}{n}}\sum _{i=1}^{n}{\frac {1}{2}}\log \left(1+{\frac {P_{i}}{N}}\right)\leq {\frac {1}{2}}\log \left(1+{\frac {1}{n}}\sum _{i=1}^{n}{\frac {P_{i}}{N}}\right)\,\!}

Because each codeword individually satisfies the power constraint, the average also satisfies the power constraint. Therefore,

1ni=1nPiN,{\displaystyle {\frac {1}{n}}\sum _{i=1}^{n}{\frac {P_{i}}{N}},\,\!}

which we may apply to simplify the inequality above and get:

12log(1+1ni=1nPiN)12log(1+PN).{\displaystyle {\frac {1}{2}}\log \left(1+{\frac {1}{n}}\sum _{i=1}^{n}{\frac {P_{i}}{N}}\right)\leq {\frac {1}{2}}\log \left(1+{\frac {P}{N}}\right).\,\!}

Therefore, it must be thatR12log(1+PN)+εn{\displaystyle R\leq {\frac {1}{2}}\log \left(1+{\frac {P}{N}}\right)+\varepsilon _{n}}. Therefore,R must be less than a value arbitrarily close to the capacity derived earlier, asεn0{\displaystyle \varepsilon _{n}\rightarrow 0}.

Effects in time domain

[edit]
Zero crossings of a noisy cosine

In serial data communications, the AWGN mathematical model is used to model the timing error caused by randomjitter (RJ).

The graph to the right shows an example of timing errors associated with AWGN. The variable Δt represents the uncertainty in the zero crossing. As the amplitude of the AWGN is increased, thesignal-to-noise ratio decreases. This results in increased uncertainty Δt.[1]

When affected by AWGN, the average number of either positive-going or negative-going zero crossings per second at the output of a narrow bandpass filter when the input is a sine wave is

positive zero crossingssecond=negative zero crossingssecond=f0SNR+1+B212f02SNR+1,{\displaystyle {\begin{aligned}&{\frac {\text{positive zero crossings}}{\text{second}}}={\frac {\text{negative zero crossings}}{\text{second}}}\\[8pt]={}&f_{0}{\sqrt {\frac {{\text{SNR}}+1+{\frac {B^{2}}{12f_{0}^{2}}}}{{\text{SNR}}+1}}},\end{aligned}}}

where

ƒ0 = the center frequency of the filter,
B = the filter bandwidth,
SNR = the signal-to-noise power ratio in linear terms.

Effects in phasor domain

[edit]
AWGN contributions in the phasor domain

In modern communication systems, bandlimited AWGN cannot be ignored. When modeling bandlimited AWGN in thephasor domain, statistical analysis reveals that the amplitudes of the real and imaginary contributions are independent variables which follow theGaussian distribution model. When combined, the resultant phasor's magnitude is aRayleigh-distributed random variable, while the phase is uniformly distributed from 0 to 2π.

The graph to the right shows an example of how bandlimited AWGN can affect a coherent carrier signal. The instantaneous response of the noise vector cannot be precisely predicted, however, its time-averaged response can be statistically predicted. As shown in the graph, we confidently predict that the noise phasor will reside about 38% of the time inside the 1σ circle, about 86% of the time inside the 2σ circle, and about 98% of the time inside the 3σ circle.[1]

See also

[edit]

References

[edit]
  1. ^abMcClaning, Kevin,Radio Receiver Design, Noble Publishing Corporation
Noise (physics and telecommunications)
General
Noise in...
Class of noise
Engineering
terms
Ratios
Related topics
Denoise
methods
General
2D (Image)
Retrieved from "https://en.wikipedia.org/w/index.php?title=Additive_white_Gaussian_noise&oldid=1181981955"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp