Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

White noise

From Wikipedia, the free encyclopedia
Type of signal in signal processing
For other uses, seeWhite Noise.
Thewaveform of aGaussian white noise signal plotted on a graph
Colors of noise

Insignal processing,white noise is a randomsignal having equal intensity at differentfrequencies, giving it a constantpower spectral density.[1] The term is used with this or similar meanings in many scientific and technical disciplines, includingphysics,acoustical engineering,telecommunications, andstatistical forecasting. White noise refers to a statistical model for signals and signal sources, not to any specific signal. White noise draws its name fromwhite light,[2] although light that appears white generally does not have a flat power spectral density over thevisible band.

An image of white noise as perceived through anunconnected analog Television

Indiscrete time, white noise is adiscrete signal whosesamples are regarded as a sequence ofserially uncorrelatedrandom variables with amean of zero and a finitevariance; a single realization of white noise is arandom shock. In some contexts, it is also required that the samples beindependent and have identicalprobability distribution (in other wordsindependent and identically distributed random variables are the simplest representation of white noise).[3] In particular, if each sample has anormal distribution with zero mean, the signal is said to beadditive white Gaussian noise.[4]

The samples of a white noise signal may besequential in time, or arranged along one or more spatial dimensions. Indigital image processing, thepixels of a white noise image are typically arranged in a rectangular grid, and are assumed to be independent random variables withuniform probability distribution over some interval. The concept can be defined also for signals spread over more complicated domains, such as asphere or atorus.

The sound of white noise

Aninfinite-bandwidth white noise signal is a purely theoretical construction. The bandwidth of white noise is limited in practice by the mechanism of noise generation, by the transmission medium and by finite observation capabilities. Thus, random signals are considered white noise if they are observed to have a flat spectrum over the range of frequencies that are relevant to the context. For anaudio signal, the relevant range is the band of audible sound frequencies (between 20 and 20,000Hz). Such a signal is heard by the human ear as a hissing sound, resembling the /h/ sound in a sustained aspiration. On the other hand, thesh sound/ʃ/ inash is a colored noise because it has aformant structure. Inmusic andacoustics, the termwhite noise may be used for any signal that has a similar hissing sound.

In the context ofphylogenetically based statistical methods, the termwhite noise can refer to a lack of phylogenetic pattern in comparative data.[5] In nontechnical contexts, it is sometimes used to mean "random talk without meaningful contents".[6][7]

Statistical properties

[edit]
This sectiondoes notcite anysources. Please helpimprove this section byadding citations to reliable sources. Unsourced material may be challenged andremoved.(January 2022) (Learn how and when to remove this message)
Spectrogram ofpink noise (left) and white noise (right), shown with linear frequency axis (vertical) versus time axis (horizontal)

Any distribution of values is possible (although it must have zeroDC component). Even a binary signal which can only take on the values 1 or -1 will be white if thesequence is statistically uncorrelated. Noise having a continuous distribution, such as anormal distribution, can of course be white.

It is often incorrectly assumed thatGaussian noise (i.e., noise with a Gaussian amplitude distribution – seenormal distribution) necessarily refers to white noise, yet neither property implies the other. Gaussianity refers to the probability distribution with respect to the value, in this context the probability of the signal falling within any particular range of amplitudes, while the term 'white' refers to the way the signal power is distributed (i.e., independently) over time or among frequencies.

One form of white noise is the generalized mean-square derivative of theWiener process orBrownian motion.

A generalization torandom elements on infinite dimensional spaces, such asrandom fields, is thewhite noise measure.

Practical applications

[edit]
This sectionneeds additional citations forverification. Please helpimprove this article byadding citations to reliable sources in this section. Unsourced material may be challenged and removed.(January 2022) (Learn how and when to remove this message)

Music

[edit]

White noise is commonly used in the production ofelectronic music, usually either directly or as an input for a filter to create other types of noise signal. It is used extensively inaudio synthesis, typically to recreate percussive instruments such ascymbals orsnare drums which have high noise content in their frequency domain.[8] A simple example of white noise is a nonexistent radio station (static).

Electronics engineering

[edit]

White noise is also used to obtain theimpulse response of an electrical circuit, in particular ofamplifiers and other audio equipment. It is not used for testing loudspeakers as its spectrum contains too great an amount of high-frequency content.Pink noise, which differs from white noise in that it has equal energy in each octave, is used for testing transducers such as loudspeakers and microphones.

Computing

[edit]

White noise is used as the basis of somerandom number generators. For example,Random.org uses a system of atmospheric antennas to generate random digit patterns from sources that can be well-modeled by white noise.[9]

Tinnitus treatment

[edit]

White noise is a common synthetic noise source used for sound masking by atinnitus masker.[10]White noise machines and other white noise sources are sold as privacy enhancers and sleep aids (seemusic and sleep) and to masktinnitus.[11] The Marpac Sleep-Mate was the first domestic use white noise machine built in 1962 by traveling salesman Jim Buckwalter.[12] Alternatively, the use of an AM radio tuned to unused frequencies ("static") is a simpler and more cost-effective source of white noise.[13] However, white noise generated from a common commercial radio receiver tuned to an unused frequency is extremely vulnerable to being contaminated with spurious signals, such as adjacent radio stations, harmonics from non-adjacent radio stations, electrical equipment in the vicinity of the receiving antenna causing interference, or even atmospheric events such as solar flares and especially lightning.

Work environment

[edit]

The effects of white noise upon cognitive function are mixed. A small study published in 2007 found that white noise background stimulation improves cognitive functioning among secondary students withattention deficit hyperactivity disorder (ADHD), while decreasing performance of non-ADHD students.[14][15] Other work indicates it is effective in improving the mood and performance of workers by masking background office noise,[16] but decreases cognitive performance in complex card sorting tasks.[17]

Similarly, an experiment was carried out on sixty-six healthy participants to observe the benefits of using white noise in a learning environment. The experiment involved the participants identifying different images whilst having different sounds in the background. Overall the experiment showed that white noise does in fact have benefits in relation to learning. The experiments showed that white noise improved the participants' learning abilities and their recognition memory slightly.[18]

Mathematical definitions

[edit]

White noise vector

[edit]

Arandom vector (that is, a random variable with values inRn) is said to be a white noise vector or white random vector if its components each have aprobability distribution with zero mean and finitevariance,[clarification needed] and arestatistically independent: that is, theirjoint probability distribution must be the product of the distributions of the individual components.[19]

A necessary (but,in general, not sufficient) condition for statistical independence of two variables is that they bestatistically uncorrelated; that is, theircovariance is zero. Therefore, thecovariance matrixR of the components of a white noise vectorw withn elements must be ann byndiagonal matrix, where each diagonal elementRii is thevariance of componentwi; and thecorrelation matrix must be then byn identity matrix.

If, in addition to being independent, every variable inw also has anormal distribution with zero mean and the same varianceσ2{\displaystyle \sigma ^{2}},w is said to be a Gaussian white noise vector. In that case, the joint distribution ofw is amultivariate normal distribution; the independence between the variables then implies that the distribution hasspherical symmetry inn-dimensional space. Therefore, anyorthogonal transformation of the vector will result in a Gaussian white random vector. In particular, under most types ofdiscrete Fourier transform, such asFFT andHartley, the transformW ofw will be a Gaussian white noise vector, too; that is, then Fourier coefficients ofw will be independent Gaussian variables with zero mean and the same varianceσ2{\displaystyle \sigma ^{2}}.

Thepower spectrumP of a random vectorw can be defined as the expected value of thesquared modulus of each coefficient of its Fourier transformW, that is,Pi = E(|Wi|2). Under that definition, a Gaussian white noise vector will have a perfectly flat power spectrum, withPi = σ2 for all i.

Ifw is a white random vector, but not a Gaussian one, its Fourier coefficientsWi will not be completely independent of each other; although for largen and common probability distributions the dependencies are very subtle, and their pairwise correlations can be assumed to be zero.

Often the weaker condition statistically uncorrelated is used in the definition of white noise, instead of statistically independent. However, some of the commonly expected properties of white noise (such as flat power spectrum) may not hold for this weaker version. Under this assumption, the stricter version can be referred to explicitly as independent white noise vector.[20]: p.60  Other authors use strongly white and weakly white instead.[21]

An example of a random vector that is Gaussian white noise in the weak but not in the strong sense isx=[x1,x2]{\displaystyle x=[x_{1},x_{2}]} wherex1{\displaystyle x_{1}} is a normal random variable with zero mean, andx2{\displaystyle x_{2}} is equal to+x1{\displaystyle +x_{1}} or tox1{\displaystyle -x_{1}}, with equal probability. These two variables are uncorrelated and individually normally distributed, but they are not jointly normally distributed and are not independent. Ifx{\displaystyle x} is rotated by 45 degrees, its two components will still be uncorrelated, but their distribution will no longer be normal.

In some situations, one may relax the definition by allowing each component of a white random vectorw{\displaystyle w} to have non-zero expected valueμ{\displaystyle \mu }. Inimage processing especially, where samples are typically restricted to positive values, one often takesμ{\displaystyle \mu } to be one half of the maximum sample value. In that case, the Fourier coefficientW0{\displaystyle W_{0}} corresponding to the zero-frequency component (essentially, the average of thewi{\displaystyle w_{i}}) will also have a non-zero expected valueμn{\displaystyle \mu {\sqrt {n}}}; and the power spectrumP{\displaystyle P} will be flat only over the non-zero frequencies.

Discrete-time white noise

[edit]

A discrete-timestochastic processW(n){\displaystyle W(n)} is a generalization of a random vector with a finite number of components to infinitely many components. A discrete-time stochastic processW(n){\displaystyle W(n)} is called white noise if its mean is equal to zero for alln{\displaystyle n} , i.e.E[W(n)]=0{\displaystyle \operatorname {E} [W(n)]=0} and if the autocorrelation functionRW(n)=E[W(k+n)W(k)]{\displaystyle R_{W}(n)=\operatorname {E} [W(k+n)W(k)]} has a nonzero value only forn=0{\displaystyle n=0}, i.e.RW(n)=σ2δ(n){\displaystyle R_{W}(n)=\sigma ^{2}\delta (n)}.[citation needed][clarification needed]

Continuous-time white noise

[edit]

In order to define the notion of white noise in the theory ofcontinuous-time signals, one must replace the concept of a random vector by a continuous-time random signal; that is, a random process that generates a functionw{\displaystyle w} of a real-valued parametert{\displaystyle t}.

Such a process is said to be white noise in the strongest sense if the valuew(t){\displaystyle w(t)} for any timet{\displaystyle t} is a random variable that is statistically independent of its entire history beforet{\displaystyle t}. A weaker definition requires independence only between the valuesw(t1){\displaystyle w(t_{1})} andw(t2){\displaystyle w(t_{2})} at every pair of distinct timest1{\displaystyle t_{1}} andt2{\displaystyle t_{2}}. An even weaker definition requires only that such pairsw(t1){\displaystyle w(t_{1})} andw(t2){\displaystyle w(t_{2})} be uncorrelated.[22] As in the discrete case, some authors adopt the weaker definition for white noise, and use the qualifier independent to refer to either of the stronger definitions. Others use weakly white and strongly white to distinguish between them.

However, a precise definition of these concepts is not trivial, because some quantities that are finite sums in the finite discrete case must be replaced by integrals that may not converge. Indeed, the set of all possible instances of a signalw{\displaystyle w} is no longer a finite-dimensional spaceRn{\displaystyle \mathbb {R} ^{n}}, but an infinite-dimensionalfunction space. Moreover, by any definition a white noise signalw{\displaystyle w} would have to be essentially discontinuous at every point; therefore even the simplest operations onw{\displaystyle w}, like integration over a finite interval, require advanced mathematical machinery.

Some authors[citation needed][clarification needed] require each valuew(t){\displaystyle w(t)} to be a real-valued random variable with expectationμ{\displaystyle \mu } and some finite varianceσ2{\displaystyle \sigma ^{2}}. Then the covarianceE(w(t1)w(t2)){\displaystyle \mathrm {E} (w(t_{1})\cdot w(t_{2}))} between the values at two timest1{\displaystyle t_{1}} andt2{\displaystyle t_{2}} is well-defined: it is zero if the times are distinct, andσ2{\displaystyle \sigma ^{2}} if they are equal. However, by this definition, the integral

W[a,a+r]=aa+rw(t)dt{\displaystyle W_{[a,a+r]}=\int _{a}^{a+r}w(t)\,dt}

over any interval with positive widthr{\displaystyle r} would be simply the width times the expectation:rμ{\displaystyle r\mu }.[clarification needed] This property renders the concept inadequate as a model of white noise signals either in a physical or mathematical sense.[clarification needed]

Therefore, most authors define the signalw{\displaystyle w} indirectly by specifying random values for the integrals ofw(t){\displaystyle w(t)} and|w(t)|2{\displaystyle |w(t)|^{2}} over each interval[a,a+r]{\displaystyle [a,a+r]}. In this approach, however, the value ofw(t){\displaystyle w(t)} at an isolated time cannot be defined as a real-valued random variable[citation needed]. Also the covarianceE(w(t1)w(t2)){\displaystyle \mathrm {E} (w(t_{1})\cdot w(t_{2}))} becomes infinite whent1=t2{\displaystyle t_{1}=t_{2}}; and theautocorrelation functionR(t1,t2){\displaystyle \mathrm {R} (t_{1},t_{2})} must be defined asNδ(t1t2){\displaystyle N\delta (t_{1}-t_{2})}, whereN{\displaystyle N} is some real constant andδ{\displaystyle \delta } is theDirac delta function.[clarification needed]

In this approach, one usually specifies that the integralWI{\displaystyle W_{I}} ofw(t){\displaystyle w(t)} over an intervalI=[a,b]{\displaystyle I=[a,b]} is a real random variable with normal distribution, zero mean, and variance(ba)σ2{\displaystyle (b-a)\sigma ^{2}}; and also that the covarianceE(WIWJ){\displaystyle \mathrm {E} (W_{I}\cdot W_{J})} of the integralsWI{\displaystyle W_{I}},WJ{\displaystyle W_{J}} isrσ2{\displaystyle r\sigma ^{2}}, wherer{\displaystyle r} is the width of the intersectionIJ{\displaystyle I\cap J} of the two intervalsI,J{\displaystyle I,J}. This model is called a Gaussian white noise signal (or process).

In the mathematical field known aswhite noise analysis, a Gaussian white noisew{\displaystyle w} is defined as a stochastic tempered distribution, i.e. a random variable with values in the spaceS(R){\displaystyle {\mathcal {S}}'(\mathbb {R} )} oftempered distributions. Analogous to the case for finite-dimensional random vectors, a probability law on the infinite-dimensional spaceS(R){\displaystyle {\mathcal {S}}'(\mathbb {R} )} can be defined via its characteristic function (existence and uniqueness are guaranteed by an extension of the Bochner–Minlos theorem, which goes under the name Bochner–Minlos–Sazanov theorem); analogously to the case of the multivariate normal distributionXNn(μ,Σ){\displaystyle X\sim {\mathcal {N}}_{n}(\mu ,\Sigma )}, which has characteristic function

kRn:E(eik,X)=eik,μ12k,Σk,{\displaystyle \forall k\in \mathbb {R} ^{n}:\quad \mathrm {E} (\mathrm {e} ^{\mathrm {i} \langle k,X\rangle })=\mathrm {e} ^{\mathrm {i} \langle k,\mu \rangle -{\frac {1}{2}}\langle k,\Sigma k\rangle },}

the white noisew:ΩS(R){\displaystyle w:\Omega \to {\mathcal {S}}'(\mathbb {R} )} must satisfy

φS(R):E(eiw,φ)=e12φ22,{\displaystyle \forall \varphi \in {\mathcal {S}}(\mathbb {R} ):\quad \mathrm {E} (\mathrm {e} ^{\mathrm {i} \langle w,\varphi \rangle })=\mathrm {e} ^{-{\frac {1}{2}}\|\varphi \|_{2}^{2}},}

wherew,φ{\displaystyle \langle w,\varphi \rangle } is the natural pairing of the tempered distributionw(ω){\displaystyle w(\omega )} with the Schwartz functionφ{\displaystyle \varphi } (i.e. we considerφ{\displaystyle \varphi } as a fixed linear function onS(R){\displaystyle {\mathcal {S}}'(\mathbb {R} )} analogous tok{\displaystyle k} above) andφ22=R|φ(x)|2dx{\displaystyle \|\varphi \|_{2}^{2}=\int _{\mathbb {R} }\vert \varphi (x)\vert ^{2}\,\mathrm {d} x}.

Mathematical applications

[edit]

Time series analysis and regression

[edit]

Instatistics andeconometrics one often assumes that an observed series of data values is the sum of the values generated by adeterministiclinear process, depending on certainindependent (explanatory) variables, and on a series of random noise values. Thenregression analysis is used to infer the parameters of the model process from the observed data, e.g. byordinary least squares, and totest the null hypothesis that each of the parameters is zero against the alternative hypothesis that it is non-zero. Hypothesis testing typically assumes that the noise values are mutually uncorrelated with zero mean and have the same Gaussian probability distribution – in other words, that the noise is Gaussian white (not just white). If there is non-zero correlation between the noise values underlying different observations then the estimated model parameters are stillunbiased, but estimates of their uncertainties (such asconfidence intervals) will be biased (not accurate on average). This is also true if the noise isheteroskedastic – that is, if it has different variances for different data points.

Alternatively, in the subset of regression analysis known astime series analysis there are often no explanatory variables other than the past values of the variable being modeled (thedependent variable). In this case the noise process is often modeled as amoving average process, in which the current value of the dependent variable depends on current and past values of a sequential white noise process.

Random vector transformations

[edit]

These two ideas are crucial in applications such aschannel estimation andchannel equalization incommunications andaudio. These concepts are also used indata compression.

In particular, by a suitable linear transformation (acoloring transformation), a white random vector can be used to produce a non-white random vector (that is, a list of random variables) whose elements have a prescribedcovariance matrix. Conversely, a random vector with known covariance matrix can be transformed into a white random vector by a suitablewhitening transformation.

Generation

[edit]

White noise may be generated digitally with adigital signal processor,microprocessor, ormicrocontroller. Generating white noise typically entails feeding an appropriate stream of random numbers to adigital-to-analog converter. The quality of the white noise will depend on the quality of the algorithm used.[23]

Informal use

[edit]

The term is sometimes used as acolloquialism to describe a backdrop of ambient sound, creating an indistinct or seamless commotion. Following are some examples:

  • Chatter from multiple conversations within the acoustics of a confined space.
  • Thepleonasticjargon used by politicians to mask a point that they don't want noticed.[24][25]
  • Music that is disagreeable, harsh, dissonant ordiscordant with nomelody.

The term can also be used metaphorically, as in the novelWhite Noise (1985) byDon DeLillo which explores the symptoms ofmodern culture that came together so as to make it difficult for an individual to actualize their ideas and personality.

See also

[edit]

References

[edit]
  1. ^Carter, Mancini, Bruce, Ron (2009).Op Amps for Everyone. Texas Instruments. pp. 10–11.ISBN 978-0-08-094948-2.{{cite book}}: CS1 maint: multiple names: authors list (link)
  2. ^Stein, Michael L. (1999).Interpolation of Spatial Data: Some Theory for Kriging. Springer Series in Statistics. Springer. p. 40.doi:10.1007/978-1-4612-1494-6.ISBN 978-1-4612-7166-6.white light is approximately an equal mixture of all visible frequencies of light, which was demonstrated by Isaac Newton
  3. ^Stein, Michael L. (1999).Interpolation of Spatial Data: Some Theory for Kriging. Springer Series in Statistics. Springer. p. 40.doi:10.1007/978-1-4612-1494-6.ISBN 978-1-4612-7166-6.The best-known generalized process is white noise, which can be thought of as a continuous time analogue to a sequence of independent and identically distributed observations.
  4. ^Diebold, Frank (2007).Elements of Forecasting (Fourth ed.).
  5. ^Fusco, G; Garland, T. Jr; Hunt, G; Hughes, NC (2011)."Developmental trait evolution in trilobites".Evolution.66 (2):314–329.doi:10.1111/j.1558-5646.2011.01447.x.PMID 22276531.S2CID 14726662.
  6. ^Claire Shipman (2005),Good Morning America: "The politicalrhetoric onSocial Security is white noise." Said onABC'sGood Morning America TV show, January 11, 2005.
  7. ^Don DeLillo (1985),White Noise
  8. ^Clark, Dexxter."Did you know all these white noise secrets? (music production tips)".www.learnhowtoproducemusic.com. Retrieved2022-07-25.
  9. ^O'Connell, Pamela LiCalzi (8 April 2004)."Lottery Numbers and Books With a Voice".The New York Times. Archived fromthe original on 26 July 2009. Retrieved25 July 2022.
  10. ^Jastreboff, P. J. (2000). "Tinnitus Habituation Therapy (THT) and Tinnitus Retraining Therapy (TRT)".Tinnitus Handbook. San Diego: Singular. pp. 357–376.
  11. ^López, HH; Bracha, AS; Bracha, HS (September 2002)."Evidence based complementary intervention for insomnia"(PDF).Hawaii Med J.61 (9): 192, 213.PMID 12422383.
  12. ^Green, Penelope (2018-12-27)."The Sound of Silence".The New York Times.ISSN 0362-4331. Retrieved2021-05-20.
  13. ^Noell, Courtney A; William L Meyerhoff (February 2003). "Tinnitus. Diagnosis and treatment of this elusive symptom".Geriatrics.58 (2):28–34.ISSN 0016-867X.PMID 12596495.
  14. ^Soderlund, Goran; Sverker Sikstrom; Jan Loftesnes; Edmund Sonuga Barke (2010)."The effects of background white noise on memory performance in inattentive school children".Behavioral and Brain Functions.6 (1): 55.doi:10.1186/1744-9081-6-55.PMC 2955636.PMID 20920224.
  15. ^Söderlund, Göran; Sverker Sikström; Andrew Smart (2007). "Listen to the noise: Noise is beneficial for cognitive performance in ADHD".Journal of Child Psychology and Psychiatry.48 (8):840–847.CiteSeerX 10.1.1.452.530.doi:10.1111/j.1469-7610.2007.01749.x.ISSN 0021-9630.PMID 17683456.
  16. ^Loewen, Laura J.; Peter Suedfeld (1992-05-01). "Cognitive and Arousal Effects of Masking Office Noise".Environment and Behavior.24 (3):381–395.Bibcode:1992EnvBe..24..381L.doi:10.1177/0013916592243006.S2CID 144443528.
  17. ^Baker, Mary Anne; Dennis H. Holding (July 1993). "The effects of noise and speech on cognitive task performance".Journal of General Psychology.120 (3):339–355.doi:10.1080/00221309.1993.9711152.ISSN 0022-1309.PMID 8138798.
  18. ^Rausch, V. H. (2014). White noise improves learning by modulating activity in dopaminergic midbrain regions and right superior temporal sulcus . Journal of cognitive neuroscience, 1469-1480
  19. ^ Jeffrey A. Fessler (1998),On Transformations of Random Vectors. Technical report 314, Dept. of Electrical Engineering and Computer Science, Univ. of Michigan. (PDF)
  20. ^Eric Zivot and Jiahui Wang (2006),Modeling Financial Time Series with S-PLUS. Second Edition. (PDF)
  21. ^Francis X. Diebold (2007),Elements of Forecasting, 4th edition. (PDF)
  22. ^White noise processArchived 2016-09-11 at theWayback Machine. By Econterms via About.com. Accessed on 2013-02-12.
  23. ^Matt Donadio."How to Generate White Gaussian Noise"(PDF). Archived fromthe original(PDF) on 2021-02-24. Retrieved2012-09-19.
  24. ^white noise, Merriam-Webster, retrieved2022-05-06
  25. ^en/articles/white_noise_for_sleep_how_does_it_work/en/authors/3."White Noise for Sleep: How Does It Work?".somnorium.com. Archived fromthe original on 3 Oct 2025. Retrieved2025-10-03.{{cite web}}: CS1 maint: numeric names: authors list (link)

External links

[edit]
Wikimedia Commons has media related toWhite noise.
Noise (physics and telecommunications)
General
Noise in...
Class of noise
Engineering
terms
Ratios
Related topics
Denoise
methods
General
2D (Image)
Discrete time
Continuous time
Both
Fields and other
Time series models
Financial models
Actuarial models
Queueing models
Properties
Limit theorems
Inequalities
Tools
Disciplines
Retrieved from "https://en.wikipedia.org/w/index.php?title=White_noise&oldid=1318685657"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp