Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Cross-correlation

From Wikipedia, the free encyclopedia
Covariance and correlation
Part of a series onStatistics
Correlation and covariance
Visual comparison ofconvolution, cross-correlation andautocorrelation. For the operations involving functionf, and assuming the height off is 1.0, the value of the result at 5 different points is indicated by the shaded area below each point. Also, the vertical symmetry off is the reasonfg{\displaystyle f*g} andfg{\displaystyle f\star g} are identical in this example.

Insignal processing,cross-correlation is ameasure of similarity of two series as a function of the displacement of one relative to the other. This is also known as aslidingdot product orsliding inner-product. It is commonly used for searching a long signal for a shorter, known feature. It has applications inpattern recognition,single particle analysis,electron tomography,averaging,cryptanalysis, andneurophysiology. The cross-correlation is similar in nature to theconvolution of two functions. In anautocorrelation, which is the cross-correlation of a signal with itself, there will always be a peak at a lag of zero, and its size will be the signal energy.

Inprobability andstatistics, the termcross-correlations refers to thecorrelations between the entries of tworandom vectorsX{\displaystyle \mathbf {X} } andY{\displaystyle \mathbf {Y} }, while thecorrelations of a random vectorX{\displaystyle \mathbf {X} } are the correlations between the entries ofX{\displaystyle \mathbf {X} } itself, those forming thecorrelation matrix ofX{\displaystyle \mathbf {X} }. If each ofX{\displaystyle \mathbf {X} } andY{\displaystyle \mathbf {Y} } is a scalar random variable which is realized repeatedly in atime series, then the correlations of the various temporal instances ofX{\displaystyle \mathbf {X} } are known asautocorrelations ofX{\displaystyle \mathbf {X} }, and the cross-correlations ofX{\displaystyle \mathbf {X} } withY{\displaystyle \mathbf {Y} } across time are temporal cross-correlations. In probability and statistics, the definition of correlation always includes a standardising factor in such a way that correlations have values between −1 and +1.

IfX{\displaystyle X} andY{\displaystyle Y} are twoindependentrandom variables withprobability density functionsf{\displaystyle f} andg{\displaystyle g}, respectively, then the probability density of the differenceYX{\displaystyle Y-X} is formally given by the cross-correlation (in the signal-processing sense)fg{\displaystyle f\star g}; however, this terminology is not used in probability and statistics. In contrast, theconvolutionfg{\displaystyle f*g} (equivalent to the cross-correlation off(t){\displaystyle f(t)} andg(t){\displaystyle g(-t)}) gives the probability density function of the sumX+Y{\displaystyle X+Y}.

Cross-correlation of deterministic signals

[edit]

For continuous functionsf{\displaystyle f} andg{\displaystyle g}, the cross-correlation is defined as:[1][2][3](fg)(τ) f(t)¯g(t+τ)dt{\displaystyle (f\star g)(\tau )\ \triangleq \int _{-\infty }^{\infty }{\overline {f(t)}}g(t+\tau )\,dt}which is equivalent to(fg)(τ) f(tτ)¯g(t)dt{\displaystyle (f\star g)(\tau )\ \triangleq \int _{-\infty }^{\infty }{\overline {f(t-\tau )}}g(t)\,dt}wheref(t)¯{\displaystyle {\overline {f(t)}}} denotes thecomplex conjugate off(t){\displaystyle f(t)}, andτ{\displaystyle \tau } is calleddisplacement orlag.

For highly-correlatedf{\displaystyle f} andg{\displaystyle g} which have a maximum cross-correlation at a particularτ{\displaystyle \tau }, a feature inf{\displaystyle f} att{\displaystyle t} also occurs later ing{\displaystyle g} att+τ{\displaystyle t+\tau }, henceg{\displaystyle g} could be described tolagf{\displaystyle f} byτ{\displaystyle \tau }.

Iff{\displaystyle f} andg{\displaystyle g} are both continuous periodic functions of periodT{\displaystyle T}, the integration from{\displaystyle -\infty } to{\displaystyle \infty } is replaced by integration over any interval[t0,t0+T]{\displaystyle [t_{0},t_{0}+T]} of lengthT{\displaystyle T}:(fg)(τ) t0t0+Tf(t)¯g(t+τ)dt{\displaystyle (f\star g)(\tau )\ \triangleq \int _{t_{0}}^{t_{0}+T}{\overline {f(t)}}g(t+\tau )\,dt}which is equivalent to(fg)(τ) t0t0+Tf(tτ)¯g(t)dt{\displaystyle (f\star g)(\tau )\ \triangleq \int _{t_{0}}^{t_{0}+T}{\overline {f(t-\tau )}}g(t)\,dt}Similarly, for discrete functions, the cross-correlation is defined as:[4][5](fg)[n] m=f[m]¯g[m+n]{\displaystyle (f\star g)[n]\ \triangleq \sum _{m=-\infty }^{\infty }{\overline {f[m]}}g[m+n]}which is equivalent to:(fg)[n] m=f[mn]¯g[m]{\displaystyle (f\star g)[n]\ \triangleq \sum _{m=-\infty }^{\infty }{\overline {f[m-n]}}g[m]}For finite discrete functionsf,gCN{\displaystyle f,g\in \mathbb {C} ^{N}}, the (circular) cross-correlation is defined as:[6](fg)[n] m=0N1f[m]¯g[(m+n)mod N]{\displaystyle (f\star g)[n]\ \triangleq \sum _{m=0}^{N-1}{\overline {f[m]}}g[(m+n)_{{\text{mod}}~N}]}which is equivalent to:(fg)[n] m=0N1f[(mn)mod N]¯g[m]{\displaystyle (f\star g)[n]\ \triangleq \sum _{m=0}^{N-1}{\overline {f[(m-n)_{{\text{mod}}~N}]}}g[m]}For finite discrete functionsfCN{\displaystyle f\in \mathbb {C} ^{N}},gCM{\displaystyle g\in \mathbb {C} ^{M}}, the kernel cross-correlation is defined as:[7](fg)[n] m=0N1f[m]¯Kg[(m+n)mod N]{\displaystyle (f\star g)[n]\ \triangleq \sum _{m=0}^{N-1}{\overline {f[m]}}K_{g}[(m+n)_{{\text{mod}}~N}]}whereKg=[k(g,T0(g)),k(g,T1(g)),,k(g,TN1(g))]{\displaystyle K_{g}=[k(g,T_{0}(g)),k(g,T_{1}(g)),\dots ,k(g,T_{N-1}(g))]} is a vector of kernel functionsk(,):CM×CMR{\displaystyle k(\cdot ,\cdot )\colon \mathbb {C} ^{M}\times \mathbb {C} ^{M}\to \mathbb {R} } andTi():CMCM{\displaystyle T_{i}(\cdot )\colon \mathbb {C} ^{M}\to \mathbb {C} ^{M}} is anaffine transform.

Specifically,Ti(){\displaystyle T_{i}(\cdot )} can be circular translation transform, rotation transform, or scale transform, etc. The kernel cross-correlation extends cross-correlation from linear space to kernel space. Cross-correlation is equivariant to translation; kernel cross-correlation is equivariant to any affine transforms, including translation, rotation, and scale, etc.

Explanation

[edit]

As an example, consider two real valued functionsf{\displaystyle f} andg{\displaystyle g} differing only by an unknown shift along the x-axis. One can use the cross-correlation to find how muchg{\displaystyle g} must be shifted along the x-axis to make it identical tof{\displaystyle f}. The formula essentially slides theg{\displaystyle g} function along the x-axis, calculating the integral of their product at each position. When the functions match, the value of(fg){\displaystyle (f\star g)} is maximized. This is because when peaks (positive areas) are aligned, they make a large contribution to the integral. Similarly, when troughs (negative areas) align, they also make a positive contribution to the integral because the product of two negative numbers is positive.

Animation of how cross-correlation is calculated. The left graph shows a green function G that is phase-shifted relative to function F by a time displacement of 𝜏. The middle graph shows the function F and the phase-shifted G represented together as aLissajous curve. Integrating F multiplied by the phase-shifted G produces the right graph, the cross-correlation across all values of 𝜏.

Withcomplex-valued functionsf{\displaystyle f} andg{\displaystyle g}, taking theconjugate off{\displaystyle f} ensures that aligned peaks (or aligned troughs) with imaginary components will contribute positively to the integral.

Ineconometrics, lagged cross-correlation is sometimes referred to as cross-autocorrelation.[8]: p. 74 

Properties

[edit]

Cross-correlation of random vectors

[edit]
Main article:Cross-correlation matrix

Definition

[edit]

Forrandom vectorsX=(X1,,Xm){\displaystyle \mathbf {X} =(X_{1},\ldots ,X_{m})} andY=(Y1,,Yn){\displaystyle \mathbf {Y} =(Y_{1},\ldots ,Y_{n})}, each containingrandom elements whoseexpected value andvariance exist, thecross-correlation matrix ofX{\displaystyle \mathbf {X} } andY{\displaystyle \mathbf {Y} } is defined by[10]: p.337 RXY E[XY]{\displaystyle \operatorname {R} _{\mathbf {X} \mathbf {Y} }\triangleq \ \operatorname {E} \left[\mathbf {X} \mathbf {Y} \right]}and has dimensionsm×n{\displaystyle m\times n}. Written component-wise:RXY=[E[X1Y1]E[X1Y2]E[X1Yn]E[X2Y1]E[X2Y2]E[X2Yn]E[XmY1]E[XmY2]E[XmYn]]{\displaystyle \operatorname {R} _{\mathbf {X} \mathbf {Y} }={\begin{bmatrix}\operatorname {E} [X_{1}Y_{1}]&\operatorname {E} [X_{1}Y_{2}]&\cdots &\operatorname {E} [X_{1}Y_{n}]\\\\\operatorname {E} [X_{2}Y_{1}]&\operatorname {E} [X_{2}Y_{2}]&\cdots &\operatorname {E} [X_{2}Y_{n}]\\\\\vdots &\vdots &\ddots &\vdots \\\\\operatorname {E} [X_{m}Y_{1}]&\operatorname {E} [X_{m}Y_{2}]&\cdots &\operatorname {E} [X_{m}Y_{n}]\end{bmatrix}}}The random vectorsX{\displaystyle \mathbf {X} } andY{\displaystyle \mathbf {Y} } need not have the same dimension, and either might be a scalar value.WhereE{\displaystyle \operatorname {E} } is theexpectation value.

Example

[edit]

For example, ifX=(X1,X2,X3){\displaystyle \mathbf {X} =\left(X_{1},X_{2},X_{3}\right)} andY=(Y1,Y2){\displaystyle \mathbf {Y} =\left(Y_{1},Y_{2}\right)} are random vectors, thenRXY{\displaystyle \operatorname {R} _{\mathbf {X} \mathbf {Y} }} is a3×2{\displaystyle 3\times 2} matrix whose(i,j){\displaystyle (i,j)}-th entry isE[XiYj]{\displaystyle \operatorname {E} [X_{i}Y_{j}]}.

Definition for complex random vectors

[edit]

IfZ=(Z1,,Zm){\displaystyle \mathbf {Z} =(Z_{1},\ldots ,Z_{m})} andW=(W1,,Wn){\displaystyle \mathbf {W} =(W_{1},\ldots ,W_{n})} arecomplex random vectors, each containing random variables whose expected value and variance exist, the cross-correlation matrix ofZ{\displaystyle \mathbf {Z} } andW{\displaystyle \mathbf {W} } is defined byRZW E[ZWH]{\displaystyle \operatorname {R} _{\mathbf {Z} \mathbf {W} }\triangleq \ \operatorname {E} [\mathbf {Z} \mathbf {W} ^{\rm {H}}]}whereH{\displaystyle {}^{\rm {H}}} denotesHermitian transposition.

Cross-correlation of stochastic processes

[edit]

Intime series analysis andstatistics, the cross-correlation of a pair ofrandom process is the correlation between values of the processes at different times, as a function of the two times. Let(Xt,Yt){\displaystyle (X_{t},Y_{t})} be a pair of random processes, andt{\displaystyle t} be any point in time (t{\displaystyle t} may be aninteger for adiscrete-time process or areal number for acontinuous-time process). ThenXt{\displaystyle X_{t}} is the value (orrealization) produced by a given run of the process at timet{\displaystyle t}.

Cross-correlation function

[edit]

Suppose that the process has meansμX(t){\displaystyle \mu _{X}(t)} andμY(t){\displaystyle \mu _{Y}(t)} and variancesσX2(t){\displaystyle \sigma _{X}^{2}(t)} andσY2(t){\displaystyle \sigma _{Y}^{2}(t)} at timet{\displaystyle t}, for eacht{\displaystyle t}. Then the definition of the cross-correlation between timest1{\displaystyle t_{1}} andt2{\displaystyle t_{2}} is[10]: p.392 RXY(t1,t2) E[Xt1Yt2¯]{\displaystyle \operatorname {R} _{XY}(t_{1},t_{2})\triangleq \ \operatorname {E} \left[X_{t_{1}}{\overline {Y_{t_{2}}}}\right]}whereE{\displaystyle \operatorname {E} } is theexpected value operator. Note that this expression may be not defined.

Cross-covariance function

[edit]

Subtracting the mean before multiplication yields the cross-covariance between timest1{\displaystyle t_{1}} andt2{\displaystyle t_{2}}:[10]: p.392 KXY(t1,t2) E[(Xt1μX(t1))(Yt2μY(t2))¯]{\displaystyle \operatorname {K} _{XY}(t_{1},t_{2})\triangleq \ \operatorname {E} \left[\left(X_{t_{1}}-\mu _{X}(t_{1})\right){\overline {(Y_{t_{2}}-\mu _{Y}(t_{2}))}}\right]}Note that this expression is not well-defined for all time series or processes, because the mean or variance may not exist.

Definition for wide-sense stationary stochastic process

[edit]

Let(Xt,Yt){\displaystyle (X_{t},Y_{t})} represent a pair ofstochastic processes that arejointly wide-sense stationary. Then thecross-covariance function and the cross-correlation function are given as follows.

Cross-correlation function

[edit]

RXY(τ) E[XtYt+τ¯]{\displaystyle \operatorname {R} _{XY}(\tau )\triangleq \ \operatorname {E} \left[X_{t}{\overline {Y_{t+\tau }}}\right]} or equivalentlyRXY(τ)=E[XtτYt¯]{\displaystyle \operatorname {R} _{XY}(\tau )=\operatorname {E} \left[X_{t-\tau }{\overline {Y_{t}}}\right]}

Cross-covariance function

[edit]

KXY(τ) E[(XtμX)(Yt+τμY)¯]{\displaystyle \operatorname {K} _{XY}(\tau )\triangleq \ \operatorname {E} \left[\left(X_{t}-\mu _{X}\right){\overline {\left(Y_{t+\tau }-\mu _{Y}\right)}}\right]} or equivalentlyKXY(τ)=E[(XtτμX)(YtμY)¯]{\displaystyle \operatorname {K} _{XY}(\tau )=\operatorname {E} \left[\left(X_{t-\tau }-\mu _{X}\right){\overline {\left(Y_{t}-\mu _{Y}\right)}}\right]}whereμX{\displaystyle \mu _{X}} andσX{\displaystyle \sigma _{X}} are the mean and standard deviation of the process(Xt){\displaystyle (X_{t})}, which are constant over time due to stationarity; and similarly for(Yt){\displaystyle (Y_{t})}, respectively.E[ ]{\displaystyle \operatorname {E} [\ ]} indicates theexpected value. That the cross-covariance and cross-correlation are independent oft{\displaystyle t} is precisely the additional information (beyond being individually wide-sense stationary) conveyed by the requirement that(Xt,Yt){\displaystyle (X_{t},Y_{t})} arejointly wide-sense stationary.

The cross-correlation of a pair of jointlywide sense stationarystochastic processes can be estimated by averaging the product of samples measured from one process and samples measured from the other (and its time shifts). The samples included in the average can be an arbitrary subset of all the samples in the signal (e.g., samples within a finite time window or asub-sampling[which?] of one of the signals). For a large number of samples, the average converges to the true cross-correlation.

Normalization

[edit]

It is common practice in some disciplines (e.g. statistics andtime series analysis) to normalize the cross-correlation function to get a time-dependentPearson correlation coefficient. However, in other disciplines (e.g. engineering) the normalization is usually dropped and the terms "cross-correlation" and "cross-covariance" are used interchangeably.

The definition of the normalized cross-correlation of a stochastic process isρXX(t1,t2)=KXX(t1,t2)σX(t1)σX(t2)=E[(Xt1μt1)(Xt2μt2)¯]σX(t1)σX(t2){\displaystyle \rho _{XX}(t_{1},t_{2})={\frac {\operatorname {K} _{XX}(t_{1},t_{2})}{\sigma _{X}(t_{1})\sigma _{X}(t_{2})}}={\frac {\operatorname {E} \left[\left(X_{t_{1}}-\mu _{t_{1}}\right){\overline {\left(X_{t_{2}}-\mu _{t_{2}}\right)}}\right]}{\sigma _{X}(t_{1})\sigma _{X}(t_{2})}}}If the functionρXX{\displaystyle \rho _{XX}} is well-defined, its value must lie in the range[1,1]{\displaystyle [-1,1]}, with 1 indicating perfect correlation and −1 indicating perfectanti-correlation.

For jointly wide-sense stationary stochastic processes, the definition isρXY(τ)=KXY(τ)σXσY=E[(XtμX)(Yt+τμY)¯]σXσY{\displaystyle \rho _{XY}(\tau )={\frac {\operatorname {K} _{XY}(\tau )}{\sigma _{X}\sigma _{Y}}}={\frac {\operatorname {E} \left[\left(X_{t}-\mu _{X}\right){\overline {\left(Y_{t+\tau }-\mu _{Y}\right)}}\right]}{\sigma _{X}\sigma _{Y}}}}The normalization is important both because the interpretation of the autocorrelation as a correlation provides a scale-free measure of the strength ofstatistical dependence, and because the normalization has an effect on the statistical properties of the estimated autocorrelations.

Properties

[edit]

Symmetry property

[edit]

For jointly wide-sense stationary stochastic processes, the cross-correlation function has the following symmetry property:[11]: p.173 RXY(t1,t2)=RYX(t2,t1)¯{\displaystyle \operatorname {R} _{XY}(t_{1},t_{2})={\overline {\operatorname {R} _{YX}(t_{2},t_{1})}}}Respectively for jointly WSS processes:RXY(τ)=RYX(τ)¯{\displaystyle \operatorname {R} _{XY}(\tau )={\overline {\operatorname {R} _{YX}(-\tau )}}}

Time delay analysis

[edit]

Cross-correlations are useful for determining the time delay between two signals, e.g., for determining time delays for the propagation of acoustic signals across a microphone array.[12][13][clarification needed] After calculating the cross-correlation between the two signals, the maximum (or minimum if the signals are negatively correlated) of the cross-correlation function indicates the point in time where the signals are best aligned; i.e., the time delay between the two signals is determined by the argument of the maximum, orarg max of the cross-correlation, as inτdelay=argmaxtR((fg)(t)){\displaystyle \tau _{\mathrm {delay} }={\underset {t\in \mathbb {R} }{\operatorname {arg\,max} }}((f\star g)(t))}Terminology in image processing

Zero-normalized cross-correlation (ZNCC)

[edit]

Forimage-processing applications in which the brightness of the image and template can vary due to lighting and exposure conditions, the images can be first normalized. This is typically done at every step by subtracting the mean and dividing by thestandard deviation. That is, the cross-correlation of a templatet(x,y){\displaystyle t(x,y)} with a subimagef(x,y){\displaystyle f(x,y)} is

1nσfσtx,y(f(x,y)μf)(t(x,y)μt){\displaystyle {\frac {1}{n\sigma _{f}\sigma _{t}}}\sum _{x,y}\left(f(x,y)-\mu _{f}\right)\left(t(x,y)-\mu _{t}\right)}

wheren{\displaystyle n} is the number of pixels int(x,y){\displaystyle t(x,y)} andf(x,y){\displaystyle f(x,y)},μf{\displaystyle \mu _{f}} is the average off{\displaystyle f} andσf{\displaystyle \sigma _{f}} isstandard deviation off{\displaystyle f}.

Infunctional analysis terms, this can be thought of as the dot product of twonormalized vectors. That is, ifF(x,y)=f(x,y)μf{\displaystyle F(x,y)=f(x,y)-\mu _{f}}andT(x,y)=t(x,y)μt{\displaystyle T(x,y)=t(x,y)-\mu _{t}}then the above sum is equal toFF,TT{\displaystyle \left\langle {\frac {F}{\|F\|}},{\frac {T}{\|T\|}}\right\rangle }where,{\displaystyle \langle \cdot ,\cdot \rangle } is theinner product and{\displaystyle \|\cdot \|} is theL² norm.Cauchy–Schwarz then implies that ZNCC has a range of[1,1]{\displaystyle [-1,1]}.

Thus, iff{\displaystyle f} andt{\displaystyle t} are real matrices, their normalized cross-correlation equals the cosine of the angle between the unit vectorsF{\displaystyle F} andT{\displaystyle T}, being thus1{\displaystyle 1} if and only ifF{\displaystyle F} equalsT{\displaystyle T} multiplied by a positive scalar.

Normalized correlation is one of the methods used fortemplate matching, a process used for finding instances of a pattern or object within an image. It is also the 2-dimensional version ofPearson product-moment correlation coefficient.

Normalized cross-correlation (NCC)

[edit]

NCC is similar to ZNCC with the only difference of not subtracting the local mean value of intensities:1nσfσtx,yf(x,y)t(x,y){\displaystyle {\frac {1}{n\sigma _{f}\sigma _{t}}}\sum _{x,y}f(x,y)t(x,y)}

Nonlinear systems

[edit]

Caution must be applied when using cross correlation function which assumes Gaussian variance for nonlinear systems. In certain circumstances, which depend on the properties of the input, cross correlation between the input and output of a system with nonlinear dynamics can be completely blind to certain nonlinear effects.[14] This problem arises because some quadratic moments can equal zero and this can incorrectly suggest that there is little "correlation" (in the sense of statistical dependence) between two signals, when in fact the two signals are strongly related by nonlinear dynamics.

See also

[edit]

References

[edit]
  1. ^Bracewell, R. "Pentagram Notation for Cross Correlation." The Fourier Transform and Its Applications. New York: McGraw-Hill, pp. 46 and 243, 1965.
  2. ^Papoulis, A. The Fourier Integral and Its Applications. New York: McGraw-Hill, pp. 244–245 and 252-253, 1962.
  3. ^Weisstein, Eric W. "Cross-Correlation." From MathWorld--A Wolfram Web Resource.http://mathworld.wolfram.com/Cross-Correlation.html
  4. ^Rabiner, L.R.; Schafer, R.W. (1978).Digital Processing of Speech Signals. Signal Processing Series. Upper Saddle River, NJ: Prentice Hall. pp. 147–148.ISBN 0132136031.
  5. ^Rabiner, Lawrence R.; Gold, Bernard (1975).Theory and Application of Digital Signal Processing. Englewood Cliffs, NJ: Prentice-Hall. pp. 401.ISBN 0139141014.
  6. ^Wang, Chen (2019).Kernel learning for visual perception, Chapter 2.2.1 (Doctoral thesis). Nanyang Technological University, Singapore. pp. 17–18.doi:10.32657/10220/47835.hdl:10356/105527.
  7. ^Wang, Chen; Zhang, Le; Yuan, Junsong; Xie, Lihua (2018)."Kernel Cross-Correlator".Proceedings of the AAAI Conference on Artificial Intelligence. The Thirty-second AAAI Conference On Artificial Intelligence.32. Association for the Advancement of Artificial Intelligence:4179–4186.doi:10.1609/aaai.v32i1.11710.S2CID 3544911.
  8. ^Campbell; Lo; MacKinlay (1996).The Econometrics of Financial Markets. NJ: Princeton University Press.ISBN 0691043019.
  9. ^Kapinchev, Konstantin; Bradu, Adrian; Barnes, Frederick; Podoleanu, Adrian (2015). "GPU implementation of cross-correlation for image generation in real time".2015 9th International Conference on Signal Processing and Communication Systems (ICSPCS). pp. 1–6.doi:10.1109/ICSPCS.2015.7391783.ISBN 978-1-4673-8118-5.S2CID 17108908.
  10. ^abcGubner, John A. (2006).Probability and Random Processes for Electrical and Computer Engineers. Cambridge University Press.ISBN 978-0-521-86470-1.
  11. ^Kun Il Park, Fundamentals of Probability and Stochastic Processes with Applications to Communications, Springer, 2018, 978-3-319-68074-3
  12. ^Rhudy, Matthew; Brian Bucci; Jeffrey Vipperman; Jeffrey Allanach; Bruce Abraham (November 2009).Microphone Array Analysis Methods Using Cross-Correlations. Proceedings of 2009 ASME International Mechanical Engineering Congress, Lake Buena Vista, FL. pp. 281–288.doi:10.1115/IMECE2009-10798.ISBN 978-0-7918-4388-8.
  13. ^Rhudy, Matthew (November 2009).Real Time Implementation of a Military Impulse Classifier (MS thesis). University of Pittsburgh.
  14. ^Billings, S. A. (2013).Nonlinear System Identification: NARMAX Methods in the Time, Frequency, and Spatio-Temporal Domains. Wiley.ISBN 978-1-118-53556-1.

Further reading

[edit]

External links

[edit]
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Retrieved from "https://en.wikipedia.org/w/index.php?title=Cross-correlation&oldid=1288065455"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp