Partial least squares (PLS) regression is astatistical method that bears some relation toprincipal components regression and is areduced rank regression;[1] instead of findinghyperplanes of maximumvariance between the response and independent variables, it finds alinear regression model by projecting thepredicted variables and theobservable variables to a new space of maximum covariance (see below). Because both theX andY data are projected to new spaces, the PLS family of methods are known as bilinear factor models. Partial least squares discriminant analysis (PLS-DA) is a variant used when theY is categorical.
PLS is used to find the fundamental relations between twomatrices (X andY), i.e. alatent variable approach to modeling thecovariance structures in these two spaces. A PLS model will try to find the multidimensional direction in theX space that explains the maximum multidimensional variance direction in theY space. PLS regression is particularly suited when the matrix of predictors has more variables than observations, and when there ismulticollinearity amongX values. By contrast, standard regression will fail in these cases (unless it isregularized).
Partial least squares was introduced by the Swedish statisticianHerman O. A. Wold, who then developed it with his son, Svante Wold. An alternative term for PLS isprojection to latent structures,[2][3] but the termpartial least squares is still dominant in many areas. Although the original applications were in the social sciences, PLS regression is today most widely used inchemometrics and related areas. It is also used inbioinformatics,sensometrics,neuroscience, andanthropology.
Core Idea of PLS. The loading vectors in the input and output space are drawn in red (not normalized for better visibility). When increases (independent of), and increase.
We are given a sample ofpaired observations.In the first step, the partial least squares regression searches for the normalized direction, that maximizes the covariance[4]
Note below, the algorithm is denoted in matrix notation.
The general underlying model of multivariate PLS with components is
where
X is an matrix of predictors
Y is an matrix of responses
T andU are matrices that are, respectively, projections ofX (theX score,component orfactor matrix) and projections ofY (theY scores)
P andQ are, respectively, andloading matrices
and matricesE andF are the error terms, assumed to be independent and identically distributed random normal variables.
The decompositions ofX andY are made so as to maximise thecovariance betweenT andU.
Note that this covariance is defined pair by pair: the covariance of columni ofT (lengthn) with the columni ofU (lengthn) is maximized. Additionally, the covariance of the column i ofT with the columnj ofU (with) is zero.
In PLSR, the loadings are thus chosen so that the scores form an orthogonal basis. This is a major difference with PCA where orthogonality is imposed onto loadings (and not the scores).
A number of variants of PLS exist for estimating the factor and loading matricesT, U, P andQ. Most of them construct estimates of the linear regression betweenX andY as. Some PLS algorithms are only appropriate for the case whereY is a column vector, while others deal with the general case of a matrixY. Algorithms also differ on whether they estimate the factor matrixT as an orthogonal (that is,orthonormal) matrix or not.[5][6][7][8][9][10] The final prediction will be the same for all these varieties of PLS, but the components will differ.
PLS is composed of iteratively repeating the following stepsk times (fork components):
finding the directions of maximal covariance in input and output space
performing least squares regression on the input score
PLS1 is a widely used algorithm appropriate for the vectorY case. It estimatesT as an orthonormal matrix.(Caution: thet vectors in the code below may not be normalized appropriately; see talk.)In pseudocode it is expressed below (capital letters are matrices, lower case letters are vectors if they are superscripted and scalars if they are subscripted).
1function PLS1(X, y, ℓ) 2 3, an initial estimate ofw. 4forto 5 6 (note this is a scalar) 7 8 9 (note this is a scalar)10if11,break thefor loop12if131415endfor16defineW to be the matrixwith columns. Do the same to form theP matrix andq vector.171819return
This form of the algorithm does not require centering of the inputX andY, as this is performed implicitly by the algorithm.This algorithm features 'deflation' of the matrixX (subtraction of), but deflation of the vectory is not performed, as it is not necessary (it can be proved that deflatingy yields the same results as not deflating[11]). The user-supplied variablel is the limit on the number of latent factors in the regression; if it equals the rank of the matrixX, the algorithm will yield the least squares regression estimates forB and
Geometric interpretation of the deflation step in the input space
In 2002 a new method was published called orthogonal projections to latent structures (OPLS). In OPLS, continuous variable data is separated into predictive and uncorrelated (orthogonal) information. This leads to improved diagnostics, as well as more easily interpreted visualization. However, these changes only improve the interpretability, not the predictivity, of the PLS models.[12] Similarly, OPLS-DA (Discriminant Analysis) may be applied when working with discrete variables, as in classification and biomarker studies.
Another extension of PLS regression, named L-PLS for its L-shaped matrices, connects 3 related data blocks to improve predictability.[14] In brief, a newZ matrix, with the same number of columns as theX matrix, is added to the PLS regression analysis and may be suitable for including additional background information on the interdependence of the predictor variables.
In 2015 partial least squares was related to a procedure called the three-pass regression filter (3PRF).[15] Supposing the number of observations and variables are large, the 3PRF (and hence PLS) is asymptotically normal for the "best" forecast implied by a linear latent factor model. In stock market data, PLS has been shown to provide accurate out-of-sample forecasts of returns and cash-flow growth.[16]
A PLS version based onsingular value decomposition (SVD) provides a memory efficient implementation that can be used to address high-dimensional problems, such as relating millions of genetic markers to thousands of imaging features in imaging genetics, on consumer-grade hardware.[17]
PLS correlation (PLSC) is another methodology related to PLS regression,[18] which has been used in neuroimaging[18][19][20] and sport science,[21] to quantify the strength of the relationship between data sets. Typically, PLSC divides the data into two blocks (sub-groups) each containing one or more variables, and then usessingular value decomposition (SVD) to establish the strength of any relationship (i.e. the amount of shared information) that might exist between the two component sub-groups.[22] It does this by using SVD to determine the inertia (i.e. the sum of the singular values) of the covariance matrix of the sub-groups under consideration.[22][18]
^Wold, S; Sjöström, M.; Eriksson, L. (2001). "PLS-regression: a basic tool of chemometrics".Chemometrics and Intelligent Laboratory Systems.58 (2):109–130.doi:10.1016/S0169-7439(01)00155-1.S2CID11920190.
^de Jong, S. (1993). "SIMPLS: an alternative approach to partial least squares regression".Chemometrics and Intelligent Laboratory Systems.18 (3):251–263.doi:10.1016/0169-7439(93)85002-X.
^Rannar, S.; Lindgren, F.; Geladi, P.; Wold, S. (1994). "A PLS Kernel Algorithm for Data Sets with Many Variables and Fewer Objects. Part 1: Theory and Algorithm".J. Chemometrics.8 (2):111–125.doi:10.1002/cem.1180080204.S2CID121613293.
^Abdi, H. (2010). "Partial least squares regression and projection on latent structure regression (PLS-Regression)".Wiley Interdisciplinary Reviews: Computational Statistics.2:97–106.doi:10.1002/wics.51.S2CID122685021.
^Sæbøa, S.; Almøya, T.; Flatbergb, A.; Aastveita, A.H.; Martens, H. (2008). "LPLS-regression: a method for prediction and classification under the influence of background information on predictor variables".Chemometrics and Intelligent Laboratory Systems.91 (2):121–132.doi:10.1016/j.chemolab.2007.10.006.
^Kelly, Bryan; Pruitt, Seth (2015-06-01). "The three-pass regression filter: A new approach to forecasting using many predictors".Journal of Econometrics. High Dimensional Problems in Econometrics.186 (2):294–316.doi:10.1016/j.jeconom.2015.02.011.
^abcKrishnan, Anjali; Williams, Lynne J.; McIntosh, Anthony Randal; Abdi, Hervé (May 2011). "Partial Least Squares (PLS) methods for neuroimaging: A tutorial and review".NeuroImage.56 (2):455–475.doi:10.1016/j.neuroimage.2010.07.034.PMID20656037.S2CID8796113.
^abAbdi, Hervé; Williams, Lynne J. (2013), Reisfeld, Brad; Mayeno, Arthur N. (eds.), "Partial Least Squares Methods: Partial Least Squares Correlation and Partial Least Square Regression",Computational Toxicology, vol. 930, Humana Press, pp. 549–579,doi:10.1007/978-1-62703-059-5_23,ISBN978-1-62703-058-8,PMID23086857
Kramer, R. (1998).Chemometric Techniques for Quantitative Analysis. Marcel-Dekker.ISBN978-0-8247-0198-7.
Frank, Ildiko E.; Friedman, Jerome H. (1993). "A Statistical View of Some Chemometrics Regression Tools".Technometrics.35 (2):109–148.doi:10.1080/00401706.1993.10485033.
Haenlein, Michael; Kaplan, Andreas M. (2004). "A Beginner's Guide to Partial Least Squares Analysis".Understanding Statistics.3 (4):283–297.doi:10.1207/s15328031us0304_4.
Henseler, Jörg; Fassott, Georg (2010). "Testing Moderating Effects in PLS Path Models: An Illustration of Available Procedures". In Vinzi, Vincenzo Esposito; Chin, Wynne W.; Henseler, Jörg; Wang, Huiwen (eds.).Handbook of Partial Least Squares: Concepts, Methods and Applications. Springer. pp. 713–735.doi:10.1007/978-3-540-32827-8_31.ISBN978-3-540-32827-8.
Lingjærde, Ole-Christian; Christophersen, Nils (2000). "Shrinkage Structure of Partial Least Squares".Scandinavian Journal of Statistics.27 (3):459–473.doi:10.1111/1467-9469.00201.S2CID121489764.
Tenenhaus, Michel (1998).La Régression PLS: Théorie et Pratique. Paris: Technip.
Rosipal, Roman; Krämer, Nicole (2006). "Overview and Recent Advances in Partial Least Squares". In Saunders, Craig; Grobelnik, Marko; Gunn, Steve; Shawe-Taylor, John (eds.).Subspace, Latent Structure and Feature Selection: Statistical and Optimization Perspectives Workshop, SLSFS 2005, Bohinj, Slovenia, February 23–25, 2005, Revised Selected Papers. Lecture Notes in Computer Science. Springer. pp. 34–51.doi:10.1007/11752790_2.ISBN978-3-540-34138-3.
Helland, Inge S. (1990). "PLS regression and statistical models".Scandinavian Journal of Statistics.17 (2):97–114.JSTOR4616159.
Wold, Herman (1966). "Estimation of principal components and related models by iterative least squares". In Krishnaiaah, P.R. (ed.).Multivariate Analysis. New York: Academic Press. pp. 391–420.
Wold, Herman (1981).The fix-point approach to interdependent systems. Amsterdam: North Holland.
Wold, Herman (1985). "Partial least squares". In Kotz, Samuel; Johnson, Norman L. (eds.).Encyclopedia of statistical sciences. Vol. 6. New York: Wiley. pp. 581–591.
Wold, Svante; Ruhe, Axel; Wold, Herman; Dunn, W.J. (1984). "The collinearity problem in linear regression. the partial least squares (PLS) approach to generalized inverses".SIAM Journal on Scientific and Statistical Computing.5 (3):735–743.doi:10.1137/0905052.
Stone, M.; Brooks, R.J. (1990). "Continuum Regression: Cross-Validated Sequentially Constructed Prediction embracing Ordinary Least Squares, Partial Least Squares and Principal Components Regression".Journal of the Royal Statistical Society, Series B.52 (2):237–269.doi:10.1111/j.2517-6161.1990.tb01786.x.JSTOR2345437.