Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Statistical learning theory

From Wikipedia, the free encyclopedia
Framework for machine learning
This article is about statistical learning in machine learning. For its use in psychology, seeStatistical learning in language acquisition.
See also:Computational learning theory
Part of a series on
Machine learning
anddata mining

Statistical learning theory is a framework formachine learning drawing from the fields ofstatistics andfunctional analysis.[1][2][3] Statistical learning theory deals with thestatistical inference problem of finding a predictive function based on data. Statistical learning theory has led to successful applications in fields such ascomputer vision,speech recognition, andbioinformatics.

Introduction

[edit]

The goals of learning are understanding and prediction. Learning falls into many categories, includingsupervised learning,unsupervised learning,online learning, andreinforcement learning. From the perspective of statistical learning theory, supervised learning is best understood.[4] Supervised learning involves learning from atraining set of data. Every point in the training is an input–output pair, where the input maps to an output. The learning problem consists of inferring the function that maps between the input and the output, such that the learned function can be used to predict the output from future input.

Depending on the type of output, supervised learning problems are either problems ofregression or problems ofclassification. If the output takes a continuous range of values, it is a regression problem. UsingOhm's law as an example, a regression could be performed with voltage as input and current as an output. The regression would find the functional relationship between voltage and current to beR{\displaystyle R}, such thatV=IR{\displaystyle V=IR}Classification problems are those for which the output will be an element from a discrete set of labels. Classification is very common for machine learning applications. Infacial recognition, for instance, a picture of a person's face would be the input, and the output label would be that person's name. The input would be represented by a large multidimensional vector whose elements represent pixels in the picture.

After learning a function based on the training set data, that function is validated on a test set of data, data that did not appear in the training set.

Formal description

[edit]

TakeX{\displaystyle X} to be thevector space of all possible inputs, andY{\displaystyle Y} to be the vector space of all possible outputs. Statistical learning theory takes the perspective that there is some unknownprobability distribution over the product spaceZ=X×Y{\displaystyle Z=X\times Y}, i.e. there exists some unknownp(z)=p(x,y){\displaystyle p(z)=p(\mathbf {x} ,y)}. The training set is made up ofn{\displaystyle n} samples from this probability distribution, and is notatedS={(x1,y1),,(xn,yn)}={z1,,zn}{\displaystyle S=\{(\mathbf {x} _{1},y_{1}),\dots ,(\mathbf {x} _{n},y_{n})\}=\{\mathbf {z} _{1},\dots ,\mathbf {z} _{n}\}}Everyxi{\displaystyle \mathbf {x} _{i}} is an input vector from the training data, andyi{\displaystyle y_{i}} is the output that corresponds to it.

In this formalism, the inference problem consists of finding a functionf:XY{\displaystyle f:X\to Y} such thatf(x)y{\displaystyle f(\mathbf {x} )\sim y}. LetH{\displaystyle {\mathcal {H}}} be a space of functionsf:XY{\displaystyle f:X\to Y} called the hypothesis space. The hypothesis space is the space of functions the algorithm will search through. LetV(f(x),y){\displaystyle V(f(\mathbf {x} ),y)} be theloss function, a metric for the difference between the predicted valuef(x){\displaystyle f(\mathbf {x} )} and the actual valuey{\displaystyle y}. Theexpected risk is defined to beI[f]=X×YV(f(x),y)p(x,y)dxdy{\displaystyle I[f]=\int _{X\times Y}V(f(\mathbf {x} ),y)\,p(\mathbf {x} ,y)\,d\mathbf {x} \,dy}The target function, the best possible functionf{\displaystyle f} that can be chosen, is given by thef{\displaystyle f} that satisfiesf=argminhHI[h]{\displaystyle f=\mathop {\operatorname {argmin} } _{h\in {\mathcal {H}}}I[h]}

Because the probability distributionp(x,y){\displaystyle p(\mathbf {x} ,y)} is unknown, a proxy measure for the expected risk must be used. This measure is based on the training set, a sample from this unknown probability distribution. It is called theempirical riskIS[f]=1ni=1nV(f(xi),yi){\displaystyle I_{S}[f]={\frac {1}{n}}\sum _{i=1}^{n}V(f(\mathbf {x} _{i}),y_{i})}A learning algorithm that chooses the functionfS{\displaystyle f_{S}} that minimizes the empirical risk is calledempirical risk minimization.

Loss functions

[edit]

The choice of loss function is a determining factor on the functionfS{\displaystyle f_{S}} that will be chosen by the learning algorithm. The loss function also affects the convergence rate for an algorithm. It is important for the loss function to beconvex.[5]

Different loss functions are used depending on whether the problem is one of regression or one of classification.

Regression

[edit]

The most common loss function for regression is the square loss function (also known as theL2-norm). This familiar loss function is used inOrdinary Least Squares regression. The form is:V(f(x),y)=(yf(x))2{\displaystyle V(f(\mathbf {x} ),y)=(y-f(\mathbf {x} ))^{2}}

The absolute value loss (also known as theL1-norm) is also sometimes used:V(f(x),y)=|yf(x)|{\displaystyle V(f(\mathbf {x} ),y)=|y-f(\mathbf {x} )|}

Classification

[edit]
Main article:Statistical classification

In some sense the 0-1indicator function is the most natural loss function for classification. It takes the value 0 if the predicted output is the same as the actual output, and it takes the value 1 if the predicted output is different from the actual output. For binary classification withY={1,1}{\displaystyle Y=\{-1,1\}}, this is:V(f(x),y)=θ(yf(x)){\displaystyle V(f(\mathbf {x} ),y)=\theta (-yf(\mathbf {x} ))}whereθ{\displaystyle \theta } is theHeaviside step function.

Regularization

[edit]
This image represents an example of overfitting in machine learning. The red dots represent training set data. The green line represents the true functional relationship, while the blue line shows the learned function, which has been overfitted to the training set data.

In machine learning problems, a major problem that arises is that ofoverfitting. Because learning is a prediction problem, the goal is not to find a function that most closely fits the (previously observed) data, but to find one that will most accurately predict output from future input.Empirical risk minimization runs this risk of overfitting: finding a function that matches the data exactly but does not predict future output well.

Overfitting is symptomatic of unstable solutions; a small perturbation in the training set data would cause a large variation in the learned function. It can be shown that if the stability for the solution can be guaranteed, generalization and consistency are guaranteed as well.[6][7]Regularization can solve the overfitting problem and give the problem stability.

Regularization can be accomplished by restricting the hypothesis spaceH{\displaystyle {\mathcal {H}}}. A common example would be restrictingH{\displaystyle {\mathcal {H}}} to linear functions: this can be seen as a reduction to the standard problem oflinear regression.H{\displaystyle {\mathcal {H}}} could also be restricted to polynomial of degreep{\displaystyle p}, exponentials, or bounded functions onL1. Restriction of the hypothesis space avoids overfitting because the form of the potential functions are limited, and so does not allow for the choice of a function that gives empirical risk arbitrarily close to zero.

One example of regularization isTikhonov regularization. This consists of minimizing1ni=1nV(f(xi),yi)+γfH2{\displaystyle {\frac {1}{n}}\sum _{i=1}^{n}V(f(\mathbf {x} _{i}),y_{i})+\gamma \left\|f\right\|_{\mathcal {H}}^{2}}whereγ{\displaystyle \gamma } is a fixed and positive parameter, the regularization parameter. Tikhonov regularization ensures existence, uniqueness, and stability of the solution.[8]

Bounding empirical risk

[edit]

Consider a binary classifierf:X{0,1}{\displaystyle f:{\mathcal {X}}\to \{0,1\}}. We can applyHoeffding's inequality to bound the probability that the empirical risk deviates from the true risk to be aSub-Gaussian distribution.P(|R^(f)R(f)|ϵ)2e2nϵ2{\displaystyle \mathbb {P} (|{\hat {R}}(f)-R(f)|\geq \epsilon )\leq 2e^{-2n\epsilon ^{2}}}But generally, when we do empirical risk minimization, we are not given a classifier; we must choose it. Therefore, a more useful result is to bound the probability of the supremum of the difference over the whole class.P(supfF|R^(f)R(f)|ϵ)2S(F,n)enϵ2/8ndenϵ2/8{\displaystyle \mathbb {P} {\bigg (}\sup _{f\in {\mathcal {F}}}|{\hat {R}}(f)-R(f)|\geq \epsilon {\bigg )}\leq 2S({\mathcal {F}},n)e^{-n\epsilon ^{2}/8}\approx n^{d}e^{-n\epsilon ^{2}/8}}whereS(F,n){\displaystyle S({\mathcal {F}},n)} is theshattering number andn{\displaystyle n} is the number of samples in your dataset. The exponential term comes from Hoeffding but there is an extra cost of taking the supremum over the whole class, which is the shattering number.

See also

[edit]

References

[edit]
  1. ^Vapnik, Vladimir N. (1995).The Nature of Statistical Learning Theory. New York: Springer.ISBN 978-1-475-72440-0.
  2. ^Hastie, Trevor; Tibshirani, Robert; Friedman, Jerome H. (2009).The Elements of Statistical Learning: Data Mining, Inference, and Prediction. Springer Series in Statistics. New York, NY: Springer.ISBN 978-0-387-84857-0.
  3. ^Mohri, Mehryar; Rostamizadeh, Afshin; Talwalkar, Ameet (2012).Foundations of Machine Learning. US, Massachusetts: MIT Press.ISBN 9780262018258.
  4. ^Tomaso Poggio, Lorenzo Rosasco, et al.Statistical Learning Theory and Applications, 2012,Class 1
  5. ^Rosasco, Lorenzo; De Vito, Ernesto; Caponnetto, Andrea; Piana, Michele; Verri, Alessandro (2004-05-01)."Are Loss Functions All the Same?".Neural Computation.16 (5):1063–1076.doi:10.1162/089976604773135104.hdl:11380/4590.ISSN 0899-7667.PMID 15070510.
  6. ^Vapnik, V.N. and Chervonenkis, A.Y. 1971.On the uniform convergence of relative frequencies of events to their probabilities.Theory of Probability and Its Applications Vol 16, pp 264-280.
  7. ^Mukherjee, S., Niyogi, P. Poggio, T., and Rifkin, R. 2006.Learning theory: stability is sufficient for generalization and necessary and sufficient for consistency of empirical risk minimization.Advances in Computational Mathematics. Vol 25, pp 161-193.
  8. ^Tomaso Poggio, Lorenzo Rosasco, et al.Statistical Learning Theory and Applications, 2012,Class 2
Differentiable computing
General
Hardware
Software libraries
Retrieved from "https://en.wikipedia.org/w/index.php?title=Statistical_learning_theory&oldid=1296238628"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp