Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Perceptron

From Wikipedia, the free encyclopedia
Algorithm for supervised learning of binary classifiers
"Perceptrons" redirects here. For the 1969 book, seePerceptrons (book).
Part of a series on
Machine learning
anddata mining

Inmachine learning, theperceptron is an algorithm forsupervised learning ofbinary classifiers. A binary classifier is a function that can decide whether or not an input, represented by a vector of numbers, belongs to some specific class.[1] It is a type oflinear classifier, i.e. a classification algorithm that makes its predictions based on alinear predictor function combining a set ofweights with thefeature vector.

History

[edit]
See also:History of artificial intelligence § Perceptrons
Mark I Perceptron machine, the first implementation of the perceptron algorithm. It was connected to a camera with 20×20cadmium sulfidephotocells to make a 400-pixel image. The main visible feature is the sensory-to-association plugboard, which sets different combinations of input features. To the right are arrays ofpotentiometers that implemented the adaptive weights.[2]: 213 
The Mark 1 Perceptron, being adjusted by Charles Wightman (Mark I Perceptron project engineer).[3] Sensory units at left, association units in center, and control panel and response units at far right. The sensory-to-association plugboard is behind the closed panel to the right of the operator. The letter "C" on the front panel is a display of the current state of the sensory input.[4]

The artificial neuron network was invented in 1943 byWarren McCulloch andWalter Pitts inA logical calculus of the ideas immanent in nervous activity.[5]

In 1957,Frank Rosenblatt was at theCornell Aeronautical Laboratory. He simulated the perceptron on anIBM 704.[6][7] Later, he obtained funding by the Information Systems Branch of the United StatesOffice of Naval Research and theRome Air Development Center, to build a custom-made computer, theMark I Perceptron. It was first publicly demonstrated on 23 June 1960.[8] The machine was "part of a previously secret four-year NPIC [the US'National Photographic Interpretation Center] effort from 1963 through 1966 to develop this algorithm into a useful tool for photo-interpreters".[9]

Rosenblatt described the details of the perceptron in a 1958 paper.[10] His organization of a perceptron is constructed of three kinds of cells ("units"): AI, AII, R, which stand for "projection", "association" and "response". He presented at the first international symposium on AI,Mechanisation of Thought Processes, which took place in 1958 November.[11]

Rosenblatt's project was funded under Contract Nonr-401(40) "Cognitive Systems Research Program", which lasted from 1959 to 1970,[12] and Contract Nonr-2381(00) "Project PARA" ("PARA" means "Perceiving and Recognition Automata"), which lasted from 1957[6] to 1963.[13]

In 1959, the Institute for Defense Analysis awarded his group a $10,000 contract. By September 1961, the ONR awarded further $153,000 worth of contracts, with $108,000 committed for 1962.[14]

The ONR research manager, Marvin Denicoff, stated that ONR, instead ofARPA, funded the Perceptron project, because the project was unlikely to produce technological results in the near or medium term. Funding from ARPA go up to the order of millions dollars, while from ONR are on the order of 10,000 dollars. Meanwhile, the head ofIPTO at ARPA,J.C.R. Licklider, was interested in 'self-organizing', 'adaptive' and other biologically-inspired methods in the 1950s; but by the mid-1960s he was openly critical of these, including the perceptron. Instead he strongly favored the logical AI approach ofSimon andNewell.[15]

Mark I Perceptron machine

[edit]
Main article:Mark I Perceptron
Organization of a biological brain and a perceptron

The perceptron was intended to be a machine, rather than a program, and while its first implementation was in software for theIBM 704, it was subsequently implemented in custom-built hardware as theMark I Perceptron with the project name "Project PARA",[16] designed forimage recognition. The machine is currently inSmithsonian National Museum of American History.[17]

The Mark I Perceptron had three layers. One version was implemented as follows:

  • An array of 400photocells arranged in a 20x20 grid, named "sensory units" (S-units), or "input retina". Each S-unit can connect to up to 40 A-units.
  • A hidden layer of 512 perceptrons, named "association units" (A-units).
  • An output layer of eight perceptrons, named "response units" (R-units).

Rosenblatt called this three-layered perceptron network thealpha-perceptron, to distinguish it from other perceptron models he experimented with.[8]

The S-units are connected to the A-units randomly (according to a table of random numbers) via a plugboard (see photo), to "eliminate any particular intentional bias in the perceptron". The connection weights are fixed, not learned. Rosenblatt was adamant about the random connections, as he believed the retina was randomly connected to the visual cortex, and he wanted his perceptron machine to resemble human visual perception.[18]

The A-units are connected to the R-units, with adjustable weights encoded inpotentiometers, and weight updates during learning were performed by electric motors.[2]: 193 The hardware details are in an operators' manual.[16]

Components of the Mark I Perceptron. From the operator's manual.[16]

In a 1958 press conference organized by the US Navy, Rosenblatt made statements about the perceptron that caused a heated controversy among the fledglingAI community; based on Rosenblatt's statements,The New York Times reported the perceptron to be "the embryo of an electronic computer that [the Navy] expects will be able to walk, talk, see, write, reproduce itself and be conscious of its existence."[19]

The Photo Division ofCentral Intelligence Agency, from 1960 to 1964, studied the use of Mark I Perceptron machine for recognizing militarily interesting silhouetted targets (such as planes and ships) inaerial photos.[20][21]

Principles of Neurodynamics (1962)

[edit]

Rosenblatt described his experiments with many variants of the Perceptron machine in a bookPrinciples of Neurodynamics (1962). The book is a published version of the 1961 report.[22]

Among the variants are:

  • "cross-coupling" (connections between units within the same layer) with possibly closed loops,
  • "back-coupling" (connections from units in a later layer to units in a previous layer),
  • four-layer perceptrons where the last two layers have adjustable weights (and thus a proper multilayer perceptron),
  • incorporating time-delays to perceptron units, to allow for processing sequential data,
  • analyzing audio (instead of images).

The machine was shipped from Cornell to Smithsonian in 1967, under a government transfer administered by the Office of Naval Research.[9]

Perceptrons (1969)

[edit]
Main article:Perceptrons (book)

Although the perceptron initially seemed promising, it was quickly proved that perceptrons could not be trained to recognise many classes of patterns. This caused the field ofneural network research to stagnate for many years, before it was recognised that afeedforward neural network with two or more layers (also called amultilayer perceptron) had greater processing power than perceptrons with one layer (also called asingle-layer perceptron).

Single-layer perceptrons are only capable of learninglinearly separable patterns.[23] For a classification task with some step activation function, a single node will have a single line dividing the data points forming the patterns. More nodes can create more dividing lines, but those lines must somehow be combined to form more complex classifications. A second layer of perceptrons, or even linear nodes, are sufficient to solve many otherwise non-separable problems.

In 1969, a famous book entitledPerceptrons byMarvin Minsky andSeymour Papert showed that it was impossible for these classes of network to learn anXOR function. It is often incorrectly believed that they also conjectured that a similar result would hold for a multi-layer perceptron network. However, this is not true, as both Minsky and Papert already knew that multi-layer perceptrons were capable of producing an XOR function. (See the page onPerceptrons (book) for more information.) Nevertheless, the often-miscited Minsky and Papert text caused a significant decline in interest and funding of neural network research. It took ten more years until neural network research experienced a resurgence in the 1980s.[23][verification needed] This text was reprinted in 1987 as "Perceptrons - Expanded Edition" where some errors in the original text are shown and corrected.

Subsequent work

[edit]

Rosenblatt continued working on perceptrons despite diminishing funding. The last attempt was Tobermory, built between 1961 and 1967, built for speech recognition.[24] It occupied an entire room.[25] It had 4 layers with 12,000 weights implemented by toroidalmagnetic cores. By the time of its completion, simulation on digital computers had become faster than purpose-built perceptron machines.[26] He died in a boating accident in 1971.

A simulation program for neural networks was written forIBM 7090/7094, and was used to study various pattern recognition applications, such ascharacter recognition, particle tracks inbubble-chamber photographs; phoneme, isolated word, and continuousspeech recognition;speaker verification; andcenter-of-attention mechanisms for image processing.[27][28]

Isometric view of Tobermory, Phase I.[25]

Thekernel perceptron algorithm was already introduced in 1964 by Aizerman et al.[29] Margin bounds guarantees were given for the Perceptron algorithm in the general non-separable case first byFreund andSchapire (1998),[1] and more recently byMohri and Rostamizadeh (2013) who extend previous results and give new and more favorable L1 bounds.[30][31]

The perceptron is a simplified model of a biologicalneuron. While the complexity ofbiological neuron models is often required to fully understand neural behavior, research suggests a perceptron-like linear model can produce some behavior seen in real neurons.[32]

The solution spaces of decision boundaries for all binary functions and learning behaviors are studied in.[33]

Definition

[edit]
The appropriate weights are applied to the inputs, and the resulting weighted sum passed to a function that produces the output o.

In the modern sense, the perceptron is an algorithm for learning a binary classifier called athreshold function: a function that maps its inputx{\displaystyle \mathbf {x} } (a real-valuedvector) to an output valuef(x){\displaystyle f(\mathbf {x} )} (a singlebinary value):

f(x)=h(wx+b){\displaystyle f(\mathbf {x} )=h(\mathbf {w} \cdot \mathbf {x} +b)}

whereh{\displaystyle h} is theHeaviside step-function (where an input of>0{\textstyle >0} outputs 1; otherwise 0 is the output ),w{\displaystyle \mathbf {w} } is a vector of real-valued weights,wx{\displaystyle \mathbf {w} \cdot \mathbf {x} } is thedot producti=1mwixi{\textstyle \sum _{i=1}^{m}w_{i}x_{i}}, wherem is the number of inputs to the perceptron, andb is thebias. The bias shifts the decision boundary away from the origin and does not depend on any input value.

Equivalently, sincewx+b=(w,b)(x,1){\displaystyle \mathbf {w} \cdot \mathbf {x} +b=(\mathbf {w} ,b)\cdot (\mathbf {x} ,1)}, we can add the bias termb{\displaystyle b} as another weightwm+1{\displaystyle \mathbf {w} _{m+1}} and add a coordinate1{\displaystyle 1} to each inputx{\displaystyle \mathbf {x} }, and then write it as a linear classifier that passes the origin:f(x)=h(wx){\displaystyle f(\mathbf {x} )=h(\mathbf {w} \cdot \mathbf {x} )}

The binary value off(x){\displaystyle f(\mathbf {x} )} (0 or 1) is used to perform binary classification onx{\displaystyle \mathbf {x} } as either a positive or a negative instance. Spatially, the bias shifts the position (though not the orientation) of the planardecision boundary.

In the context of neural networks, a perceptron is anartificial neuron using theHeaviside step function as the activation function. The perceptron algorithm is also termed thesingle-layer perceptron, to distinguish it from amultilayer perceptron, which is a misnomer for a more complicated neural network. As a linear classifier, the single-layer perceptron is the simplestfeedforward neural network.

Power of representation

[edit]

Information theory

[edit]

From aninformation theory point of view, a single perceptron withK inputs has a capacity of2Kbits of information.[34] This result is due toThomas Cover.[35]

Specifically letT(N,K){\displaystyle T(N,K)} be the number of ways to linearly separateN points inK dimensions, thenT(N,K)={2NKN2k=0K1(N1k)K<N{\displaystyle T(N,K)=\left\{{\begin{array}{cc}2^{N}&K\geq N\\2\sum _{k=0}^{K-1}\left({\begin{array}{c}N-1\\k\end{array}}\right)&K<N\end{array}}\right.}WhenK is large,T(N,K)/2N{\displaystyle T(N,K)/2^{N}} is very close to one whenN2K{\displaystyle N\leq 2K}, but very close to zero whenN>2K{\displaystyle N>2K}. In words, one perceptron unit can almost certainly memorize a random assignment of binary labels on N points whenN2K{\displaystyle N\leq 2K}, but almost certainly not whenN>2K{\displaystyle N>2K}.

Boolean function

[edit]

When operating on only binary inputs, a perceptron is called alinearly separable Boolean function, or threshold Boolean function. The sequence of numbers of threshold Boolean functions on n inputs isOEISA000609. The value is only known exactly up ton=9{\displaystyle n=9} case, but the order of magnitude is known quite exactly: it has upper bound2n2nlog2n+O(n){\displaystyle 2^{n^{2}-n\log _{2}n+O(n)}} and lower bound2n2nlog2nO(n){\displaystyle 2^{n^{2}-n\log _{2}n-O(n)}}.[36]

Any Boolean linear threshold function can be implemented with only integer weights. Furthermore, the number of bits necessary and sufficient for representing a single integer weight parameter isΘ(nlnn){\displaystyle \Theta (n\ln n)}.[36]

Universal approximation theorem

[edit]
Main article:Universal approximation theorem

A single perceptron can learn to classify any half-space. It cannot solve any linearly nonseparable vectors, such as the Booleanexclusive-or problem (the famous "XOR problem").

A perceptron network withone hidden layer can learn to classify any compact subset arbitrarily closely. Similarly, it can also approximate anycompactly-supportedcontinuous function arbitrarily closely. This is essentially a special case of thetheorems by George Cybenko and Kurt Hornik.

Conjunctively local perceptron

[edit]
Main article:Perceptrons (book)

Perceptrons (Minsky and Papert, 1969) studied the kind of perceptron networks necessary to learn various Boolean functions.

Consider a perceptron network withn{\displaystyle n} input units, one hidden layer, and one output, similar to the Mark I Perceptron machine. It computes a Boolean function of typef:2n2{\displaystyle f:2^{n}\to 2}. They call a functionconjunctively local of orderk{\displaystyle k}, iff there exists a perceptron network such that each unit in the hidden layer connects to at mostk{\displaystyle k} input units.

Theorem. (Theorem 3.1.1): The parity function is conjunctively local of ordern{\displaystyle n}.

Theorem. (Section 5.5): The connectedness function is conjunctively local of orderΩ(n1/2){\displaystyle \Omega (n^{1/2})}.

Learning algorithm for a single-layer perceptron

[edit]
A diagram showing a perceptron updating its linear boundary as more training examples are added

Below is an example of a learning algorithm for a single-layer perceptron with a single output unit. For a single-layer perceptron with multiple output units, since the weights of one output unit are completely separate from all the others', the same algorithm can be run for each output unit.

Formultilayer perceptrons, where a hidden layer exists, more sophisticated algorithms such asbackpropagation must be used. If the activation function or the underlying process being modeled by the perceptron isnonlinear, alternative learning algorithms such as thedelta rule can be used as long as the activation function isdifferentiable. Nonetheless, the learning algorithm described in the steps below will often work, even for multilayer perceptrons with nonlinear activation functions.

When multiple perceptrons are combined in an artificial neural network, each output neuron operates independently of all the others; thus, learning each output can be considered in isolation.

Definitions

[edit]

We first define some variables:

We show the values of the features as follows:

To represent the weights:

To show the time-dependence ofw{\displaystyle \mathbf {w} }, we use:

Steps

[edit]
  1. Initialize the weights. Weights may be initialized to 0 or to a small random value. In the example below, we use 0. Set the time countert{\displaystyle t} to 0.
  2. For each examplej in our training setD, perform the following steps over the inputxj{\displaystyle \mathbf {x} _{j}} and desired outputdj{\displaystyle d_{j}}:
    1. Calculate the output with the current weight vectorw(t)=(w0(t),w1(t),,wn(t)){\displaystyle \mathbf {w} (t)=(w_{0}(t),w_{1}(t),\dots ,w_{n}(t))}:
      yj(t)=h[w0(t)xj,0+w1(t)xj,1+w2(t)xj,2++wn(t)xj,n]{\displaystyle {\begin{aligned}y_{j}(t)&=h[w_{0}(t)x_{j,0}+w_{1}(t)x_{j,1}+w_{2}(t)x_{j,2}+\dotsb +w_{n}(t)x_{j,n}]\end{aligned}}}
    2. Update the weights:
      wi(t+1)=wi(t)+r(djyj(t))xj,i{\displaystyle w_{i}(t+1)=w_{i}(t)\;{\boldsymbol {+}}\;r\cdot (d_{j}-y_{j}(t))x_{j,i}}, for all features0in{\displaystyle 0\leq i\leq n},r{\displaystyle r} is thelearning rate.
    3. Increment the time counter:t:=t+1{\displaystyle t:=t+1}

Foroffline learning, the second step may be repeated until the iteration error1sj=1s|djyj(t)|{\displaystyle {\frac {1}{s}}\sum _{j=1}^{s}|d_{j}-y_{j}(t)|} is less than a user-specified error thresholdγ{\displaystyle \gamma }, or a predetermined number of iterations have been completed, wheres is again the size of the sample set.

The algorithm updates the weights after every training sample in step 2b, though one may note that the weights remain unchanged wheneverdj=yj(t){\displaystyle d_{j}=y_{j}(t)}.

Convergence of one perceptron on a linearly separable dataset

[edit]
Illustration of the perceptron convergence. In the picture,γ=0.01,R=1,r=1{\displaystyle \gamma =0.01,R=1,r=1}. All data points havey=+1{\displaystyle y=+1}, since the negative samples are equivalent toy=+1{\displaystyle y=+1} after reflection through the origin. As the learning proceeds, the weight vector performs a somewhat random walk in the space of weights. Each step is at least 90 degrees away from its current direction, thus increasing its norm-square by at mostR{\displaystyle R}. Each step adds tow{\displaystyle w} by a point in the samples, and since all the samples havex10.01{\displaystyle x_{1}\geq 0.01}, the weight vector must move alongx1{\displaystyle x_{1}} by at least0.01{\displaystyle 0.01}. Since the norm grows liket{\displaystyle {\sqrt {t}}} but thex1{\displaystyle x_{1}}-component grows liket{\displaystyle t}, this would eventually force the weight vector to point almost entirely in thex1{\displaystyle x_{1}} direction, and thus achieve convergence.

A single perceptron is alinear classifier. It can only reach a stable state if all input vectors are classified correctly. In case the training setD isnotlinearly separable, i.e. if the positive examples cannot be separated from the negative examples by a hyperplane, then the algorithm would not converge since there is no solution. Hence, if linear separability of the training set is not known a priori, one of the training variants below should be used. Detailed analysis and extensions to the convergence theorem are in Chapter 11 ofPerceptrons (1969).

Linear separability is testable in timemin(O(nd/2),O(d2n),O(nd1lnn)){\displaystyle \min(O(n^{d/2}),O(d^{2n}),O(n^{d-1}\ln n))}, wheren{\displaystyle n} is the number of data points, andd{\displaystyle d} is the dimension of each point.[37]

If the training setis linearly separable, then the perceptron is guaranteed to converge after making finitely many mistakes.[38] The theorem is proved by Rosenblatt et al.

Perceptron convergence theoremGiven a datasetD{\textstyle D}, such thatmax(x,y)Dx2=R{\textstyle \max _{(x,y)\in D}\|x\|_{2}=R}, and it is linearly separable by some unit vectorw{\textstyle w^{*}}, with marginγ{\textstyle \gamma }:γ:=min(x,y)Dy(wx){\displaystyle \gamma :=\min _{(x,y)\in D}y(w^{*}\cdot x)}

Then the perceptron 0-1 learning algorithm converges after making at most(R/γ)2{\textstyle (R/\gamma )^{2}} mistakes, for any learning rate, and any method of sampling from the dataset.

The following simple proof is due to Novikoff (1962). The idea of the proof is that the weight vector is always adjusted by a bounded amount in a direction with which it has a negativedot product, and thus can be bounded above byO(t), wheret is the number of changes to the weight vector. However, it can also be bounded below byO(t) because if there exists an (unknown) satisfactory weight vector, then every change makes progress in this (unknown) direction by a positive amount that depends only on the input vector.

Proof

Suppose at stept{\textstyle t}, the perceptron with weightwt{\textstyle w_{t}} makes a mistake on data point(x,y){\textstyle (x,y)}, then it updates towt+1=wt+r(yfwt(x))x{\textstyle w_{t+1}=w_{t}+r(y-f_{w_{t}}(x))x}.

Ify=0{\textstyle y=0}, the argument is symmetric, so we omit it.

WLOG,y=1{\textstyle y=1}, thenfwt(x)=0{\textstyle f_{w_{t}}(x)=0},fw(x)=1{\textstyle f_{w^{*}}(x)=1}, andwt+1=wt+rx{\textstyle w_{t+1}=w_{t}+rx}.

By assumption, we have separation with margins:wxγ{\displaystyle w^{*}\cdot x\geq \gamma } Thus,
wwt+1wwt=w(rx)rγ{\displaystyle w^{*}\cdot w_{t+1}-w^{*}\cdot w_{t}=w^{*}\cdot (rx)\geq r\gamma }

Alsowt+122wt22=wt+rx22wt22=2r(wtx)+r2x22{\displaystyle \|w_{t+1}\|_{2}^{2}-\|w_{t}\|_{2}^{2}=\|w_{t}+rx\|_{2}^{2}-\|w_{t}\|_{2}^{2}=2r(w_{t}\cdot x)+r^{2}\|x\|_{2}^{2}} and since the perceptron made a mistake,wtx0{\textstyle w_{t}\cdot x\leq 0}, and so
wt+122wt22x22r2R2{\displaystyle \|w_{t+1}\|_{2}^{2}-\|w_{t}\|_{2}^{2}\leq \|x\|_{2}^{2}\leq r^{2}R^{2}}

Since we started withw0=0{\textstyle w_{0}=0}, after makingN{\textstyle N} mistakes,w2Nr2R2{\displaystyle \|w\|_{2}\leq {\sqrt {Nr^{2}R^{2}}}} but also
w2wwNrγ{\displaystyle \|w\|_{2}\geq w\cdot w^{*}\geq Nr\gamma }

Combining the two, we haveN(R/γ)2{\textstyle N\leq (R/\gamma )^{2}}

Two classes of points, and two of the infinitely many linear boundaries that separate them. Even though the boundaries are at nearly right angles to one another, the perceptron algorithm has no way of choosing between them.

While the perceptron algorithm is guaranteed to converge onsome solution in the case of a linearly separable training set, it may still pickany solution and problems may admit many solutions of varying quality.[39] Theperceptron of optimal stability, nowadays better known as the linearsupport-vector machine, was designed to solve this problem (Krauth andMezard, 1987).[40]

Perceptron cycling theorem

[edit]

When the dataset is not linearly separable, then there is no way for a single perceptron to converge. However, we still have[41]

Perceptron cycling theoremIf the datasetD{\displaystyle D} has only finitely many points, then there exists an upper bound numberM{\displaystyle M}, such that for any starting weight vectorw0{\displaystyle w_{0}} all weight vectorwt{\displaystyle w_{t}} has norm bounded bywtw0+M{\displaystyle \|w_{t}\|\leq \|w_{0}\|+M}

This is proved first byBradley Efron.[42]

Learning a Boolean function

[edit]

Consider a dataset where thex{\displaystyle x} are from{1,+1}n{\displaystyle \{-1,+1\}^{n}}, that is, the vertices of an n-dimensional hypercube centered at origin, andy=θ(xi){\displaystyle y=\theta (x_{i})}. That is, all data points with positivexi{\displaystyle x_{i}} havey=1{\displaystyle y=1}, and vice versa. By the perceptron convergence theorem, a perceptron would converge after making at mostn{\displaystyle n} mistakes.

If we were to write a logical program to perform the same task, each positive example shows that one of the coordinates is the right one, and each negative example shows that itscomplement is a positive example. By collecting all the known positive examples, we eventually eliminate all but one coordinate, at which point the dataset is learned.[43]

This bound is asymptotically tight in terms of the worst-case. In the worst-case, the first presented example is entirely new, and givesn{\displaystyle n} bits of information, but each subsequent example would differ minimally from previous examples, and gives 1 bit each. Aftern+1{\displaystyle n+1} examples, there are2n{\displaystyle 2n} bits of information, which is sufficient for the perceptron (with2n{\displaystyle 2n} bits of information).[34]

However, it is not tight in terms of expectation if the examples are presented uniformly at random, since the first would given{\displaystyle n} bits, the secondn/2{\displaystyle n/2} bits, and so on, takingO(lnn){\displaystyle O(\ln n)} examples in total.[43]

Variants

[edit]

The pocket algorithm with ratchet (Gallant, 1990) solves the stability problem of perceptron learning by keeping the best solution seen so far "in its pocket". The pocket algorithm then returns the solution in the pocket, rather than the last solution. It can be used also for non-separable data sets, where the aim is to find a perceptron with a small number of misclassifications. However, these solutions appear purely stochastically and hence the pocket algorithm neither approaches them gradually in the course of learning, nor are they guaranteed to show up within a given number of learning steps.

The Maxover algorithm (Wendemuth, 1995) is"robust" in the sense that it will converge regardless of (prior) knowledge of linear separability of the data set.[44] In the linearly separable case, it will solve the training problem – if desired, even with optimal stability (maximum margin between the classes). For non-separable data sets, it will return a solution with a computable small number of misclassifications.[45] In all cases, the algorithm gradually approaches the solution in the course of learning, without memorizing previous states and without stochastic jumps. Convergence is to global optimality for separable data sets and to local optimality for non-separable data sets.

The Voted Perceptron (Freund and Schapire, 1999), is a variant using multiple weighted perceptrons. The algorithm starts a new perceptron every time an example is wrongly classified, initializing the weights vector with the final weights of the last perceptron. Each perceptron will also be given another weight corresponding to how many examples do they correctly classify before wrongly classifying one, and at the end the output will be a weighted vote on all perceptrons.

In separable problems, perceptron training can also aim at finding the largest separating margin between the classes. The so-called perceptron of optimal stability can be determined by means of iterative training and optimization schemes, such as the Min-Over algorithm (Krauth and Mezard, 1987)[40] or the AdaTron (Anlauf and Biehl, 1989)).[46] AdaTron uses the fact that the corresponding quadratic optimization problem is convex. The perceptron of optimal stability, together with thekernel trick, are the conceptual foundations of thesupport-vector machine.

Theα{\displaystyle \alpha }-perceptron further used a pre-processing layer of fixed random weights, with thresholded output units. This enabled the perceptron to classifyanalogue patterns, by projecting them into abinary space. In fact, for a projection space of sufficiently high dimension, patterns can become linearly separable.

Another way to solve nonlinear problems without using multiple layers is to use higher order networks (sigma-pi unit). In this type of network, each element in the input vector is extended with each pairwise combination of multiplied inputs (second order). This can be extended to ann-order network.

It should be kept in mind, however, that the best classifier is not necessarily that which classifies all the training data perfectly. Indeed, if we had the prior constraint that the data come from equi-variant Gaussian distributions, the linear separation in the input space is optimal, and the nonlinear solution isoverfitted.

Other linear classification algorithms includeWinnow,support-vector machine, andlogistic regression.

Multiclass perceptron

[edit]

Like most other techniques for training linear classifiers, the perceptron generalizes naturally tomulticlass classification. Here, the inputx{\displaystyle x} and the outputy{\displaystyle y} are drawn from arbitrary sets. A feature representation functionf(x,y){\displaystyle f(x,y)} maps each possible input/output pair to a finite-dimensional real-valued feature vector. As before, the feature vector is multiplied by a weight vectorw{\displaystyle w}, but now the resulting score is used to choose among many possible outputs:

y^=argmaxyf(x,y)w.{\displaystyle {\hat {y}}=\operatorname {argmax} _{y}f(x,y)\cdot w.}

Learning again iterates over the examples, predicting an output for each, leaving the weights unchanged when the predicted output matches the target, and changing them when it does not. The update becomes:

wt+1=wt+f(x,y)f(x,y^).{\displaystyle w_{t+1}=w_{t}+f(x,y)-f(x,{\hat {y}}).}

This multiclass feedback formulation reduces to the original perceptron whenx{\displaystyle x} is a real-valued vector,y{\displaystyle y} is chosen from{0,1}{\displaystyle \{0,1\}}, andf(x,y)=yx{\displaystyle f(x,y)=yx}.

For certain problems, input/output representations and features can be chosen so thatargmaxyf(x,y)w{\displaystyle \mathrm {argmax} _{y}f(x,y)\cdot w} can be found efficiently even thoughy{\displaystyle y} is chosen from a very large or even infinite set.

Since 2002, perceptron training has become popular in the field ofnatural language processing for such tasks aspart-of-speech tagging andsyntactic parsing (Collins, 2002). It has also been applied to large-scale machine learning problems in adistributed computing setting.[47]

References

[edit]
  1. ^abFreund, Y.;Schapire, R. E. (1999)."Large margin classification using the perceptron algorithm"(PDF).Machine Learning.37 (3):277–296.doi:10.1023/A:1007662407062.S2CID 5885617.
  2. ^abBishop, Christopher M. (2006).Pattern Recognition and Machine Learning. Springer.ISBN 0-387-31073-8.
  3. ^Hecht-Nielsen, Robert (1991).Neurocomputing (Reprint. with corrections ed.). Reading (Mass.) Menlo Park (Calif.) New York [etc.]: Addison-Wesley. p. 6, Figure 1.3 caption.ISBN 978-0-201-09355-1.
  4. ^Block, H. D. (1962-01-01)."The Perceptron: A Model for Brain Functioning. I".Reviews of Modern Physics.34 (1):123–135.Bibcode:1962RvMP...34..123B.doi:10.1103/RevModPhys.34.123.ISSN 0034-6861.
  5. ^McCulloch, W; Pitts, W (1943)."A Logical Calculus of Ideas Immanent in Nervous Activity".Bulletin of Mathematical Biophysics.5 (4):115–133.doi:10.1007/BF02478259.
  6. ^abRosenblatt, Frank (1957)."The Perceptron—a perceiving and recognizing automaton"(PDF).Report 85-460-1. Cornell Aeronautical Laboratory.
  7. ^Rosenblatt, Frank (March 1960). "Perceptron Simulation Experiments".Proceedings of the IRE.48 (3):301–309.doi:10.1109/JRPROC.1960.287598.ISSN 0096-8390.
  8. ^abNilsson, Nils J. (2009). "4.2.1. Perceptrons".The Quest for Artificial Intelligence. Cambridge: Cambridge University Press.ISBN 978-0-521-11639-8.
  9. ^abO’Connor, Jack (2022-06-21)."Undercover Algorithm: A Secret Chapter in the Early History of Artificial Intelligence and Satellite Imagery".International Journal of Intelligence and CounterIntelligence:1–15.doi:10.1080/08850607.2022.2073542.ISSN 0885-0607.S2CID 249946000.
  10. ^Rosenblatt, F. (1958)."The perceptron: A probabilistic model for information storage and organization in the brain".Psychological Review.65 (6):386–408.doi:10.1037/h0042519.ISSN 1939-1471.PMID 13602029.
  11. ^Frank Rosenblatt, ‘Two Theorems of Statistical Separability in the Perceptron’, Symposium on the Mechanization of Thought, National Physical Laboratory, Teddington, UK, November 1958, vol. 1, H. M. Stationery Office, London, 1959.
  12. ^Rosenblatt, Frank, and CORNELL UNIV ITHACA NY.Cognitive Systems Research Program. Technical report, Cornell University, 72, 1971.
  13. ^Muerle, John Ludwig, and CORNELL AERONAUTICAL LAB INC BUFFALO NY.Project Para, Perceiving and Recognition Automata. Cornell Aeronautical Laboratory, Incorporated, 1963.
  14. ^Penn, Jonathan (2021-01-11).Inventing Intelligence: On the History of Complex Information Processing and Artificial Intelligence in the United States in the Mid-Twentieth Century (Thesis). [object Object].doi:10.17863/cam.63087.
  15. ^Guice, Jon (1998)."Controversy and the State: Lord ARPA and Intelligent Computing".Social Studies of Science.28 (1):103–138.doi:10.1177/030631298028001004.ISSN 0306-3127.JSTOR 285752.PMID 11619937.
  16. ^abcHay, John Cameron (1960).Mark I perceptron operators' manual (Project PARA) /(PDF). Buffalo: Cornell Aeronautical Laboratory. Archived fromthe original(PDF) on 2023-10-27.
  17. ^"Perceptron, Mark I".National Museum of American History. Retrieved2023-10-30.
  18. ^Anderson, James A.; Rosenfeld, Edward, eds. (2000).Talking Nets: An Oral History of Neural Networks. The MIT Press.doi:10.7551/mitpress/6626.003.0004.ISBN 978-0-262-26715-1.
  19. ^Olazaran, Mikel (1996). "A Sociological Study of the Official History of the Perceptrons Controversy".Social Studies of Science.26 (3):611–659.doi:10.1177/030631296026003005.JSTOR 285702.S2CID 16786738.
  20. ^"Perception Concepts to Photo-Interpretation".www.cia.gov. Retrieved2024-11-14.
  21. ^Irwin, Julia A. (2024-09-11)."Artificial Worlds and Perceptronic Objects: The CIA's Mid-century Automatic Target Recognition".Grey Room (97):6–35.doi:10.1162/grey_a_00415.ISSN 1526-3819.
  22. ^Principles of neurodynamics: Perceptrons and the theory of brain mechanisms, by Frank Rosenblatt, Report Number VG-1196-G-8, Cornell Aeronautical Laboratory, published on 15 March 1961. The work reported in this volume has been carried out under Contract Nonr-2381 (00) (Project PARA) at C.A.L. and Contract Nonr-401(40), at Cornell Univensity.
  23. ^abSejnowski, Terrence J. (2018).The Deep Learning Revolution. MIT Press. p. 47.ISBN 978-0-262-03803-4.
  24. ^Rosenblatt, Frank (1962). “A Description of the Tobermory Perceptron.” Cognitive Research Program. Report No. 4. Collected Technical Papers, Vol. 2. Edited by Frank Rosenblatt. Ithaca, NY: Cornell University.
  25. ^abNagy, George. 1963.System and circuit designs for the Tobermory perceptron. Technical report number 5, Cognitive Systems Research Program, Cornell University, Ithaca New York.
  26. ^Nagy, George. "Neural networks-then and now."IEEE Transactions on Neural Networks 2.2 (1991): 316-318.
  27. ^Barker, Trevor H. (1966-07-14).A Computer Program for Simulation of Perceptrons and Similar Neural Networks: Users' Manual (Technical Report). NONR 401 (40) and NSF GK-250. Ithaca, NY: Cornell University (published 1966).
  28. ^Nagy, George (March 1991)."Neural networks—then and now"(PDF).IEEE Transactions on Neural Networks.2 (2):316–318.doi:10.1109/72.80343.ISSN 1941-0093.PMID 18276386.
  29. ^Aizerman, M. A.; Braverman, E. M.; Rozonoer, L. I. (1964). "Theoretical foundations of the potential function method in pattern recognition learning".Automation and Remote Control.25:821–837.
  30. ^Mohri, Mehryar; Rostamizadeh, Afshin (2013). "Perceptron Mistake Bounds".arXiv:1305.0208 [cs.LG].
  31. ^[1] Foundations of Machine Learning, MIT Press (Chapter 8).
  32. ^Cash, Sydney; Yuste, Rafael (1999)."Linear Summation of Excitatory Inputs by CA1 Pyramidal Neurons".Neuron.22 (2):383–394.doi:10.1016/S0896-6273(00)81098-3.PMID 10069343.
  33. ^Liou, D.-R.; Liou, J.-W.; Liou, C.-Y. (2013).Learning Behaviors of Perceptron. iConcept Press.ISBN 978-1-477554-73-9.
  34. ^abMacKay, David (2003-09-25).Information Theory, Inference and Learning Algorithms.Cambridge University Press. p. 483.ISBN 9780521642989.
  35. ^Cover, Thomas M. (June 1965). "Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition".IEEE Transactions on Electronic Computers. EC-14 (3):326–334.doi:10.1109/PGEC.1965.264137.ISSN 0367-7508.
  36. ^abŠíma, Jiří; Orponen, Pekka (2003-12-01)."General-Purpose Computation with Neural Networks: A Survey of Complexity Theoretic Results".Neural Computation.15 (12):2727–2778.doi:10.1162/089976603322518731.ISSN 0899-7667.PMID 14629867.
  37. ^"Introduction to Machine Learning, Chapter 3: Perceptron".openlearninglibrary.mit.edu. Retrieved2023-10-27.
  38. ^Novikoff, Albert J. (1963). "On convergence proofs for perceptrons".Office of Naval Research.
  39. ^Bishop, Christopher M (2006-08-17). "Chapter 4. Linear Models for Classification".Pattern Recognition and Machine Learning. Springer Science+Business Media, LLC. p. 194.ISBN 978-0387-31073-2.
  40. ^abKrauth, W.; Mezard, M. (1987). "Learning algorithms with optimal stability in neural networks".Journal of Physics A: Mathematical and General.20 (11):L745 –L752.Bibcode:1987JPhA...20L.745K.doi:10.1088/0305-4470/20/11/013.
  41. ^Block, H. D.; Levin, S. A. (1970)."On the boundedness of an iterative procedure for solving a system of linear inequalities".Proceedings of the American Mathematical Society.26 (2):229–235.doi:10.1090/S0002-9939-1970-0265383-5.ISSN 0002-9939.
  42. ^Efron, Bradley. "The perceptron correction procedure in nonseparable situations."Rome Air Dev. Center Tech. Doc. Rept (1964).
  43. ^abSimon, Herbert A.; Laird, John E. (2019-08-13). "Limits on Speed of Concept Attainment".The Sciences of the Artificial, reissue of the third edition with a new introduction by John Laird (Reissue ed.). Cambridge, Massachusetts London, England: The MIT Press.ISBN 978-0-262-53753-7.
  44. ^Wendemuth, A. (1995). "Learning the Unlearnable".Journal of Physics A: Mathematical and General.28 (18):5423–5436.Bibcode:1995JPhA...28.5423W.doi:10.1088/0305-4470/28/18/030.
  45. ^Wendemuth, A. (1995). "Performance of robust training algorithms for neural networks".Journal of Physics A: Mathematical and General.28 (19):5485–5493.Bibcode:1995JPhA...28.5485W.doi:10.1088/0305-4470/28/19/006.
  46. ^Anlauf, J. K.; Biehl, M. (1989). "The AdaTron: an Adaptive Perceptron algorithm".Europhysics Letters.10 (7):687–692.Bibcode:1989EL.....10..687A.doi:10.1209/0295-5075/10/7/014.S2CID 250773895.
  47. ^McDonald, R.; Hall, K.; Mann, G. (2010)."Distributed Training Strategies for the Structured Perceptron"(PDF).Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the ACL. Association for Computational Linguistics. pp. 456–464.

Further reading

[edit]

External links

[edit]
Differentiable computing
General
Hardware
Software libraries
International
National
Other
Retrieved from "https://en.wikipedia.org/w/index.php?title=Perceptron&oldid=1314995336"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp