Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Quantum finite automaton

From Wikipedia, the free encyclopedia
Quantum analog of probabilistic automata

Inquantum computing,quantum finite automata (QFA) orquantum state machines are a quantum analog ofprobabilistic automata or aMarkov decision process. They provide a mathematical abstraction of real-worldquantum computers. Several types of automata may be defined, includingmeasure-once andmeasure-many automata. Quantum finite automata can also be understood as the quantization ofsubshifts of finite type, or as a quantization ofMarkov chains. QFAs are, in turn, special cases ofgeometric finite automata ortopological finite automata.

The automata work by receiving a finite-lengthstringσ=(σ0,σ1,,σk){\displaystyle \sigma =(\sigma _{0},\sigma _{1},\cdots ,\sigma _{k})} of lettersσi{\displaystyle \sigma _{i}} from a finitealphabetΣ{\displaystyle \Sigma }, and assigning to each such string aprobabilityPr(σ){\displaystyle \operatorname {Pr} (\sigma )} indicating the probability of the automaton being in anaccept state; that is, indicating whether the automaton accepted or rejected the string.

Thelanguages accepted by QFAs are not theregular languages ofdeterministic finite automata, nor are they thestochastic languages ofprobabilistic finite automata. Study of thesequantum languages remains an active area of research.

Informal description

[edit]

There is a simple, intuitive way of understanding quantum finite automata. One begins with agraph-theoretic interpretation ofdeterministic finite automata (DFA). A DFA can be represented as adirected graph, with states as nodes in the graph, and arrows representing state transitions. Each arrow is labelled with a possible input symbol, so that, given a specific state and an input symbol, the arrow points at the next state. One way of representing such a graph is by means of a set ofadjacency matrices, with one matrix for each input symbol. In this case, the list of possible DFA states is written as acolumn vector. For a given input symbol, the adjacency matrix indicates how any given state (row in the state vector) will transition to the next state; a state transition is given bymatrix multiplication.

One needs a distinct adjacency matrix for each possible input symbol, since each input symbol can result in a different transition. The entries in the adjacency matrix must be zero's and one's. For any given column in the matrix, only one entry can be non-zero: this is the entry that indicates the next (unique) state transition. Similarly, the state of the system is a column vector, in which only one entry is non-zero: this entry corresponds to the current state of the system. LetΣ{\displaystyle \Sigma } denote the set of input symbols. For a given input symbolαΣ{\displaystyle \alpha \in \Sigma }, writeUα{\displaystyle U_{\alpha }} as the adjacency matrix that describes the evolution of the DFA to its next state. The set{Uα|αΣ}{\displaystyle \{U_{\alpha }|\alpha \in \Sigma \}} then completely describes the state transition function of the DFA. LetQ represent the set of possible states of the DFA. If there areN states inQ, then each matrixUα{\displaystyle U_{\alpha }} isN byN-dimensional. The initial stateq0Q{\displaystyle q_{0}\in Q} corresponds to a column vector with a one in theq0'th row. A general stateq is then a column vector with a one in theq'th row. Byabuse of notation, letq0 andq also denote these two vectors. Then, after reading input symbolsαβγ{\displaystyle \alpha \beta \gamma \cdots } from the input tape, the state of the DFA will be given byq=UγUβUαq0.{\displaystyle q=\cdots U_{\gamma }U_{\beta }U_{\alpha }q_{0}.} The state transitions are given by ordinarymatrix multiplication (that is, multiplyq0 byUα{\displaystyle U_{\alpha }},etc.); the order of application is 'reversed' only because we follow the standard notation oflinear algebra.

The above description of a DFA, in terms oflinear operators and vectors, almost begs for generalization, by replacing the state-vectorq by some general vector, and the matrices{Uα}{\displaystyle \{U_{\alpha }\}} by some general operators. This is essentially what a QFA does: it replacesq by aunit vector, and the{Uα}{\displaystyle \{U_{\alpha }\}} byunitary matrices. Other, similar generalizations also become obvious: the vectorq can be somedistribution on amanifold; the set of transition matrices becomeautomorphisms of the manifold; this defines a topological finite automaton. Similarly, the matrices could be taken as automorphisms of ahomogeneous space; this defines a geometric finite automaton.

Before moving on to the formal description of a QFA, there are two noteworthy generalizations that should be mentioned and understood. The first is thenon-deterministic finite automaton (NFA). In this case, the vectorq is replaced by a vector that can have more than one entry that is non-zero. Such a vector then represents an element of thepower set ofQ; it’s just anindicator function onQ. Likewise, the state transition matrices{Uα}{\displaystyle \{U_{\alpha }\}} are defined in such a way that a given column can have several non-zero entries in it. Equivalently, the multiply-add operations performed during component-wise matrix multiplication should be replaced by Boolean and-or operations, that is, so that one is working with aring ofcharacteristic 2.

A well-known theorem states that, for each DFA, there is an equivalent NFA, andvice versa. This implies that the set oflanguages that can be recognized by DFA's and NFA's are the same; these are theregular languages. In the generalization to QFAs, the set of recognized languages will be different. Describing that set is one of the outstanding research problems in QFA theory.

Another generalization that should be immediately apparent is to use astochastic matrix for the transition matrices, and aprobability vector for the state; this gives aprobabilistic finite automaton. The entries in the state vector must be real numbers, positive, and sum to one, in order for the state vector to be interpreted as a probability. The transition matrices must preserve this property: this is why they must be stochastic. Each state vector should be imagined as specifying a point in asimplex; thus, this is a topological automaton, with the simplex being the manifold, and the stochastic matrices being linear automorphisms of the simplex onto itself. Since each transition is (essentially) independent of the previous (if we disregard the distinction between accepted and rejected languages), the PFA essentially becomes a kind ofMarkov chain.

By contrast, in a QFA, the manifold iscomplex projective spaceCPN{\displaystyle \mathbb {C} P^{N}}, and the transition matrices are unitary matrices. Each point inCPN{\displaystyle \mathbb {C} P^{N}} corresponds to a (pure)quantum-mechanical state; the unitary matrices can be thought of as governing the time evolution of the system (viz in theSchrödinger picture). The generalization from pure states tomixed states should be straightforward: A mixed state is simply ameasure-theoreticprobability distribution onCPN{\displaystyle \mathbb {C} P^{N}}.

A worthy point to contemplate is the distributions that result on the manifold during the input of a language. In order for an automaton to be 'efficient' in recognizing a language, that distribution should be 'as uniform as possible'. This need for uniformity is the underlying principle behindmaximum entropy methods: these simply guarantee crisp, compact operation of the automaton. Put in other words, themachine learning methods used to trainhidden Markov models generalize to QFAs as well: theViterbi algorithm and theforward–backward algorithm generalize readily to the QFA.

Although the study of QFA was popularized in the work of Kondacs and Watrous in 1997[1] and later by Moore and Crutchfeld,[2] they were described as early as 1971, byIon Baianu.[3][4]

Measure-once automata

[edit]

Measure-once automata were introduced byCris Moore andJames P. Crutchfield.[2] They may be defined formally as follows.

As with an ordinaryfinite automaton, the quantum automaton is considered to haveN{\displaystyle N} possible internal states, represented in this case by anN{\displaystyle N}-statequdit|ψ{\displaystyle |\psi \rangle }. More precisely, theN{\displaystyle N}-state qudit|ψP(CN){\displaystyle |\psi \rangle \in P(\mathbb {C} ^{N})} is an element of(N1){\displaystyle (N-1)}-dimensionalcomplex projective space, carrying aninner product{\displaystyle \Vert \cdot \Vert } that is theFubini–Study metric.

Thestate transitions,transition matrices orde Bruijn graphs are represented by a collection ofN×N{\displaystyle N\times N}unitary matricesUα{\displaystyle U_{\alpha }}, with one unitary matrix for each letterαΣ{\displaystyle \alpha \in \Sigma }. That is, given an input letterα{\displaystyle \alpha }, the unitary matrix describes the transition of the automaton from its current state|ψ{\displaystyle |\psi \rangle } to its next state|ψ{\displaystyle |\psi ^{\prime }\rangle }:

|ψ=Uα|ψ{\displaystyle |\psi ^{\prime }\rangle =U_{\alpha }|\psi \rangle }

Thus, the triple(P(CN),Σ,{Uα|αΣ}){\displaystyle (P(\mathbb {C} ^{N}),\Sigma ,\{U_{\alpha }\;\vert \;\alpha \in \Sigma \})} form aquantum semiautomaton.

Theaccept state of the automaton is given by anN×N{\displaystyle N\times N}projection matrixP{\displaystyle P}, so that, given aN{\displaystyle N}-dimensional quantum state|ψ{\displaystyle |\psi \rangle }, the probability of|ψ{\displaystyle |\psi \rangle } being in the accept state is

ψ|P|ψ=P|ψ2{\displaystyle \langle \psi |P|\psi \rangle =\Vert P|\psi \rangle \Vert ^{2}}

The probability of the state machine accepting a given finite input stringσ=(σ0,σ1,,σk){\displaystyle \sigma =(\sigma _{0},\sigma _{1},\cdots ,\sigma _{k})} is given by

Pr(σ)=PUσkUσ1Uσ0|ψ2{\displaystyle \operatorname {Pr} (\sigma )=\Vert PU_{\sigma _{k}}\cdots U_{\sigma _{1}}U_{\sigma _{0}}|\psi \rangle \Vert ^{2}}

Here, the vector|ψ{\displaystyle |\psi \rangle } is understood to represent the initial state of the automaton, that is, the state the automaton was in before it started accepting the string input. The empty string{\displaystyle \varnothing } is understood to be just the unit matrix, so that

Pr()=P|ψ2{\displaystyle \operatorname {Pr} (\varnothing )=\Vert P|\psi \rangle \Vert ^{2}}

is just the probability of the initial state being an accepted state.

Because the left-action ofUα{\displaystyle U_{\alpha }} on|ψ{\displaystyle |\psi \rangle } reverses the order of the letters in the stringσ{\displaystyle \sigma }, it is not uncommon for QFAs to be defined using a right action on theHermitian transpose states, simply in order to keep the order of the letters the same.

Alanguage over the alphabetΣ{\displaystyle \Sigma } is accepted with probabilityp{\displaystyle p} by a quantum finite automaton (and a given, fixed initial state|ψ{\displaystyle |\psi \rangle }), if, for all sentencesσ{\displaystyle \sigma } in the language, one haspPr(σ){\displaystyle p\leq \operatorname {Pr} (\sigma )}.

Example

[edit]

Consider the classicaldeterministic finite automaton given by thestate transition table

State Transition Table
  Input
State
10
S1S1S2
S2S2S1
 State Diagram
DFAexample.svg

The quantum state is a vector, inbra–ket notation

|ψ=a1|S1+a2|S2=[a1a2]{\displaystyle |\psi \rangle =a_{1}|S_{1}\rangle +a_{2}|S_{2}\rangle ={\begin{bmatrix}a_{1}\\a_{2}\end{bmatrix}}}

with thecomplex numbersa1,a2{\displaystyle a_{1},a_{2}} normalized so that

[a1a2][a1a2]=a1a1+a2a2=1{\displaystyle {\begin{bmatrix}a_{1}^{*}\;\;a_{2}^{*}\end{bmatrix}}{\begin{bmatrix}a_{1}\\a_{2}\end{bmatrix}}=a_{1}^{*}a_{1}+a_{2}^{*}a_{2}=1}

The unitary transition matrices are

U0=[0110]{\displaystyle U_{0}={\begin{bmatrix}0&1\\1&0\end{bmatrix}}}

and

U1=[1001]{\displaystyle U_{1}={\begin{bmatrix}1&0\\0&1\end{bmatrix}}}

TakingS1{\displaystyle S_{1}} to be the accept state, the projection matrix is

P=[1000]{\displaystyle P={\begin{bmatrix}1&0\\0&0\end{bmatrix}}}

As should be readily apparent, if the initial state is the pure state|S1{\displaystyle |S_{1}\rangle } or|S2{\displaystyle |S_{2}\rangle }, then the result of running the machine will be exactly identical to the classical deterministic finite state machine. In particular, there is a language accepted by this automaton with probability one, for these initial states, and it is identical to theregular language for the classical DFA, and is given by theregular expression:

(1(010)){\displaystyle (1^{*}(01^{*}0)^{*})^{*}\,\!}

The non-classical behaviour occurs if botha1{\displaystyle a_{1}} anda2{\displaystyle a_{2}} are non-zero. More subtle behaviour occurs when the matricesU0{\displaystyle U_{0}} andU1{\displaystyle U_{1}} are not so simple; see, for example, thede Rham curve as an example of a quantum finite state machine acting on the set of all possible finite binary strings.

Measure-many automata

[edit]

Measure-many automata were introduced by Kondacs and Watrous in 1997.[1] The general framework resembles that of the measure-once automaton, except that instead of there being one projection, at the end, there is a projection, orquantum measurement, performed after each letter is read. A formal definition follows.

TheHilbert spaceHQ{\displaystyle {\mathcal {H}}_{Q}} is decomposed into threeorthogonal subspaces

HQ=HacceptHrejectHnon-halting{\displaystyle {\mathcal {H}}_{Q}={\mathcal {H}}_{\text{accept}}\oplus {\mathcal {H}}_{\text{reject}}\oplus {\mathcal {H}}_{\text{non-halting}}}

In the literature, these orthogonal subspaces are usually formulated in terms of the setQ{\displaystyle Q} of orthogonal basis vectors for the Hilbert spaceHQ{\displaystyle {\mathcal {H}}_{Q}}. This set of basis vectors is divided up into subsetsQaccQ{\displaystyle Q_{\text{acc}}\subseteq Q} andQrejQ{\displaystyle Q_{\text{rej}}\subseteq Q}, such that

Haccept=span{|q:|qQacc}{\displaystyle {\mathcal {H}}_{\text{accept}}=\operatorname {span} \{|q\rangle :|q\rangle \in Q_{\text{acc}}\}}

is thelinear span of the basis vectors in the accept set. The reject space is defined analogously, and the remaining space is designated thenon-halting subspace. There are three projection matrices,Pacc{\displaystyle P_{\text{acc}}},Prej{\displaystyle P_{\text{rej}}} andPnon{\displaystyle P_{\text{non}}}, each projecting to the respective subspace:

Pacc:HQHaccept{\displaystyle P_{\text{acc}}:{\mathcal {H}}_{Q}\to {\mathcal {H}}_{\text{accept}}}

and so on. The parsing of the input string proceeds as follows. Consider the automaton to be in a state|ψ{\displaystyle |\psi \rangle }. After reading an input letterα{\displaystyle \alpha }, the automaton will be in the state

|ψ=Uα|ψ{\displaystyle |\psi ^{\prime }\rangle =U_{\alpha }|\psi \rangle }

At this point, a measurement whose three possible outcomes have eigenspacesHaccept{\displaystyle {\mathcal {H}}_{\text{accept}}},Hreject{\displaystyle {\mathcal {H}}_{\text{reject}}},Hnon-halting{\displaystyle {\mathcal {H}}_{\text{non-halting}}} is performed on the state|ψ{\displaystyle |\psi ^{\prime }\rangle }, at which time its wave-function collapses into one of the three subspacesHaccept{\displaystyle {\mathcal {H}}_{\text{accept}}} orHreject{\displaystyle {\mathcal {H}}_{\text{reject}}} orHnon-halting{\displaystyle {\mathcal {H}}_{\text{non-halting}}}. The probability of collapse to the "accept" subspace is given by

Pracc(σ)=Pacc|ψ2,{\displaystyle \operatorname {Pr} _{\text{acc}}(\sigma )=\Vert P_{\text{acc}}|\psi ^{\prime }\rangle \Vert ^{2},}

and analogously for the other two spaces.

If the wave function has collapsed to either the "accept" or "reject" subspaces, then further processing halts. Otherwise, processing continues, with the next letter read from the input, and applied to what must be an eigenstate ofPnon{\displaystyle P_{\text{non}}}. Processing continues until the whole string is read, or the machine halts. Often, additional symbolsκ{\displaystyle \kappa } and $ are adjoined to the alphabet, to act as the left and right end-markers for the string.

In the literature, the measure-many automaton is often denoted by the tuple(Q;Σ;δ;q0;Qacc;Qrej){\displaystyle (Q;\Sigma ;\delta ;q_{0};Q_{\text{acc}};Q_{\text{rej}})}. Here,Q{\displaystyle Q},Σ{\displaystyle \Sigma },Qacc{\displaystyle Q_{\text{acc}}} andQrej{\displaystyle Q_{\text{rej}}} are as defined above. The initial state is denoted by|q0{\displaystyle |q_{0}\rangle }. The unitary transformations are denoted by the mapδ{\displaystyle \delta },

δ:Q×Σ×QC{\displaystyle \delta :Q\times \Sigma \times Q\to \mathbb {C} }

so that

Uα|qi=qjQδ(qi,α,qj)|qj{\displaystyle U_{\alpha }|q_{i}\rangle =\sum _{q_{j}\in Q}\delta (q_{i},\alpha ,q_{j})|q_{j}\rangle }

Relation to quantum computing

[edit]

As of 2019, mostquantum computers are implementations of measure-once quantum finite automata, and the software systems for programming them expose the state-preparation of|ψ{\displaystyle |\psi \rangle }, measurementP{\displaystyle P} and a choice of unitary transformationsUα{\displaystyle U_{\alpha }}, such thecontrolled NOT gate, theHadamard transform and otherquantum logic gates, directly to the programmer.

The primary difference between real-world quantum computers and the theoretical framework presented above is that the initial state preparation cannot ever result in a point-likepure state, nor can the unitary operators be precisely applied. Thus, the initial state must be taken as amixed state

ρ=p(x)|ψxdx{\displaystyle \rho =\int p(x)|\psi _{x}\rangle dx}

for some probability distributionp(x){\displaystyle p(x)} characterizing the ability of the machinery to prepare an initial state close to the desired initial pure state|ψ{\displaystyle |\psi \rangle }. This state is not stable, but suffers from some amount ofquantum decoherence over time. Precise measurements are also not possible, and one instead usespositive operator-valued measures to describe the measurement process. Finally, each unitary transformation is not a single, sharply defined quantum logic gate, but rather a mixture

Uα,(ρ)=pα(x)Uα,xdx{\displaystyle U_{\alpha ,(\rho )}=\int p_{\alpha }(x)U_{\alpha ,x}dx}

for some probability distributionpα(x){\displaystyle p_{\alpha }(x)} describing how well the machinery can effect the desired transformationUα{\displaystyle U_{\alpha }}.

As a result of these effects, the actual time evolution of the state cannot be taken as an infinite-precision pure point, operated on by a sequence of arbitrarily sharp transformations, but rather as anergodic process, or more accurately, amixing process that only concatenates transformations onto a state, but also smears the state over time.

There is no quantum analog to thepush-down automaton orstack machine. This is due to theno-cloning theorem: there is no way to make a copy of the current state of the machine, push it onto a stack for later reference, and then return to it.

Geometric generalizations

[edit]

The above constructions indicate how the concept of a quantum finite automaton can be generalized to arbitrarytopological spaces. For example, one may take some (N-dimensional)Riemann symmetric space to take the place ofCPN{\displaystyle \mathbb {C} P^{N}}. In place of the unitary matrices, one uses theisometries of the Riemannian manifold, or, more generally, some set ofopen functions appropriate for the given topological space. The initial state may be taken to be a point in the space. The set of accept states can be taken to be some arbitrary subset of the topological space. One then says that aformal language is accepted by thistopological automaton if the point, after iteration by the homeomorphisms, intersects the accept set. But, of course, this is nothing more than the standard definition of anM-automaton. The behaviour of topological automata is studied in the field oftopological dynamics.

The quantum automaton differs from the topological automaton in that, instead of having a binary result (is the iterated point in, or not in, the final set?), one has a probability. The quantum probability is the (square of) the initial state projected onto some final stateP; that isPr=|P|ψ|2{\displaystyle \mathbf {Pr} =\vert \langle P\vert \psi \rangle \vert ^{2}}. But thisprobability amplitude is just a very simple function of the distance between the point|P{\displaystyle \vert P\rangle } and the point|ψ{\displaystyle \vert \psi \rangle } inCPN{\displaystyle \mathbb {C} P^{N}}, under the distancemetric given by theFubini–Study metric. To recap, the quantum probability of a language being accepted can be interpreted as a metric, with the probability of accept being unity, if the metric distance between the initial and final states is zero, and otherwise the probability of accept is less than one, if the metric distance is non-zero. Thus, it follows that the quantum finite automaton is just a special case of ageometric automaton or ametric automaton, whereCPN{\displaystyle \mathbb {C} P^{N}} is generalized to somemetric space, and the probability measure is replaced by a simple function of the metric on that space.

See also

[edit]

Notes

[edit]
  1. ^abKondacs, A.;Watrous, J. (1997), "On the power of quantum finite state automata",Proceedings of the 38th Annual Symposium on Foundations of Computer Science, pp. 66–75
  2. ^abC. Moore, J. Crutchfield, "Quantum automata and quantum grammars",Theoretical Computer Science,237 (2000) pp 275-306.
  3. ^I. Baianu, "Organismic Supercategories and Qualitative Dynamics of Systems" (1971), Bulletin of Mathematical Biophysics,33 pp.339-354.
  4. ^I. Baianu, "Categories, Functors and Quantum Automata Theory" (1971). The 4th Intl. Congress LMPS, August-Sept.1971
General
Theorems
Quantum
communication
Quantum cryptography
Quantum algorithms
Quantum
complexity theory
Quantum
processor benchmarks
Quantum
computing models
Quantum
error correction
Physical
implementations
Quantum optics
Ultracold atoms
Spin-based
Superconducting
Quantum
programming
Retrieved from "https://en.wikipedia.org/w/index.php?title=Quantum_finite_automaton&oldid=1285525510"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp