Intheoretical computer science, aprobabilistic Turing machine is anon-deterministic Turing machine that chooses between the availabletransitions at each point according to someprobability distribution. As a consequence, a probabilistic Turing machine can (unlike a deterministic Turing machine) havestochastic results; that is, on a given input and instruction state machine, it may have different run times, or it may not halt at all; furthermore, it may accept an input in one execution and reject the same input in another execution.
In the case of equal probabilities for the transitions, probabilistic Turing machines can be defined as deterministicTuring machines having an additional "write" instruction where the value of the write isuniformly distributed in the Turing machine's alphabet (generally, an equal likelihood of writing a "1" or a "0" on to the tape). Another common reformulation is simply adeterministic Turing machine with an added tape full of random bits called the "random tape".
Aquantum computer (orquantum Turing machine) is anothermodel of computation that is inherentlyprobabilistic.
A probabilistic Turing machine is a type ofnondeterministic Turing machine in which each nondeterministic step is a "coin-flip", that is, at each step there are two possible next moves and the Turing machine probabilistically selects which move to take.[1]
A probabilistic Turing machine can be formally defined as the 7-tuple, where
At each step, the Turing machine probabilistically applies either the transition function or the transition function.[2] This choice is made independently of all prior choices. In this way, the process of selecting a transition function at each step of the computation resembles a coin flip.
The probabilistic selection of the transition function at each step introduces error into the Turing machine; that is, strings which the Turing machine is meant to accept may on some occasions be rejected and strings which the Turing machine is meant to reject may on some occasions be accepted. To accommodate this, a language is said to be recognizedwith error probability by a probabilistic Turing machine if:
As a result of the error introduced by utilizing probabilistic coin tosses, the notion of acceptance of a string by a probabilistic Turing machine can be defined in different ways. One such notion that includes several important complexity classes is allowing for an error probability of 1/3. For instance, the complexity classBPP is defined as the class of languages recognized by a probabilistic Turing machine inpolynomial time with an error probability of 1/3. Another class defined using this notion of acceptance isBPL, which is the same asBPP but places the additional restriction that languages must be solvable inlogarithmicspace.
Complexity classes arising from other definitions of acceptance includeRP,co-RP, andZPP. If the machine is restricted to logarithmic space instead of polynomial time, the analogousRL,co-RL, andZPL complexity classes are obtained. By enforcing both restrictions,RLP,co-RLP,BPLP, andZPLP are yielded.
Probabilistic computation is also critical for the definition of most classes ofinteractive proof systems, in which the verifier machine depends on randomness to avoid being predicted and tricked by the all-powerful prover machine. For example, the classIP equalsPSPACE, but if randomness is removed from the verifier, we are left with onlyNP, which is not known but widely believed to be a considerably smaller class.
One of the central questions of complexity theory is whether randomness adds power; that is, is there a problem that can be solved in polynomial time by a probabilistic Turing machine but not a deterministic Turing machine? Or can deterministic Turing machines efficiently simulate all probabilistic Turing machines with at most a polynomial slowdown? It is known thatP ⊆BPP, since a deterministic Turing machine is just a special case of a probabilistic Turing machine. However, it is uncertain whether (but widely suspected that)BPP ⊆P, implying thatBPP =P. The same question for log space instead of polynomial time (doesL =BPLP?) is even more widely believed to be true. On the other hand, the power randomness gives to interactive proof systems, as well as the simple algorithms it creates for difficult problems such as polynomial-time primality testing and log-space graph connectedness testing, suggests that randomness may add power.