Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Computational complexity theory

From Wikipedia, the free encyclopedia
(Redirected fromTractable problem)
Inherent difficulty of computational problems

Intheoretical computer science and mathematics,computational complexity theory focuses on classifyingcomputational problems according to their resource usage, and explores the relationships between these classifications. A computational problem is a task solved by a computer. A computation problem is solvable by mechanical application of mathematical steps, such as analgorithm.

A problem is regarded as inherently difficult if its solution requires significant resources, whatever the algorithm used. The theory formalizes this intuition, by introducing mathematicalmodels of computation to study these problems and quantifying theircomputational complexity, i.e., the amount of resources needed to solve them, such as time and storage. Other measures of complexity are also used, such as the amount of communication (used incommunication complexity), the number ofgates in a circuit (used incircuit complexity) and the number of processors (used inparallel computing). One of the roles of computational complexity theory is to determine the practical limits on what computers can and cannot do. TheP versus NP problem, one of the sevenMillennium Prize Problems,[1] is part of the field of computational complexity.

Closely related fields intheoretical computer science areanalysis of algorithms andcomputability theory. A key distinction between analysis of algorithms and computational complexity theory is that the former is devoted to analyzing the amount of resources needed by a particular algorithm to solve a problem, whereas the latter asks a more general question about all possible algorithms that could be used to solve the same problem. More precisely, computational complexity theory tries to classify problems that can or cannot be solved with appropriately restricted resources. In turn, imposing restrictions on the available resources is what distinguishes computational complexity from computability theory: the latter theory asks what kinds of problems can, in principle, be solved algorithmically.

Computational problems

[edit]
A traveling salesman tour through 14 German cities

Problem instances

[edit]

Acomputational problem can be viewed as an infinite collection ofinstances together with a set (possibly empty) ofsolutions for every instance. The input string for a computational problem is referred to as a problem instance, and should not be confused with the problem itself. In computational complexity theory, a problem refers to the abstract question to be solved. In contrast, an instance of this problem is a rather concrete utterance, which can serve as the input for a decision problem. For example, consider the problem ofprimality testing. The instance is a number (e.g., 15) and the solution is "yes" if the number is prime and "no" otherwise (in this case, 15 is not prime and the answer is "no"). Stated another way, theinstance is a particular input to the problem, and thesolution is the output corresponding to the given input.

To further highlight the difference between a problem and an instance, consider the following instance of the decision version of thetravelling salesman problem: Is there a route of at most 2000 kilometres passing through all of Germany's 14 largest cities? The quantitative answer to this particular problem instance is of little use for solving other instances of the problem, such as asking for a round trip through all sites inMilan whose total length is at most 10 km. For this reason, complexity theory addresses computational problems and not particular problem instances.

Representing problem instances

[edit]

When considering computational problems, a problem instance is astring over analphabet. Usually, the alphabet is taken to be the binary alphabet (i.e., the set {0,1}), and thus the strings arebitstrings. As in a real-worldcomputer, mathematical objects other than bitstrings must be suitably encoded. For example,integers can be represented inbinary notation, andgraphs can be encoded directly via theiradjacency matrices, or by encoding theiradjacency lists in binary.

Even though some proofs of complexity-theoretic theorems regularly assume some concrete choice of input encoding, one tries to keep the discussion abstract enough to be independent of the choice of encoding. This can be achieved by ensuring that different representations can be transformed into each other efficiently.

Decision problems as formal languages

[edit]
Adecision problem has only two possible outputs,yes orno (or alternately 1 or 0) on any input.

Decision problems are one of the central objects of study in computational complexity theory. A decision problem is a type of computational problem where the answer is eitheryes orno (alternatively, 1 or 0). A decision problem can be viewed as aformal language, where the members of the language are instances whose output is yes, and the non-members are those instances whose output is no. The objective is to decide, with the aid of analgorithm, whether a given input string is a member of the formal language under consideration. If the algorithm deciding this problem returns the answeryes, the algorithm is said to accept the input string, otherwise it is said to reject the input.

An example of a decision problem is the following. The input is an arbitrarygraph. The problem consists in deciding whether the given graph isconnected or not. The formal language associated with this decision problem is then the set of all connected graphs — to obtain a precise definition of this language, one has to decide how graphs are encoded as binary strings.

Function problems

[edit]

Afunction problem is a computational problem where a single output (of atotal function) is expected for every input, but the output is more complex than that of adecision problem—that is, the output is not just yes or no. Notable examples include thetraveling salesman problem and theinteger factorization problem.

It is tempting to think that the notion of function problems is much richer than the notion of decision problems. However, this is not really the case, since function problems can be recast as decision problems. For example, the multiplication of two integers can be expressed as the set of triples(a,b,c){\displaystyle (a,b,c)} such that the relationa×b=c{\displaystyle a\times b=c} holds. Deciding whether a given triple is a member of this set corresponds to solving the problem of multiplying two numbers.

Measuring the size of an instance

[edit]

To measure the difficulty of solving a computational problem, one may wish to see how much time the best algorithm requires to solve the problem. However, the running time may, in general, depend on the instance. In particular, larger instances will require more time to solve. Thus the time required to solve a problem (or the space required, or any measure of complexity) is calculated as a function of the size of the instance. The input size is typically measured in bits. Complexity theory studies how algorithms scale as input size increases. For instance, in the problem of finding whether a graph is connected, how much more time does it take to solve a problem for a graph with2n{\displaystyle 2n} vertices compared to the time taken for a graph withn{\displaystyle n} vertices?

If the input size isn{\displaystyle n}, the time taken can be expressed as a function ofn{\displaystyle n}. Since the time taken on different inputs of the same size can be different, the worst-case time complexityT(n){\displaystyle T(n)} is defined to be the maximum time taken over all inputs of sizen{\displaystyle n}. IfT(n){\displaystyle T(n)} is a polynomial inn{\displaystyle n}, then the algorithm is said to be apolynomial time algorithm.Cobham's thesis argues that a problem can be solved with a feasible amount of resources if it admits a polynomial-time algorithm.

Machine models and complexity measures

[edit]

Turing machine

[edit]
Main article:Turing machine
An illustration of a Turing machine

A Turing machine is a mathematical model of a general computing machine. It is a theoretical device that manipulates symbols contained on a strip of tape. Turing machines are not intended as a practical computing technology, but rather as a general model of a computing machine—anything from an advanced supercomputer to a mathematician with a pencil and paper. It is believed that if a problem can be solved by an algorithm, there exists a Turing machine that solves the problem. Indeed, this is the statement of theChurch–Turing thesis. Furthermore, it is known that everything that can be computed on other models of computation known to us today, such as aRAM machine,Conway's Game of Life,cellular automata,lambda calculus or any programming language can be computed on a Turing machine. Since Turing machines are easy to analyze mathematically, and are believed to be as powerful as any other model of computation, the Turing machine is the most commonly used model in complexity theory.

Many types of Turing machines are used to define complexity classes, such asdeterministic Turing machines,probabilistic Turing machines,non-deterministic Turing machines,quantum Turing machines,symmetric Turing machines andalternating Turing machines. They are all equally powerful in principle, but when resources (such as time or space) are bounded, some of these may be more powerful than others.

A deterministic Turing machine is the most basic Turing machine, which uses a fixed set of rules to determine its future actions. A probabilistic Turing machine is a deterministic Turing machine with an extra supply of random bits. The ability to make probabilistic decisions often helps algorithms solve problems more efficiently. Algorithms that use random bits are calledrandomized algorithms. A non-deterministic Turing machine is a deterministic Turing machine with an added feature of non-determinism, which allows a Turing machine to have multiple possible future actions from a given state. One way to view non-determinism is that the Turing machine branches into many possible computational paths at each step, and if it solves the problem in any of these branches, it is said to have solved the problem. Clearly, this model is not meant to be a physically realizable model, it is just a theoretically interesting abstract machine that gives rise to particularly interesting complexity classes. For examples, seenon-deterministic algorithm.

Other machine models

[edit]

Many machine models different from the standardmulti-tape Turing machines have been proposed in the literature, for examplerandom-access machines. Perhaps surprisingly, each of these models can be converted to another without providing any extra computational power. The time and memory consumption of these alternate models may vary.[2] What all these models have in common is that the machines operatedeterministically.

However, some computational problems are easier to analyze in terms of more unusual resources. For example, a non-deterministic Turing machine is a computational model that is allowed to branch out to check many different possibilities at once. The non-deterministic Turing machine has very little to do with how we physically want to compute algorithms, but its branching exactly captures many of the mathematical models we want to analyze, so thatnon-deterministic time is a very important resource in analyzing computational problems.

Complexity measures

[edit]

For a precise definition of what it means to solve a problem using a given amount of time and space, a computational model such as thedeterministic Turing machine is used. The time required by a deterministic Turing machineM{\displaystyle M} on inputx{\displaystyle x} is the total number of state transitions, or steps, the machine makes before it halts and outputs the answer ("yes" or "no"). A Turing machineM{\displaystyle M} is said to operate within timef(n){\displaystyle f(n)} if the time required byM{\displaystyle M} on each input of lengthn{\displaystyle n} is at mostf(n){\displaystyle f(n)}. A decision problemA{\displaystyle A} can be solved in timef(n){\displaystyle f(n)} if there exists a Turing machine operating in timef(n){\displaystyle f(n)} that solves the problem. Since complexity theory is interested in classifying problems based on their difficulty, one defines sets of problems based on some criteria. For instance, the set of problems solvable within timef(n){\displaystyle f(n)} on a deterministic Turing machine is then denoted byDTIME(f(n){\displaystyle f(n)}).

Analogous definitions can be made for space requirements. Although time and space are the most well-known complexity resources, anycomplexity measure can be viewed as a computational resource. Complexity measures are very generally defined by theBlum complexity axioms. Other complexity measures used in complexity theory includecommunication complexity,circuit complexity, anddecision tree complexity.

The complexity of an algorithm is often expressed usingbig O notation.

Best, worst and average case complexity

[edit]
Visualization of thequicksortalgorithm, which hasaverage case performanceO(nlogn){\displaystyle {\mathcal {O}}(n\log n)}

Thebest, worst and average case complexity refer to three different ways of measuring the time complexity (or any other complexity measure) of different inputs of the same size. Since some inputs of sizen{\displaystyle n} may be faster to solve than others, we define the following complexities:

  1. Best-case complexity: This is the complexity of solving the problem for the best input of sizen{\displaystyle n}.
  2. Average-case complexity: This is the complexity of solving the problem on an average. This complexity is only defined with respect to aprobability distribution over the inputs. For instance, if all inputs of the same size are assumed to be equally likely to appear, the average case complexity can be defined with respect to the uniform distribution over all inputs of sizen{\displaystyle n}.
  3. Amortized analysis: Amortized analysis considers both the costly and less costly operations together over the whole series of operations of the algorithm.
  4. Worst-case complexity: This is the complexity of solving the problem for the worst input of sizen{\displaystyle n}.

The order from cheap to costly is: Best, average (ofdiscrete uniform distribution), amortized, worst.

For example, the deterministic sorting algorithmquicksort addresses the problem of sorting a list of integers. The worst-case is when the pivot is always the largest or smallest value in the list (so the list is never divided). In this case, the algorithm takes timeO(n2{\displaystyle n^{2}}). If we assume that all possible permutations of the input list are equally likely, the average time taken for sorting isO(nlogn){\displaystyle O(n\log n)}. The best case occurs when each pivoting divides the list in half, also needingO(nlogn){\displaystyle O(n\log n)} time.

Upper and lower bounds on the complexity of problems

[edit]

To classify the computation time (or similar resources, such as space consumption), it is helpful to demonstrate upper and lower bounds on the maximum amount of time required by the most efficient algorithm to solve a given problem. The complexity of an algorithm is usually taken to be its worst-case complexity unless specified otherwise. Analyzing a particular algorithm falls under the field ofanalysis of algorithms. To show an upper boundT(n){\displaystyle T(n)} on the time complexity of a problem, one needs to show only that there is a particular algorithm with running time at mostT(n){\displaystyle T(n)}. However, proving lower bounds is much more difficult, since lower bounds make a statement about all possible algorithms that solve a given problem. The phrase "all possible algorithms" includes not just the algorithms known today, but any algorithm that might be discovered in the future. To show a lower bound ofT(n){\displaystyle T(n)} for a problem requires showing that no algorithm can have time complexity lower thanT(n){\displaystyle T(n)}.

Upper and lower bounds are usually stated using thebig O notation, which hides constant factors and smaller terms. This makes the bounds independent of the specific details of the computational model used. For instance, ifT(n)=7n2+15n+40{\displaystyle T(n)=7n^{2}+15n+40}, in big O notation one would writeT(n)O(n2){\displaystyle T(n)\in O(n^{2})}.


Complexity classes

[edit]
Main article:Complexity class

Defining complexity classes

[edit]

Acomplexity class is a set of problems of related complexity. Simpler complexity classes are defined by the following factors:

  • The type of computational problem: The most commonly used problems are decision problems. However, complexity classes can be defined based onfunction problems,counting problems,optimization problems,promise problems, etc.
  • The model of computation: The most common model of computation is the deterministic Turing machine, but many complexity classes are based on non-deterministic Turing machines,Boolean circuits,quantum Turing machines,monotone circuits, etc.
  • The resource (or resources) that is being bounded and the bound: These two properties are usually stated together, such as "polynomial time", "logarithmic space", "constant depth", etc.

Some complexity classes have complicated definitions that do not fit into this framework. Thus, a typical complexity class has a definition like the following:

The set of decision problems solvable by a deterministic Turing machine within timef(n){\displaystyle f(n)}. (This complexity class is known as DTIME(f(n){\displaystyle f(n)}).)

But bounding the computation time above by some concrete functionf(n){\displaystyle f(n)} often yields complexity classes that depend on the chosen machine model. For instance, the language{xxx is any binary string}{\displaystyle \{xx\mid x{\text{ is any binary string}}\}} can be solved inlinear time on a multi-tape Turing machine, but necessarily requires quadratic time in the model of single-tape Turing machines. If we allow polynomial variations in running time,Cobham-Edmonds thesis states that "the time complexities in any two reasonable and general models of computation are polynomially related" (Goldreich 2008, Chapter 1.2). This forms the basis for the complexity classP, which is the set of decision problems solvable by a deterministic Turing machine within polynomial time. The corresponding set of function problems isFP.

Important complexity classes

[edit]
A representation of the relation among complexity classes; L would be another step "inside" NL

Many important complexity classes can be defined by bounding the time or space used by the algorithm. Some important complexity classes of decision problems defined in this manner are the following:

ResourceDeterminismComplexity classResource constraint
SpaceNon-DeterministicNSPACE(f(n){\displaystyle f(n)})O(f(n)){\displaystyle O(f(n))}
NLO(logn){\displaystyle O(\log n)}
NPSPACEO(poly(n)){\displaystyle O({\text{poly}}(n))}
NEXPSPACEO(2poly(n)){\displaystyle O(2^{{\text{poly}}(n)})}
DeterministicDSPACE(f(n){\displaystyle f(n)})O(f(n)){\displaystyle O(f(n))}
LO(logn){\displaystyle O(\log n)}
PSPACEO(poly(n)){\displaystyle O({\text{poly}}(n))}
EXPSPACEO(2poly(n)){\displaystyle O(2^{{\text{poly}}(n)})}
TimeNon-DeterministicNTIME(f(n){\displaystyle f(n)})O(f(n)){\displaystyle O(f(n))}
NPO(poly(n)){\displaystyle O({\text{poly}}(n))}
NEXPTIMEO(2poly(n)){\displaystyle O(2^{{\text{poly}}(n)})}
DeterministicDTIME(f(n){\displaystyle f(n)})O(f(n)){\displaystyle O(f(n))}
PO(poly(n)){\displaystyle O({\text{poly}}(n))}
EXPTIMEO(2poly(n)){\displaystyle O(2^{{\text{poly}}(n)})}

Logarithmic-space classes do not account for the space required to represent the problem.

It turns out that PSPACE = NPSPACE and EXPSPACE = NEXPSPACE bySavitch's theorem.

Other important complexity classes includeBPP,ZPP andRP, which are defined usingprobabilistic Turing machines;AC andNC, which are defined using Boolean circuits; andBQP andQMA, which are defined using quantum Turing machines.#P is an important complexity class of counting problems (not decision problems). Classes likeIP andAM are defined usingInteractive proof systems.ALL is the class of all decision problems.

Hierarchy theorems

[edit]
Main articles:time hierarchy theorem andspace hierarchy theorem

For the complexity classes defined in this way, it is desirable to prove that relaxing the requirements on (say) computation time indeed defines a bigger set of problems. In particular, although DTIME(n{\displaystyle n}) is contained in DTIME(n2{\displaystyle n^{2}}), it would be interesting to know if the inclusion is strict. For time and space requirements, the answer to such questions is given by the time and space hierarchy theorems respectively. They are called hierarchy theorems because they induce a proper hierarchy on the classes defined by constraining the respective resources. Thus there are pairs of complexity classes such that one is properly included in the other. Having deduced such proper set inclusions, we can proceed to make quantitative statements about how much more additional time or space is needed in order to increase the number of problems that can be solved.

More precisely, thetime hierarchy theorem states thatDTIME(o(f(n)))DTIME(f(n)log(f(n))){\displaystyle {\mathsf {DTIME}}{\big (}o(f(n)){\big )}\subsetneq {\mathsf {DTIME}}{\big (}f(n)\cdot \log(f(n)){\big )}}.

Thespace hierarchy theorem states thatDSPACE(o(f(n)))DSPACE(f(n)){\displaystyle {\mathsf {DSPACE}}{\big (}o(f(n)){\big )}\subsetneq {\mathsf {DSPACE}}{\big (}f(n){\big )}}.

The time and space hierarchy theorems form the basis for most separation results of complexity classes. For instance, the time hierarchy theorem tells us that P is strictly contained in EXPTIME, and the space hierarchy theorem tells us that L is strictly contained in PSPACE.

Reduction

[edit]
Main article:Reduction (complexity)

Many complexity classes are defined using the concept of a reduction. A reduction is a transformation of one problem into another problem. It captures the informal notion of a problem being at most as difficult as another problem. For instance, if a problemX{\displaystyle X} can be solved using an algorithm forY{\displaystyle Y},X{\displaystyle X} is no more difficult thanY{\displaystyle Y}, and we say thatX{\displaystyle X}reduces toY{\displaystyle Y}. There are many different types of reductions, based on the method of reduction, such as Cook reductions, Karp reductions and Levin reductions, and the bound on the complexity of reductions, such aspolynomial-time reductions orlog-space reductions.

The most commonly used reduction is a polynomial-time reduction. This means that the reduction process takes polynomial time. For example, the problem of squaring an integer can be reduced to the problem of multiplying two integers. This means an algorithm for multiplying two integers can be used to square an integer. Indeed, this can be done by giving the same input to both inputs of the multiplication algorithm. Thus we see that squaring is not more difficult than multiplication, since squaring can be reduced to multiplication.

This motivates the concept of a problem being hard for a complexity class. A problemX{\displaystyle X} ishard for a class of problemsC{\displaystyle C} if every problem inC{\displaystyle C} can be reduced toX{\displaystyle X}. Thus no problem inC{\displaystyle C} is harder thanX{\displaystyle X}, since an algorithm forX{\displaystyle X} allows us to solve any problem inC{\displaystyle C}. The notion of hard problems depends on the type of reduction being used. For complexity classes larger than P, polynomial-time reductions are commonly used. In particular, the set of problems that are hard for NP is the set ofNP-hard problems.

If a problemX{\displaystyle X} is inC{\displaystyle C} and hard forC{\displaystyle C}, thenX{\displaystyle X} is said to becomplete forC{\displaystyle C}. This means thatX{\displaystyle X} is the hardest problem inC{\displaystyle C}. (Since many problems could be equally hard, one might say thatX{\displaystyle X} is one of the hardest problems inC{\displaystyle C}.) Thus the class ofNP-complete problems contains the most difficult problems in NP, in the sense that they are the ones most likely not to be in P. Because the problem P = NP is not solved, being able to reduce a known NP-complete problem,Π2{\displaystyle \Pi _{2}}, to another problem,Π1{\displaystyle \Pi _{1}}, would indicate that there is no known polynomial-time solution forΠ1{\displaystyle \Pi _{1}}. This is because a polynomial-time solution toΠ1{\displaystyle \Pi _{1}} would yield a polynomial-time solution toΠ2{\displaystyle \Pi _{2}}. Similarly, because all NP problems can be reduced to the set, finding anNP-complete problem that can be solved in polynomial time would mean that P = NP.[3]

Important open problems

[edit]
Diagram of complexity classes provided that P ≠ NP. The existence of problems in NP outside both P and NP-complete in this case was established by Ladner.[4]

P versus NP problem

[edit]
Main article:P versus NP problem

The complexity class P is often seen as a mathematical abstraction modeling those computational tasks that admit an efficient algorithm. This hypothesis is called theCobham–Edmonds thesis. The complexity classNP, on the other hand, contains many problems that people would like to solve efficiently, but for which no efficient algorithm is known, such as theBoolean satisfiability problem, theHamiltonian path problem and thevertex cover problem. Since deterministic Turing machines are special non-deterministic Turing machines, it is easily observed that each problem in P is also member of the class NP.

The question of whether P equals NP is one of the most important open questions in theoretical computer science because of the wide implications of a solution.[3] If the answer is yes, many important problems can be shown to have more efficient solutions. These include various types ofinteger programming problems inoperations research, many problems inlogistics,protein structure prediction inbiology,[5] and the ability to find formal proofs ofpure mathematics theorems.[6] The P versus NP problem is one of theMillennium Prize Problems proposed by theClay Mathematics Institute. There is a US$1,000,000 prize for resolving the problem.[7]

Problems in NP not known to be in P or NP-complete

[edit]

It was shown by Ladner that ifPNP{\displaystyle P\neq NP} then there exist problems inNP{\displaystyle NP} that are neither inP{\displaystyle P} norNP{\displaystyle NP}-complete.[4] Such problems are calledNP-intermediate problems. Thegraph isomorphism problem, thediscrete logarithm problem and theinteger factorization problem are examples of problems believed to be NP-intermediate. They are some of the very few NP problems not known to be inP{\displaystyle P} or to beNP{\displaystyle NP}-complete.

Thegraph isomorphism problem is the computational problem of determining whether two finitegraphs areisomorphic. An important unsolved problem in complexity theory is whether the graph isomorphism problem is inP{\displaystyle P},NP{\displaystyle NP}-complete, or NP-intermediate. The answer is not known, but it is believed that the problem is at least not NP-complete.[8] If graph isomorphism is NP-complete, thepolynomial time hierarchy collapses to its second level.[9] Since it is widely believed that the polynomial hierarchy does not collapse to any finite level, it is believed that graph isomorphism is not NP-complete. The best algorithm for this problem, due toLászló Babai andEugene Luks has run timeO(2nlogn){\displaystyle O(2^{\sqrt {n\log n}})} for graphs withn{\displaystyle n} vertices, although some recent work by Babai offers some potentially new perspectives on this.[10]

Theinteger factorization problem is the computational problem of determining theprime factorization of a given integer. Phrased as a decision problem, it is the problem of deciding whether the input has a prime factor less thank{\displaystyle k}. No efficient integer factorization algorithm is known, and this fact forms the basis of several modern cryptographic systems, such as theRSA algorithm. The integer factorization problem is inNP{\displaystyle NP} and inco-NP{\displaystyle co{\text{-}}NP} (and even in UP and co-UP[11]). If the problem isNP{\displaystyle NP}-complete, the polynomial time hierarchy will collapse to its first level (i.e.,NP{\displaystyle NP} will equalco-NP{\displaystyle co{\text{-}}NP}). The best known algorithm for integer factorization is thegeneral number field sieve, which takes timeO(e(6493)(logn)3(loglogn)23){\displaystyle O(e^{\left({\sqrt[{3}]{\frac {64}{9}}}\right){\sqrt[{3}]{(\log n)}}{\sqrt[{3}]{(\log \log n)^{2}}}})}[12] to factor an odd integern{\displaystyle n}. However, the best knownquantum algorithm for this problem,Shor's algorithm, does run in polynomial time. Unfortunately, this fact doesn't say much about where the problem lies with respect to non-quantum complexity classes.

Separations between other complexity classes

[edit]

Many known complexity classes are suspected to be unequal, but this has not been proved. For instancePNPPPPSPACE{\displaystyle P\subseteq NP\subseteq PP\subseteq PSPACE}, but it is possible thatP=PSPACE{\displaystyle P=PSPACE}. IfP{\displaystyle P} is not equal toNP{\displaystyle NP}, thenP{\displaystyle P} is not equal toPSPACE{\displaystyle PSPACE} either. Since there are many known complexity classes betweenP{\displaystyle P} andPSPACE{\displaystyle PSPACE}, such asRP{\displaystyle RP},BPP{\displaystyle BPP},PP{\displaystyle PP},BQP{\displaystyle BQP},MA{\displaystyle MA},PH{\displaystyle PH}, etc., it is possible that all these complexity classes collapse to one class. Proving that any of these classes are unequal would be a major breakthrough in complexity theory.

Along the same lines,co-NP{\displaystyle co{\text{-}}NP} is the class containing thecomplement problems (i.e. problems with theyes/no answers reversed) ofNP{\displaystyle NP} problems. It is believed[13] thatNP{\displaystyle NP} is not equal toco-NP{\displaystyle co{\text{-}}NP}; however, it has not yet been proven. It is clear that if these two complexity classes are not equal thenP{\displaystyle P} is not equal toNP{\displaystyle NP}, sinceP=co-P{\displaystyle P=co{\text{-}}P}. Thus ifP=NP{\displaystyle P=NP} we would haveco-P=co-NP{\displaystyle co{\text{-}}P=co{\text{-}}NP} whenceNP=P=co-P=co-NP{\displaystyle NP=P=co{\text{-}}P=co{\text{-}}NP}.

Similarly, it is not known ifL{\displaystyle L} (the set of all problems that can be solved in logarithmic space) is strictly contained inP{\displaystyle P} or equal toP{\displaystyle P}. Again, there are many complexity classes between the two, such asNL{\displaystyle NL} andNC{\displaystyle NC}, and it is not known if they are distinct or equal classes.

It is suspected thatP{\displaystyle P} andBPP{\displaystyle BPP} are equal. However, it is currently open ifBPP=NEXP{\displaystyle BPP=NEXP}.

Intractability

[edit]
See also:Combinatorial explosion

A problem that can theoretically be solved, but requires impractical and finite resources (e.g., time) to do so, is known as anintractable problem.[14] Conversely, a problem that can be solved in practice is called atractable problem, literally "a problem that can be handled". The terminfeasible (literally "cannot be done") is sometimes used interchangeably withintractable,[15] though this risks confusion with afeasible solution inmathematical optimization.[16]

Tractable problems are frequently identified with problems that have polynomial-time solutions (P{\displaystyle P},PTIME{\displaystyle PTIME}); this is known as theCobham–Edmonds thesis. Problems that are known to be intractable in this sense include those that areEXPTIME-hard. IfNP{\displaystyle NP} is not the same asP{\displaystyle P}, thenNP-hard problems are also intractable in this sense.

However, this identification is inexact: a polynomial-time solution with large degree or large leading coefficient grows quickly, and may be impractical for practical size problems; conversely, an exponential-time solution that grows slowly may be practical on realistic input, or a solution that takes a long time in the worst case may take a short time in most cases or the average case, and thus still be practical. Saying that a problem is not inP{\displaystyle P} does not imply that all large cases of the problem are hard or even that most of them are. For example, the decision problem inPresburger arithmetic has been shown not to be inP{\displaystyle P}, yet algorithms have been written that solve the problem in reasonable times in most cases. Similarly, algorithms can solve the NP-completeknapsack problem over a wide range of sizes in less than quadratic time andSAT solvers routinely handle large instances of the NP-completeBoolean satisfiability problem.

To see why exponential-time algorithms are generally unusable in practice, consider a program that makes2n{\displaystyle 2^{n}} operations before halting. For smalln{\displaystyle n}, say 100, and assuming for the sake of example that the computer does1012{\displaystyle 10^{12}} operations each second, the program would run for about4×1010{\displaystyle 4\times 10^{10}} years, which is the same order of magnitude as theage of the universe. Even with a much faster computer, the program would only be useful for very small instances and in that sense the intractability of a problem is somewhat independent of technological progress. However, an exponential-time algorithm that takes1.0001n{\displaystyle 1.0001^{n}} operations is practical untiln{\displaystyle n} gets relatively large.

Similarly, a polynomial time algorithm is not always practical. If its running time is, say,n15{\displaystyle n^{15}}, it is unreasonable to consider it efficient and it is still useless except on small instances. Indeed, in practice evenn3{\displaystyle n^{3}} orn2{\displaystyle n^{2}} algorithms are often impractical on realistic sizes of problems.

Continuous complexity theory

[edit]

Continuous complexity theory can refer to complexity theory of problems that involve continuous functions that are approximated by discretizations, as studied innumerical analysis. One approach to complexity theory of numerical analysis[17] isinformation based complexity.

Continuous complexity theory can also refer to complexity theory of the use ofanalog computation, which uses continuousdynamical systems anddifferential equations.[18]Control theory can be considered a form of computation and differential equations are used in the modelling of continuous-time and hybrid discrete-continuous-time systems.[19]

History

[edit]

An early example of algorithm complexity analysis is the running time analysis of theEuclidean algorithm done byGabriel Lamé in 1844.

Before the actual research explicitly devoted to the complexity of algorithmic problems started off, numerous foundations were laid out by various researchers. Most influential among these was the definition of Turing machines byAlan Turing in 1936, which turned out to be a very robust and flexible simplification of a computer.

The beginning of systematic studies in computational complexity is attributed to the seminal 1965 paper "On the Computational Complexity of Algorithms" byJuris Hartmanis andRichard E. Stearns, which laid out the definitions oftime complexity andspace complexity, and proved the hierarchy theorems.[20] In addition, in 1965Edmonds suggested to consider a "good" algorithm to be one with running time bounded by a polynomial of the input size.[21]

Earlier papers studying problems solvable by Turing machines with specific bounded resources include[20]John Myhill's definition oflinear bounded automata (Myhill 1960),Raymond Smullyan's study of rudimentary sets (1961), as well asHisao Yamada's paper[22] on real-time computations (1962). Somewhat earlier,Boris Trakhtenbrot (1956), a pioneer in the field from the USSR, studied another specific complexity measure.[23] As he remembers:

However, [my] initial interest [in automata theory] was increasingly set aside in favor of computational complexity, an exciting fusion of combinatorial methods, inherited fromswitching theory, with the conceptual arsenal of the theory of algorithms. These ideas had occurred to me earlier in 1955 when I coined the term "signalizing function", which is nowadays commonly known as "complexity measure".[24]

In 1967,Manuel Blum formulated a set ofaxioms (now known asBlum axioms) specifying desirable properties of complexity measures on the set of computable functions and proved an important result, the so-calledspeed-up theorem. The field began to flourish in 1971 whenStephen Cook andLeonid Levinproved the existence of practically relevant problems that areNP-complete. In 1972,Richard Karp took this idea a leap forward with his landmark paper, "Reducibility Among Combinatorial Problems", in which he showed that 21 diversecombinatorial andgraph theoretical problems, each infamous for its computational intractability, are NP-complete.[25]

See also

[edit]

Works on complexity

[edit]

References

[edit]

Citations

[edit]
  1. ^"P vs NP Problem | Clay Mathematics Institute".www.claymath.org. Archived fromthe original on July 6, 2018. RetrievedJuly 6, 2018.
  2. ^SeeArora & Barak 2009, Chapter 1: The computational model and why it doesn't matter
  3. ^abSeeSipser 2006, Chapter 7: Time complexity
  4. ^abLadner, Richard E. (1975), "On the structure of polynomial time reducibility",Journal of the ACM,22 (1):151–171,doi:10.1145/321864.321877,S2CID 14352974.
  5. ^Berger, Bonnie A.;Leighton, T (1998), "Protein folding in the hydrophobic-hydrophilic (HP) model is NP-complete",Journal of Computational Biology,5 (1):27–40,CiteSeerX 10.1.1.139.5547,doi:10.1089/cmb.1998.5.27,PMID 9541869.
  6. ^Cook, Stephen (April 2000),The P versus NP Problem(PDF),Clay Mathematics Institute, archived fromthe original(PDF) on December 12, 2010, retrievedOctober 18, 2006.
  7. ^Jaffe, Arthur M. (2006),"The Millennium Grand Challenge in Mathematics"(PDF),Notices of the AMS,53 (6),archived(PDF) from the original on June 12, 2006, retrievedOctober 18, 2006.
  8. ^Arvind, Vikraman; Kurur, Piyush P. (2006), "Graph isomorphism is in SPP",Information and Computation,204 (5):835–852,doi:10.1016/j.ic.2006.02.002.
  9. ^Schöning, Uwe (1988), "Graph Isomorphism is in the Low Hierarchy",Journal of Computer and System Sciences,37 (3):312–323,doi:10.1016/0022-0000(88)90010-4
  10. ^Babai, László (2016). "Graph Isomorphism in Quasipolynomial Time".arXiv:1512.03547 [cs.DS].
  11. ^Fortnow, Lance (September 13, 2002)."Computational Complexity Blog: Factoring".weblog.fortnow.com.
  12. ^Wolfram MathWorld:Number Field Sieve
  13. ^Boaz Barak's course on Computational ComplexityLecture 2
  14. ^Hopcroft, J.E., Motwani, R. and Ullman, J.D. (2007)Introduction to Automata Theory, Languages, and Computation, Addison Wesley, Boston/San Francisco/New York (page 368)
  15. ^Meurant, Gerard (2014).Algorithms and Complexity. Elsevier. p. p. 4.ISBN 978-0-08093391-7.
  16. ^Zobel, Justin (2015).Writing for Computer Science. Springer. p. 132.ISBN 978-1-44716639-9.
  17. ^Smale, Steve (1997). "Complexity Theory and Numerical Analysis".Acta Numerica.6. Cambridge Univ Press:523–551.Bibcode:1997AcNum...6..523S.CiteSeerX 10.1.1.33.4678.doi:10.1017/s0962492900002774.S2CID 5949193.
  18. ^Babai, László; Campagnolo, Manuel (2009). "A Survey on Continuous Time Computations".arXiv:0907.3117 [cs.CC].
  19. ^Tomlin, Claire J.; Mitchell, Ian; Bayen, Alexandre M.;Oishi, Meeko (July 2003). "Computational Techniques for the Verification of Hybrid Systems".Proceedings of the IEEE.91 (7):986–1001.CiteSeerX 10.1.1.70.4296.doi:10.1109/jproc.2003.814621.
  20. ^abFortnow & Homer (2003)
  21. ^Richard M. Karp, "Combinatorics, Complexity, and Randomness", 1985 Turing Award Lecture
  22. ^Yamada, H. (1962). "Real-Time Computation and Recursive Functions Not Real-Time Computable".IEEE Transactions on Electronic Computers. EC-11 (6):753–760.doi:10.1109/TEC.1962.5219459.
  23. ^Trakhtenbrot, B.A.: Signalizing functions and tabular operators. Uchionnye ZapiskiPenzenskogo Pedinstituta (Transactions of the Penza Pedagogoical Institute) 4, 75–87 (1956) (in Russian)
  24. ^Boris Trakhtenbrot, "From Logic to Theoretical Computer Science – An Update". In:Pillars of Computer Science, LNCS 4800, Springer 2008.
  25. ^Richard M. Karp (1972),"Reducibility Among Combinatorial Problems"(PDF), in R. E. Miller; J. W. Thatcher (eds.),Complexity of Computer Computations, New York: Plenum, pp. 85–103, archived fromthe original(PDF) on June 29, 2011, retrievedSeptember 28, 2009

Textbooks

[edit]

Surveys

[edit]

External links

[edit]
Look uptractable,feasible,intractability, orinfeasible in Wiktionary, the free dictionary.
Wikimedia Commons has media related toComputational complexity theory.
Considered feasible
Suspected infeasible
Considered infeasible
Class hierarchies
Families of classes
Note: This template roughly follows the 2012ACM Computing Classification System.
Hardware
Computer systems organization
Networks
Software organization
Software notations andtools
Software development
Theory of computation
Algorithms
Mathematics ofcomputing
Information systems
Security
Human–computer interaction
Concurrency
Artificial intelligence
Machine learning
Graphics
Applied computing
Authority control databases: NationalEdit this at Wikidata
Retrieved from "https://en.wikipedia.org/w/index.php?title=Computational_complexity_theory&oldid=1283580007#tractable_problem"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp