Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

CYK algorithm

From Wikipedia, the free encyclopedia
Parsing algorithm for context-free grammars
"CYK" redirects here. For other uses, seeCyk (disambiguation).
Cocke–Younger–Kasami algorithm (CYK)
ClassParsing withcontext-free grammars
Data structureString
Worst-caseperformanceO(n3|G|){\displaystyle {\mathcal {O}}\left(n^{3}\cdot \left|G\right|\right)}, where:

Incomputer science, theCocke–Younger–Kasami algorithm (alternatively calledCYK, orCKY) is aparsingalgorithm forcontext-free grammars published by Itiroo Sakai in 1961.[1][2] The algorithm is named after some of its rediscoverers:John Cocke, Daniel Younger,Tadao Kasami, andJacob T. Schwartz. It employsbottom-up parsing anddynamic programming.

The standard version of CYK operates only on context-free grammars given inChomsky normal form (CNF). However any context-free grammar may be algorithmically transformed into a CNF grammar expressing the same language (Sipser 1997).

The importance of the CYK algorithm stems from its high efficiency in certain situations. UsingbigO notation, theworst case running time of CYK isO(n3|G|){\displaystyle {\mathcal {O}}\left(n^{3}\cdot \left|G\right|\right)}, wheren{\displaystyle n} is the length of the parsed string and|G|{\displaystyle \left|G\right|} is the size of the CNF grammarG{\displaystyle G} (Hopcroft & Ullman 1979, p. 140). This makes it one of the most efficient[citation needed] parsing algorithms in terms of worst-caseasymptotic complexity, although other algorithms exist with better average running time in many practical scenarios.

Standard form

[edit]

Thedynamic programming algorithm requires the context-free grammar to be rendered intoChomsky normal form (CNF), because it tests for possibilities to split the current sequence into two smaller sequences. Any context-free grammar that does not generate the empty string can be represented in CNF using onlyproduction rules of the formsAα{\displaystyle A\rightarrow \alpha } andABC{\displaystyle A\rightarrow BC}; to allow for the empty string, one can explicitly allowSε{\displaystyle S\to \varepsilon }, whereS{\displaystyle S} is the start symbol.[3]

Algorithm

[edit]

As pseudocode

[edit]

The algorithm inpseudocode is as follows:

let the input be a stringI consisting ofn characters:a1 ...an.let the grammar containr nonterminal symbolsR1 ...Rr, with start symbolR1.letP[n,n,r] be an array of booleans. Initialize all elements ofP to false.letback[n,n,r] be an array of lists of backpointing triples. Initialize all elements ofback to the empty list.for eachs = 1 tonfor each unit productionRvassetP[1,s,v] = truefor eachl = 2 ton-- Length of spanfor eachs = 1 ton-l+1-- Start of spanfor eachp = 1 tol-1-- Partition of spanfor each productionRaRbRcifP[p,s,b] andP[l-p,s+p,c]thensetP[l,s,a] = true,                     append <p,b,c> toback[l,s,a]ifP[n,1,1] is truethenI is member of languagereturnback -- byretracing the steps through back, one can easily construct all possible parse trees of the string.elsereturn "not a member of language"

Probabilistic CYK (for finding the most probable parse)

[edit]

Allows to recover the most probable parse given the probabilities of all productions.

let the input be a stringI consisting ofn characters:a1 ...an.let the grammar containr nonterminal symbolsR1 ...Rr, with start symbolR1.letP[n,n,r] be an array of real numbers. Initialize all elements ofP to zero.letback[n,n,r] be an array of backpointing triples.for eachs = 1 tonfor each unit productionRvassetP[1,s,v] = Pr(Rvas)for eachl = 2 ton-- Length of spanfor eachs = 1 ton-l+1-- Start of spanfor eachp = 1 tol-1-- Partition of spanfor each productionRaRbRc        prob_splitting = Pr(RaRbRc) *P[p,s,b] *P[l-p,s+p,c]if prob_splitting >P[l,s,a]thensetP[l,s,a] = prob_splittingsetback[l,s,a] = <p,b,c>ifP[n,1,1] > 0then    find the parse tree by retracing throughbackreturn the parse treeelsereturn "not a member of language"

As prose

[edit]

In informal terms, this algorithm considers every possible substring of the input string and setsP[l,s,v]{\displaystyle P[l,s,v]} to be true if the substring of lengthl{\displaystyle l} starting froms{\displaystyle s} can be generated from the nonterminalRv{\displaystyle R_{v}}. Once it has considered substrings of length 1, it goes on to substrings of length 2, and so on. For substrings of length 2 and greater, it considers every possible partition of the substring into two parts, and checks to see if there is some productionABC{\displaystyle A\to B\;C} such thatB{\displaystyle B} matches the first part andC{\displaystyle C} matches the second part. If so, it recordsA{\displaystyle A} as matching the whole substring. Once this process is completed, the input string is generated by the grammar if the substring containing the entire input string is matched by the start symbol.

Example

[edit]
Sentence parsing using the CYK algorithm

This is an example grammar:

S NP VPVP VP PPVP V NPVP eatsPP P NPNP Det NNP sheV eatsP withN fishN forkDet a{\displaystyle {\begin{aligned}{\ce {S}}&\ {\ce {->NP\ VP}}\\{\ce {VP}}&\ {\ce {->VP\ PP}}\\{\ce {VP}}&\ {\ce {->V\ NP}}\\{\ce {VP}}&\ {\ce {->eats}}\\{\ce {PP}}&\ {\ce {->P\ NP}}\\{\ce {NP}}&\ {\ce {->Det\ N}}\\{\ce {NP}}&\ {\ce {->she}}\\{\ce {V}}&\ {\ce {->eats}}\\{\ce {P}}&\ {\ce {->with}}\\{\ce {N}}&\ {\ce {->fish}}\\{\ce {N}}&\ {\ce {->fork}}\\{\ce {Det}}&\ {\ce {->a}}\end{aligned}}}

Now the sentenceshe eats a fish with a fork is analyzed using the CYK algorithm. In the following table, inP[i,j,k]{\displaystyle P[i,j,k]},i is the number of the row (starting at the bottom at 1), andj is the number of the column (starting at the left at 1).

CYK table
S
VP
 
S
VPPP
SNPNP
NPV, VPDet.NPDetN
sheeatsafishwithafork

For readability, the CYK table forP is represented here as a 2-dimensional matrixM containing a set of non-terminal symbols, such thatRk is inM[i,j]{\displaystyle M[i,j]} if, and only if,P[i,j,k]{\displaystyle P[i,j,k]}.In the above example, since a start symbolS is inM[7,1]{\displaystyle M[7,1]}, the sentence can be generated by the grammar.

Extensions

[edit]

Generating a parse tree

[edit]

The above algorithm is arecognizer that will only determine if a sentence is in the language. It is simple to extend it into aparser that also constructs aparse tree, by storing parse tree nodes as elements of the array, instead of the boolean 1. The node is linked to the array elements that were used to produce it, so as to build the tree structure. Only one such node in each array element is needed if only one parse tree is to be produced. However, if all parse trees of an ambiguous sentence are to be kept, it is necessary to store in the array element a list of all the ways the corresponding node can be obtained in the parsing process. This is sometimes done with a second table B[n,n,r] of so-calledbackpointers.The end result is then a shared-forest of possible parse trees, where common trees parts are factored between the various parses. This shared forest can conveniently be read as anambiguous grammar generating only the sentence parsed, but with the same ambiguity as the original grammar, and the same parse trees up to a very simple renaming of non-terminals, as shown byLang (1994).

Parsing non-CNF context-free grammars

[edit]

As pointed out byLange & Leiß (2009), the drawback of all known transformations into Chomsky normal form is that they can lead to an undesirable bloat in grammar size. The size of a grammar is the sum of the sizes of its production rules, where the size of a rule is one plus the length of its right-hand side. Usingg{\displaystyle g} to denote the size of the original grammar, the size blow-up in the worst case may range fromg2{\displaystyle g^{2}} to22g{\displaystyle 2^{2g}}, depending on the transformation algorithm used. For the use in teaching, Lange and Leiß propose a slight generalization of the CYK algorithm, "without compromising efficiency of the algorithm, clarity of its presentation, or simplicity of proofs" (Lange & Leiß 2009).

Parsing weighted context-free grammars

[edit]

It is also possible to extend the CYK algorithm to parse strings usingweighted andstochastic context-free grammars. Weights (probabilities) are then stored in the table P instead of booleans, so P[i,j,A] will contain the minimum weight (maximum probability) that the substring from i to j can be derived from A. Further extensions of the algorithm allow all parses of a string to be enumerated from lowest to highest weight (highest to lowest probability).

Numerical stability

[edit]

When the probabilistic CYK algorithm is applied to a long string, the splitting probability can become very small due to multiplying many probabilities together. This can be dealt with by summing log-probability instead of multiplying probabilities.

Valiant's algorithm

[edit]

Theworst case running time of CYK isΘ(n3|G|){\displaystyle \Theta (n^{3}\cdot |G|)}, wheren is the length of the parsed string and |G| is the size of the CNF grammarG. This makes it one of the most efficient algorithms for recognizing general context-free languages in practice.Valiant (1975) gave an extension of the CYK algorithm. His algorithm computes the same parsing tableas the CYK algorithm; yet he showed thatalgorithms for efficient multiplication ofmatrices with 0-1-entries can be utilized for performing this computation.

Using theCoppersmith–Winograd algorithm for multiplying these matrices, this gives an asymptotic worst-case running time ofO(n2.38|G|){\displaystyle O(n^{2.38}\cdot |G|)}. However, the constant term hidden by theBig O Notation is so large that the Coppersmith–Winograd algorithm is only worthwhile for matrices that are too large to handle on present-day computers (Knuth 1997), and this approach requires subtraction and so is only suitable for recognition. The dependence on efficient matrix multiplication cannot be avoided altogether:Lee (2002) has proved that any parser for context-free grammars working in timeO(n3ε|G|){\displaystyle O(n^{3-\varepsilon }\cdot |G|)} can be effectively converted into an algorithm computing the product of(n×n){\displaystyle (n\times n)}-matrices with 0-1-entries in timeO(n3ε/3){\displaystyle O(n^{3-\varepsilon /3})}, and this was extended by Abboud et al.[4] to apply to a constant-size grammar.

See also

[edit]

References

[edit]
  1. ^Grune, Dick (2008).Parsing techniques : a practical guide (2nd ed.). New York: Springer. p. 579.ISBN 978-0-387-20248-8.
  2. ^Itiroo Sakai, “Syntax in universal translation”. In Proceedings 1961 International Conference on Machine Translation of Languages and Applied Language Analysis, Her Majesty’s Stationery Office, London, p. 593-608, 1962.
  3. ^Sipser, Michael (2006).Introduction to the theory of computation (2nd ed.). Boston: Thomson Course Technology. Definition 2.8.ISBN 0-534-95097-3.OCLC 58544333.
  4. ^Abboud, Amir; Backurs, Arturs; Williams, Virginia Vassilevska (2015-11-05). "If the Current Clique Algorithms are Optimal, so is Valiant's Parser".arXiv:1504.01431 [cs.CC].

Sources

[edit]

External links

[edit]
Top-down
Bottom-up
Mixed, other
Related topics
Retrieved from "https://en.wikipedia.org/w/index.php?title=CYK_algorithm&oldid=1300921646"
Category:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp