Typically, alogic consists of a formal or informal languagetogether with a deductive system and/or a model-theoretic semantics.The language has components that correspond to a part of a naturallanguage like English or Greek. The deductive system is to capture,codify, or simply recordarguments that arevalidfor the given language, and the semantics is to capture, codify, orrecord the meanings, or truth-conditions for at least part of thelanguage.
The following sections provide the basics of a typical logic,sometimes called “classical elementary logic” or“classical first-order logic”. Section 2 develops a formallanguage, with a rigorous syntax and grammar. The formal language is arecursively defined collection of strings on a fixed alphabet. Assuch, it has no meaning, or perhaps better, the meaning of itsformulas is given by the deductive system and the semantics. Some ofthe symbols have counterparts in ordinary language. We define anargument to be a non-empty collection of sentences in theformal language, one of which is designated to be the conclusion. Theother sentences (if any) in an argument are its premises. Section 3sets up a deductive system for the language, in the spirit of naturaldeduction. An argument isderivable if there is a deductionfrom some or all of its premises to its conclusion. Section 4 providesa model-theoretic semantics. An argument isvalid if there isno interpretation (in the semantics) in which its premises are alltrue and its conclusion false. This reflects the longstanding viewthat a valid argument is truth-preserving.
In Section 5, we turn to relationships between the deductive systemand the semantics, and in particular, the relationship betweenderivability and validity. We show that an argument is derivable onlyif it is valid. This pleasant feature, calledsoundness,entails that no deduction takes one from true premises to a falseconclusion. Thus, deductions preserve truth. Then we establish aconverse, calledcompleteness, that an argument is valid onlyif it is derivable. This shows that the deductive system is richenough to provide a deduction for every valid argument. So there areenough deductions: all and only valid arguments are derivable. Webriefly indicate other features of the logic, some of which arecorollaries to soundness and completeness.
The final section, Section 6, is devoted to the a brief examination ofthe philosophical position that classical logic is “the oneright logic”.
Today, logic is a branch of mathematics and a branch of philosophy. Inmost large universities, both departments offer courses in logic, andthere is usually a lot of overlap between them. Formal languages,deductive systems, and model-theoretic semantics are mathematicalobjects and, as such, the logician is interested in their mathematicalproperties and relations. Soundness, completeness, and most of theother results reported below are typical examples. Philosophically,logic is at least closely related to the study ofcorrectreasoning. Reasoning is an epistemic, mental activity. So logicis at least closely allied with epistemology. Logic is also a centralbranch of computer science, due, in part, to interesting computationalrelations in logical systems, and, in part, to the close connectionbetween formal deductive argumentation and reasoning (see the entriesonrecursive functions,computability and complexity, andphilosophy of computer science).
This raises questions concerning the philosophical relevance of thevarious mathematical aspects of logic. How do deducibility andvalidity, as properties of formal languages – sets of strings ona fixed alphabet – relate to correct reasoning? What do themathematical results reported below have to do with the originalphilosophical issues concerning valid reasoning? This is an instanceof the philosophical problem of explaining how mathematics applies tonon-mathematical reality.
Typically, ordinary deductive reasoning takes place in a naturallanguage, or perhaps a natural language augmented with somemathematical symbols. So our question begins with the relationshipbetween a natural language and a formal language. Without attemptingto be comprehensive, it may help to sketch several options on thismatter.
One view is that the formal languages accurately exhibit actualfeatures of certain fragments of a natural language. Some philosophersclaim that declarative sentences of natural language have underlyinglogical forms and that these forms are displayed by formulasof a formal language. Other writers hold that (successful) declarativesentences expresspropositions; and formulas of formallanguages somehow display the forms of these propositions. On viewslike this, the components of a logic provide the underlying deepstructure of correct reasoning. A chunk of reasoning in naturallanguage is correct if the forms underlying the sentences constitute avalid or deducible argument. See for example, Montague [1974],Davidson [1984], Lycan [1984] (and the entry onlogical form).
Another view, held at least in part by Gottlob Frege and WilhelmLeibniz, is that because natural languages are fraught with vaguenessand ambiguity, they should bereplaced by formal languages. Asimilar view, held by W. V. O. Quine (e.g., [1960], [1986]), is that anatural language should beregimented, cleaned up for seriousscientific and metaphysical work. One desideratum of the enterprise isthat the logical structures in the regimented language should betransparent. It should be easy to “read off” the logicalproperties of each sentence. A regimented language is similar to aformal language regarding, for example, the explicitly presented rigorof its syntax and its truth conditions.
On a view like this, deducibility and validity representidealizations of correct reasoning in natural language. Achunk of reasoning is correct to the extent that it corresponds to, orcan be regimented by, a valid or deducible argument in a formallanguage.
When mathematicians and many philosophers engage in deductivereasoning, they occasionally invoke formulas in a formal language tohelp disambiguate, or otherwise clarify what they mean. In otherwords, sometimes formulas in a formal language areused inordinary reasoning. This suggests that one might think of a formallanguage as anaddendum to a natural language. Then ourpresent question concerns the relationship between this addendum andthe original language. What do deducibility and validity, as sharplydefined on the addendum, tell us about correct deductive reasoning ingeneral?
Another view is that a formal language is amathematicalmodel of a natural language in roughly the same sense as, say, acollection of point masses is a model of a system of physical objects,and the Bohr construction is a model of an atom. In other words, aformal language displays certain features of natural languages, oridealizations thereof, while ignoring or simplifying other features.The purpose of mathematical models is to shed light on what they aremodels of, without claiming that the model is accurate in all respectsor that the model should replace what it is a model of. On a view likethis, deducibility and validity represent mathematical models of(perhaps different aspects of) correct reasoning in natural languages.Correct chunks of deductive reasoning correspond, more or less, tovalid or deducible arguments; incorrect chunks of reasoning roughlycorrespond to invalid or non-deducible arguments. See, for example,Corcoran [1973], Shapiro [1998], and Cook [2002].
There is no need to adjudicate this matter here. Perhaps the truthlies in a combination of the above options, or maybe some other optionis the correct, or most illuminating one. We raise the matter only tolend some philosophical perspective to the formal treatment thatfollows.
Here we develop the basics of a formal language, or to be precise, aclass of formal languages. Again, a formal language is a recursivelydefined set of strings on a fixed alphabet. Some aspects of the formallanguages correspond to, or have counterparts in, natural languageslike English. Technically, this “counterpart relation” isnot part of the formal development, but we will mention it from timeto time, to motivate some of the features and results.
We begin with analogues ofsingular terms, linguistic itemswhose function is to denote a person or object. We call theseterms. We assume a stock ofindividual constants.These are lower-case letters, near the beginning of the Romanalphabet, with or without numerical subscripts:
\[a, a_1, b_{23}, c, d_{22}, \text{etc}.\]We envisage a potential infinity of individual constants. In thepresent system each constant is a single character, and so individualconstants do not have an internal syntax. Thus we have an infinitealphabet. This could be avoided by taking a constant like \(d_{22}\),for example, to consist of three characters, a lowercase“\(d\)” followed by a pair of subscript“2”s.
We also assume a stock ofindividual variables. These arelower-case letters, near the end of the alphabet, with or withoutnumerical subscripts:
\[w, x, y_{12}, z, z_4, \text{etc}. \]In ordinary mathematical reasoning, there are two functions terms needto fulfill. We need to be able to denote specific, but unspecified (orarbitrary) objects, and sometimes we need to express generality. Inour system, we use some constants in the role of unspecified referenceand variables to express generality. Both uses are recapitulated inthe formal treatment below. Some logicians employ different symbolsfor unspecified objects (sometimes called “individualparameters”) and variables used to express generality.
Constants and variables are the only terms in our formal language, soall of our terms are simple, corresponding to proper names and someuses of pronouns. We call a term closed if it is not a variable. Ingeneral, we use \(v\) to represent variables, and \(t\) to represent aclosed term, an individual constant. Some authors also introducefunction letters, which allow complex terms corresponding to:“\(7+4\)” and “the wife of Bill Clinton”, orcomplex terms containing variables, like “the father of\(x\)” and “\(x/y\)”. Logic books aimed atmathematicians are likely to contain function letters, probably due tothe centrality of functions in mathematical discourse. Books aimed ata more general audience (or at philosophy students), may leave outfunction letters, since it simplifies the syntax and theory. We followthe latter route here. This is an instance of a general tradeoffbetween presenting a system with greater expressive resources, at thecost of making its formal treatment more complex.
For each natural number \(n\), we introduce a stock of \(n\)-placepredicate letters. These are upper-case letters at thebeginning or middle of the alphabet. A superscript indicates thenumber of places, and there may or may not be a subscript. Forexample,
\[A^3, B^{3}_2, P^3, \text{etc}.\]are three-place predicate letters. We often omit the superscript, whenno confusion will result. We also add a special two-place predicatesymbol “\(=\)” for identity.
Zero-place predicate letters are sometimes called “sentenceletters”. They correspond to free-standing sentences whoseinternal structure does not matter. One-place predicate letters,called “monadic predicate letters”, correspond tolinguistic items denoting properties, like “being a man”,“being red”, or “being a prime number”.Two-place predicate letters, called “binary predicateletters”, correspond to linguistic items denoting binaryrelations, like “is a parent of” or “is greaterthan”. Three-place predicate letters correspond to three-placerelations, like “lies on a straight line between”. And soon.
Thenon-logical terminology of the language consists of itsindividual constants and predicate letters. The symbol“\(=\)”, for identity, is not a non-logical symbol. Intaking identity to be logical, we provide explicit treatment for it inthe deductive system and in the model-theoretic semantics. Mostauthors do the same, but there is some controversy over the issue(Quine [1986, Chapter 5]). If \(K\) is a set of constants andpredicate letters, then we give the fundamentals of a language\(\LKe\) built on this set of non-logical terminology. It may becalled thefirst-order language with identity on \(K\). Asimilar language that lacks the symbol for identity (or which takesidentity to be non-logical) may be called \(\mathcal{L}1K\), thefirst-order language without identity on \(K\).
If \(V\) is an \(n\)-place predicate letter in \(K\), and \(t_1,\ldots,t_n\) are terms of \(K\), then \(Vt_1 \ldots t_n\) is anatomic formula of \(\LKe\). Notice that the terms \(t_1,\ldots,t_n\) need not be distinct. Examples of atomic formulasinclude:
\[P^4 xaab, C^1 x, C^1 a, D^0, A^3 abc.\]The last one is an analogue of a statement that a certain relation\((A)\) holds between three objects \((a, b, c)\). If \(t_1\) and\(t_2\) are terms, then \(t_1 =t_2\) is also an atomic formula of\(\LKe\). It corresponds to an assertion that \(t_1\) is identical to\(t_2\).
If an atomic formula has no variables, then it is called anatomicsentence. If it does have variables, it is calledopen.In the above list of examples, the first and second are open; the restare sentences.
We now introduce the final items of the lexicon:
\[\neg, \amp, \vee, \rightarrow, \forall, \exists, (, )\]We give a recursive definition of aformula of \(\LKe\):
A formula corresponding to \(\neg \theta\) thus says that it is notthe case that \(\theta\). The symbol “\(\neg\)” is called“negation”, and is a unary connective.
The ampersand “\(\amp\)” corresponds to the English“and” (when “and” is used to connectsentences). So \((\theta \amp \psi)\) can be read “\(\theta\)and \(\psi\)”. The formula \((\theta \amp \psi)\) is called the“conjunction” of \(\theta\) and \(\psi\).
The symbol “\(\vee\)” corresponds to “either… or … or both”, so \((\theta \vee \psi)\) can beread “\(\theta\) or \(\psi\)”. The formula \((\theta \vee\psi)\) is called the “disjunction” of \(\theta\) and\(\psi\).
The arrow “\(\rightarrow\)” roughly corresponds to“if … then … ”, so \((\theta \rightarrow\psi)\) can be read “if \(\theta\) then \(\psi\)” or“\(\theta\) only if \(\psi\)”.
The symbols “\(\amp\)”, “\(\vee\)”, and“\(\rightarrow\)” are called “binaryconnectives”, since they serve to “connect” twoformulas into one. Some authors introduce \((\theta \leftrightarrow\psi)\) as an abbreviation of \(((\theta \rightarrow \psi) \amp(\psi\rightarrow \theta))\). The symbol “\(\leftrightarrow\)”is an analogue of the locution “if and only if”.
The symbol “\(\forall\)” is called auniversalquantifier, and is an analogue of “for all”; so\(\forall v\theta\) can be read “for all \(v,\theta\)”.
The symbol “\(\exists\)” is called anexistentialquantifier, and is an analogue of “there exists” or“there is”; so \(\exists v \theta\) can be read“there is a \(v\) such that \(\theta\)”.
Clause (8) allows us to do inductions on the complexity of formulas.If a certain property holds of the atomic formulas and is closed underthe operations presented in clauses (2)–(7), then the propertyholds of all formulas. Here is a simple example:
Theorem 1. Every formula of \(\LKe\) has the samenumber of left and right parentheses. Moreover, each left parenthesiscorresponds to a unique right parenthesis, which occurs to the rightof the left parenthesis. Similarly, each right parenthesis correspondsto a unique left parenthesis, which occurs to the left of the givenright parenthesis. If a parenthesis occurs between a matched pair ofparentheses, then its mate also occurs within that matched pair. Inother words, parentheses that occur within a matched pair arethemselves matched.
Proof: By clause (8), every formula is built up fromthe atomic formulas using clauses (2)–(7). The atomic formulashave no parentheses. Parentheses are introduced only in clauses(3)–(5), and each time they are introduced as a matched set. Soat any stage in the construction of a formula, the parentheses arepaired off.
We next define the notion of an occurrence of a variable beingfree orbound in a formula. A variable thatimmediately follows a quantifier (as in “\(\forall x\)”and “\(\exists y\)”) is neither free nor bound. We do noteven think of those as occurrences of the variable. All variables thatoccur in an atomic formula are free. If a variable occurs free (orbound) in \(\theta\) or in \(\psi\), then that same occurrence is free(or bound) in \(\neg \theta, (\theta \amp \psi), (\theta \vee \psi)\),and \((\theta \rightarrow \psi)\). That is, the (unary and binary)connectives do not change the status of variables that occur in them.All occurrences of the variable \(v\) in \(\theta\) are bound in\(\forall v \theta\) and \(\exists v \theta\). Anyfreeoccurrences of \(v\) in \(\theta\) are bound by the initialquantifier. All other variables that occur in \(\theta\) are free orbound in \(\forall v \theta\) and \(\exists v \theta\), as they are in\(\theta\).
For example, in the formula \((\forall\)x(Axy \(\vee Bx) \ampBx)\), the occurrences of “\(x\)” inAxy and inthe first \(Bx\) are bound by the quantifier. The occurrence of“\(y\)” and last occurrence of “\(x\)” arefree. In \(\forall x(Ax \rightarrow \exists\)xBx), the“\(x\)” in \(Ax\) is bound by the initial universalquantifier, while the other occurrence of \(x\) is bound by theexistential quantifier. The above syntax allows this“double-binding”. Although it does not create anyambiguities (see below), we will avoid such formulas, as a matter oftaste and clarity.
The syntax also allows so-called vacuous binding, as in\(\forall\)x\(Bc\). These, too, will be avoided in what follows. Sometreatments of logic rule out vacuous binding and double binding as amatter of syntax. That simplifies some of the treatments below, andcomplicates others.
Free variables correspond to place-holders, while bound variables areused to express generality. If a formula has no free variables, thenit is called asentence. If a formula has free variables, itis calledopen.
Before turning to the deductive system and semantics, we mention a fewfeatures of the language, as developed so far. This helps draw thecontrast between formal languages and natural languages likeEnglish.
We assume at the outset that all of the categories are disjoint. Forexample, no connective is also a quantifier or a variable, and thenon-logical terms are not also parentheses or connectives. Also, theitems within each category are distinct. For example, the sign fordisjunction does not do double-duty as the negation symbol, andperhaps more significantly, no two-place predicate is also a one-placepredicate.
One difference between natural languages like English and formallanguages like \(\LKe\) is that the latter are not supposed to haveany ambiguities. The policy that the different categories of symbolsdo not overlap, and that no symbol does double-duty, avoids the kindof ambiguity, sometimes called “equivocation”, that occurswhen a single word has two meanings: “I’ll meet you at thebank.” But there are other kinds of ambiguity. Consider theEnglish sentence:
John is married, and Mary is single, or Joe is crazy.
It can mean that John is married and either Mary is single or Joe iscrazy, or else it can mean that either both John is married and Maryis single, or else Joe is crazy. An ambiguity like this, due todifferent ways to parse the same sentence, is sometimes called an“amphiboly”. If our formal language did not have theparentheses in it, it would have amphibolies. For example, there wouldbe a “formula” \(A \amp B \vee\) C. Is thissupposed to be \(((A \amp B) \vee C)\), or is it \((A \amp(B \veeC))\)? The parentheses resolve what would be an amphiboly.
Can we be sure that there are no other amphibolies in our language?That is, can we be sure that each formula of \(\LKe\) can be puttogether in only one way? Our next task is to answer thisquestion.
Let us temporarily use the term “unary marker” for thenegation symbol \((\neg)\) or a quantifier followed by a variable(e.g., \(\forall x, \exists z)\).
Lemma 2. Each formula consists of a string of zero ormore unary markers followed by either an atomic formula or a formulaproduced using a binary connective, via one of clauses(3)–(5).
Proof: We proceed by induction on the complexity ofthe formula or, in other words, on the number of formation rules thatare applied. The Lemma clearly holds for atomic formulas. Let \(n\) bea natural number, and suppose that the Lemma holds for any formulaconstructed from \(n\) or fewer instances of clauses (2)–(7).Let \(\theta\) be a formula constructed from \(n+1\) instances. TheLemma holds if the last clause used to construct \(\theta\) was either(3), (4), or (5). If the last clause used to construct \(\theta\) was(2), then \(\theta\) is \(\neg \psi\). Since \(\psi\) was constructedwith \(n\) instances of the rule, the Lemma holds for \(\psi\) (by theinduction hypothesis), and so it holds for \(\theta\). Similarreasoning shows the Lemma to hold for \(\theta\) if the last clausewas (6) or (7). By clause (8), this exhausts the cases, and so theLemma holds for \(\theta\), by induction.
Lemma 3. If a formula \(\theta\) contains a leftparenthesis, then it ends with a right parenthesis, which matches theleftmost left parenthesis in \(\theta\).
Proof: Here we also proceed by induction on thenumber of instances of (2)–(7) used to construct the formula.Clearly, the Lemma holds for atomic formulas, since they have noparentheses. Suppose, then, that the Lemma holds for formulasconstructed with \(n\) or fewer instances of (2)–(7), and let\(\theta\) be constructed with \(n+1\) instances. If the last clauseapplied was (3)–(5), then the Lemma holds since \(\theta\)itself begins with a left parenthesis and ends with the matching rightparenthesis. If the last clause applied was (2), then \(\theta\) is\(\neg \psi\), and the induction hypothesis applies to \(\psi\).Similarly, if the last clause applied was (6) or (7), then \(\theta\)consists of a quantifier, a variable, and a formula to which we canapply the induction hypothesis. It follows that the Lemma holds for\(\theta\).
Lemma 4. Each formula contains at least one atomicformula.
The proof proceeds by induction on the number of instances of(2)–(7) used to construct the formula, and we leave it as anexercise.
Theorem 5. Let \(\alpha, \beta\) be nonemptysequences of characters on our alphabet, such that \(\alpha \beta\)(i.e \(\alpha\) followed by \(\beta)\) is a formula. Then \(\alpha\)isnot a formula.
Proof: By Theorem 1 and Lemma 3, if \(\alpha\)contains a left parenthesis, then the right parenthesis that matchesthe leftmost left parenthesis in \(\alpha \beta\) comes at the end of\(\alpha \beta\), and so the matching right parenthesis is in\(\beta\). So, \(\alpha\) has more left parentheses than rightparentheses. By Theorem \(1, \alpha\) is not a formula. So now supposethat \(\alpha\) does not contain any left parentheses. By Lemma \(2,\alpha \beta\) consists of a string of zero or more unary markersfollowed by either an atomic formula or a formula produced using abinary connective, via one of clauses (3)–(5). If the latterformula was produced via one of clauses (3)–(5), then it beginswith a left parenthesis. Since \(\alpha\) does not contain anyparentheses, it must be a string of unary markers. But then \(\alpha\)does not contain any atomic formulas, and so by Lemma \(4, \alpha\) isnot a formula. The only case left is where \(\alpha \beta\) consistsof a string of unary markers followed by an atomic formula, either inthe form \(t_1 =t_2\) or \(Pt_1 \ldots t_n\). Again, if \(\alpha\)just consisted of unary markers, it would not be a formula, and so\(\alpha\) must consist of the unary markers that start \(\alpha\beta\), followed by either \(t_1\) by itself, \(t_1 =\) by itself, orthe predicate letter \(P\), and perhaps some (but not all) of theterms \(t_1, \ldots,t_n\). In the first two cases, \(\alpha\) does notcontain an atomic formula, by the policy that the categories do notoverlap. Since \(P\) is an \(n\)-place predicate letter, by the policythat the predicate letters are distinct, \(P\) is not an \(m\)-placepredicate letter for any \(m \ne n\). So the part of \(\alpha\) thatconsists of \(P\) followed by the terms is not an atomic formula. Inall of these cases, then, \(\alpha\) does not contain an atomicformula. By Lemma \(4, \alpha\) is not a formula.
We are finally in position to show that there is no amphiboly in ourlanguage.
Theorem 6. Let \(\theta\) be any formula of \(\LKe\).If \(\theta\) is not atomic, then there is one and only one among(2)–(7) that was the last clause applied to construct\(\theta\). That is, \(\theta\) could not be produced by two differentclauses. Moreover, no formula produced by clauses (2)–(7) isatomic.
Proof: By Clause (8), either \(\theta\) is atomic orit was produced by one of clauses (2)–(7). Thus, the firstsymbol in \(\theta\) must be either a predicate letter, a term, aunary marker, or a left parenthesis. If the first symbol in \(\theta\)is a predicate letter or term, then \(\theta\) is atomic. In thiscase, \(\theta\) was not produced by any of (2)–(7), since allsuch formulas begin with something other than a predicate letter orterm. If the first symbol in \(\theta\) is a negation sign“\(\neg\)”, then was \(\theta\) produced by clause (2),and not by any other clause (since the other clauses produce formulasthat begin with either a quantifier or a left parenthesis). Similarly,if \(\theta\) begins with a universal quantifier, then it was producedby clause (6), and not by any other clause, and if \(\theta\) beginswith an existential quantifier, then it was produced by clause (7),and not by any other clause. The only case left is where \(\theta\)begins with a left parenthesis. In this case, it must have beenproduced by one of (3)–(5), and not by any other clause. We onlyneed to rule out the possibility that \(\theta\) was produced by morethan one of (3)–(5). To take an example, suppose that \(\theta\)was produced by (3) and (4). Then \(\theta\) is \((\psi_1 \amp\psi_2)\) and \(\theta\) is also \((\psi_3 \vee \psi_4)\), where\(\psi_1, \psi_2, \psi_3\), and \(\psi_4\) are themselves formulas.That is, \((\psi_1 \amp \psi_2)\) is the very same formula as\((\psi_3 \vee \psi_4)\). By Theorem \(5, \psi_1\) cannot be a properpart of \(\psi_3\), nor can \(\psi_3\) be a proper part of \(\psi_1\).So \(\psi_1\) must be the same formula as \(\psi_3\). But then“\(\amp\)” must be the same symbol as“\(\vee\)”, and this contradicts the policy that all ofthe symbols are different. So \(\theta\) was not produced by bothClause (3) and Clause (4). Similar reasoning takes care of the othercombinations.
This result is sometimes called “unique readability”. Itshows that each formula is produced from the atomic formulas via thevarious clauses in exactly one way. If \(\theta\) was produced byclause (2), then itsmain connective is the initial“\(\neg\)”. If \(\theta\) was produced by clauses (3),(4), or (5), then itsmain connective is the introduced“\(\amp\)”, “\(\vee\)”, or“\(\rightarrow\)”, respectively. If \(\theta\) wasproduced by clauses (6) or (7), then itsmain connective isthe initial quantifier. We apologize for the tedious details. Weincluded them to indicate the level of precision and rigor for thesyntax.
We now introduce adeductive system, \(D\), for ourlanguages. As above, we define anargument to be a non-emptycollection of sentences in the formal language, one of which isdesignated to be theconclusion. If there are any othersentences in the argument, they are itspremises.[1] By convention, we use “\(\Gamma\)”,“\(\Gamma'\)”, “\(\Gamma_1\)”, etc, to rangeover sets of sentences, and we use the letters “\(\phi\)”,“\(\psi\)”, “\(\theta\)”, uppercase orlowercase, with or without subscripts, to range over single sentences.We write “\(\Gamma, \Gamma'\)” for the union of \(\Gamma\)and \(\Gamma'\), and “\(\Gamma, \phi\)” for the union of\(\Gamma\) with \(\{\phi\}\).
We write an argument in the form \(\langle \Gamma, \phi \rangle\),where \(\Gamma\) is a set of sentences, the premises, and \(\phi\) isa single sentence, the conclusion. Remember that \(\Gamma\) may beempty. We write \(\Gamma \vdash \phi\) to indicate that \(\phi\) isdeducible from \(\Gamma\), or, in other words, that the argument\(\langle \Gamma, \phi \rangle\) is deducible in \(D\). We may write\(\Gamma \vdash_D \phi\) to emphasize the deductive system \(D\). Wewrite \(\vdash \phi\) or \(\vdash_D \phi\) to indicate that \(\phi\)can be deduced (in \(D)\) from the empty set of premises.
The rules in \(D\) are chosen to match logical relations concerningthe English analogues of the logical terminology in the language.Again, we define the deducibility relation by recursion. We start witha rule of assumptions:
We thus have that \(\{\phi \}\vdash \phi\); each premise follows fromitself. We next present two clauses for each connective andquantifier. The clauses indicate how to “introduce” and“eliminate” sentences in which each symbol is the mainconnective.
First, recall that “\(\amp\)” is an analogue of theEnglish connective “and”. Intuitively, one can deduce asentence in the form \((\theta \amp \psi)\) if one has deduced\(\theta\) and one has deduced \(\psi\). Conversely, one can deduce\(\theta\) from \((\theta \amp \psi)\) and one can deduce \(\psi\)from \((\theta \amp \psi)\):
The name “&I” stands for“&-introduction”; “&E” stands for“&-elimination”.
Since, the symbol “\(\vee\)” corresponds to the English“or”, \((\theta \vee \psi)\) should be deducible from\(\theta\), and \((\theta \vee \psi)\) should also be deducible from\(\psi\):
The elimination rule is a bit more complicated. Suppose that“\(\theta\) or \(\psi\)” is true. Suppose also that\(\phi\) follows from \(\theta\) and that \(\phi\) follows from\(\psi\). One can reason that if \(\theta\) is true, then \(\phi\) istrue. If instead \(\psi\) is true, we still have that \(\phi\) istrue. So either way, \(\phi\) must be true.
For the next clauses, recall that the symbol,“\(\rightarrow\)”, is an analogue of the English “if… then … ” construction. If one knows, or assumes\((\theta \rightarrow \psi)\) and also knows, or assumes \(\theta\),then one can conclude \(\psi\). Conversely, if one deduces \(\psi\)from an assumption \(\theta\), then one can conclude that \((\theta\rightarrow \psi)\).
This elimination rule is sometimes called “modus ponens”.In some logic texts, the introduction rule is proved as a“deduction theorem”.
Our next clauses are for the negation sign, “\(\neg\)”.The underlying idea is that a sentence \(\psi\) is inconsistent withits negation \(\neg \psi\). They cannot both be true. We call a pairof sentences \(\psi, \neg \psi\)contradictory opposites. Ifone can deduce such a pair from an assumption \(\theta\), then one canconclude that \(\theta\) is false, or, in other words, one canconclude \(\neg \theta\).
By (As), we have that \(\{A,\neg A\}\vdash A\) and\(\{\)A,\(\neg\)A\(\}\vdash \neg A\). So by \(\neg\)I we havethat \(\{A\}\vdash \neg \neg A\). However, we do not have the converseyet. Intuitively, \(\neg \neg \theta\) corresponds to “it is notthe case that it is not the case that” . One might think thatthis last is equivalent to \(\theta\), and we have a rule to thateffect:
The name DNE stands for “double-negation elimination”.There is some controversy over this inference. It is rejected byphilosophers and mathematicians who do not hold that each meaningfulsentence is either true or not true.Intuitionistic logicdoes not sanction the inference in question (see, for example Dummett[2000], or the entry onintuitionistic logic, orhistory of intuitionistic logic), but, again, classical logic does.
To illustrate the parts of the deductive system \(D\) presented thusfar, we show that \(\vdash(A \vee \neg A)\):
The principle \((\theta \vee \neg \theta)\) is sometimes called thelaw of excluded middle. It is not valid in intuitionisticlogic.
Let \(\theta, \neg \theta\) be a pair of contradictory opposites, andlet \(\psi\) be any sentence at all. By (As) we have \(\{\theta, \neg\theta, \neg \psi \}\vdash \theta\) and \(\{\theta, \neg \theta, \neg\psi \}\vdash \neg \theta\). So by \((\neg\)I), \(\{\theta, \neg\theta \}\vdash \neg \neg \psi\). So, by (DNE) we have \(\{\theta ,\neg \theta \}\vdash \psi\) . That is, anything at all follows from apair of contradictory opposites. Some logicians introduce a rule tocodify a similar inference:
If \(\Gamma_1 \vdash \theta\) and \(\Gamma_2 \vdash \neg \theta\),then for any sentence \(\psi, \Gamma_1, \Gamma_2 \vdash \psi\)
The inference is sometimes calledex falso quodlibet or, morecolorfully,explosion. Some call it“\(\neg\)-elimination”, but perhaps this stretches thenotion of “elimination” a bit. We do not officiallyincludeex falso quodlibet as a separate rule in \(D\), butas will be shown below (Theorem 10), each instance of it is derivablein our system \(D\).
Some logicians object toex falso quodlibet, on the groundthat the sentence \(\psi\) may beirrelevant to any of thepremises in \(\Gamma\). Suppose, for example, that one starts withsome premises \(\Gamma\) about human nature and facts about certainpeople, and then deduces both the sentence “Clinton hadextra-marital sexual relations” and “Clinton did not haveextra-marital sexual relations”. One can perhaps conclude thatthere is something wrong with the premises \(\Gamma\). But should webe allowed to then deduceanything at all from \(\Gamma\)?Should we be allowed to deduce “The economy is sound”?
A small minority of logicians, calleddialetheists, hold thatsome contradictions are actually true. For them,ex falsoquodlibet is not truth-preserving.
Deductive systems that demur fromex falso quodlibet arecalledparaconsistent. Most relevant logics areparaconsistent. See the entries onrelevance logic,paraconsistent logic, anddialetheism. Or see Anderson and Belnap [1975], Anderson, Belnap, and Dunn [1992],and Tennant [1997] for fuller overviews of relevant logic; and Priest[2006a,b], for dialetheism. Deep philosophical issues concerning thenature oflogical consequence are involved. Far be it for an article in a philosophy encyclopediato avoid philosophical issues, but space considerations preclude afuller treatment of this issue here. Suffice it to note that theinferenceex falso quodlibet is sanctioned in systems ofclassical logic, the subject of this article. It is essentialto establishing the balance between the deductive system and thesemantics (see §5 below).
The next pieces of \(D\) are the clauses for the quantifiers. Let\(\theta\) be a formula, \(v\) a variable, and \(t\) a term (i.e., avariable or a constant). Then define \(\theta(v|t)\) to be the resultof substituting \(t\) for eachfree occurrence of \(v\) in\(\theta\). So, if \(\theta\) is \((Qx \amp \exists\)xPxy),then \(\theta(x|c)\) is \((Qc \amp \exists\)xPxy). The lastoccurrence of \(x\) is not free.
A sentence in the form \(\forall v \theta\) is an analogue of theEnglish “for every \(v, \theta\) holds”. So one should beable to infer \(\theta(v|t)\) from \(\forall v \theta\) for any closedterm \(t\). Recall that the only closed terms in our system areconstants.
The idea here is that if \(\forall v \theta\) is true, then \(\theta\)should hold of \(t\), no matter what \(t\) is.
The introduction clause for the universal quantifier is a bit morecomplicated. Suppose that a sentence \(\theta\) contains a closed term\(t\), and that \(\theta\) has been deduced from a set of premises\(\Gamma\). If the closed term \(t\) does not occur in any member of\(\Gamma\), then \(\theta\) will hold no matter which object \(t\) maydenote. That is, \(\forall v \theta\) follows.
This rule \((\forall \mathbf{I})\) corresponds to a common inferencein mathematics. Suppose that a mathematician says “let \(n\) bea natural number” and goes on to show that \(n\) has a certainproperty \(P\), without assuming anything about \(n\) (except that itis a natural number). She then reminds the reader that \(n\) is“arbitrary”, and concludes that \(P\) holds forall natural numbers. The condition that the term \(t\) notoccur in any premise is what guarantees that it is indeed“arbitrary”. It could be any object, and so anything weconclude about it holds for all objects.
The existential quantifier is an analogue of the English expression“there exists”, or perhaps just “there is”. Ifwe have established (or assumed) that a given object \(t\) has a givenproperty, then it follows that there is something that has thatproperty.
The elimination rule for \(\exists\) is not quite as simple:
This elimination rule also corresponds to a common inference. Supposethat a mathematician assumes or somehow concludes that there is anatural number with a given property \(P\). She then says “let\(n\) be such a natural number, so that \(Pn\)”, and goes on toestablish a sentence \(\phi\), which does not mention the number\(n\). If the derivation of \(\phi\) does not invoke anything about\(n\) (other than the assumption that it has the given property\(P)\), then \(n\) could have been any number that has the property\(P\). That is, \(n\) is anarbitrary number with property\(P\). It does not matter which number \(n\) is. Since \(\phi\) doesnot mention \(n\), it follows from the assertion that something hasproperty \(P\). The provisions added to \((\exists\)E) are toguarantee that \(t\) is “arbitrary”.
The final items are the rules for the identity sign “=”.The introduction rule is about a simple as can be:
This “inference” corresponds to the truism that everythingis identical to itself. The elimination rule corresponds to aprinciple that if \(a\) is identical to \(b\), then anything true of\(a\) is also true of \(b\).
The rule \(({=}\mathrm{E})\) indicates a certain restriction in theexpressive resources of our language. Suppose, for example, that Harryis identical to Donald (since his mischievous parents gave him twonames). According to most people’s intuitions, it would notfollow from this and “Dick knows that Harry is wicked”that “Dick knows that Donald is wicked”, for the reasonthat Dick might not know that Harry is identical to Donald. Contextslike this, in which identicals cannot safely be substituted for eachother, are called “opaque”. We assume that our language\(\LKe\) has no opaque contexts.
One final clause completes the description of the deductive system\(D\):
Again, this clause allows proofs by induction on the rules used toestablish an argument. If a property of arguments holds of allinstances of (As) and \(({=}\mathrm{I})\), and if the other rulespreserve the property, then every argument that is deducible in \(D\)enjoys the property in question.
Before moving on to the model theory for \(\LKe\), we pause to note afew features of the deductive system. To illustrate the level ofrigor, we begin with a lemma that if a sentence does not contain aparticular closed term, we can make small changes to the set ofsentences we prove it from without problems. We allow ourselves theliberty here of extending some previous notation: for any terms \(t\)and \(t'\), and any formula \(\theta\), we say that \(\theta(t|t')\)is the result of replacing all free occurrences of \(t\) in \(\theta\)with \(t'\).
Lemma 7. If \(\Gamma_1\) and \(\Gamma_2\) differ onlyin that wherever \(\Gamma_1\) contains \(\theta\), \(\Gamma_2\)contains \(\theta(t|t')\), then for any sentence \(\phi\) notcontaining \(t\) or \(t'\), if \(\Gamma_1\vdash\phi\) then\(\Gamma_2\vdash\phi\).
Proof: The proof proceeds by induction on the numberof steps in the proof of \(\phi\). Crucial to this proof is the factthat \(\theta=\theta(t|t')\) whenever \(\theta\) does not contain\(t\) or \(t'\). When the number of steps in the proof of \(\phi\) isone, this means that the last (and only) rule applied is (As) or (=I).Then, since \(\phi\) does not contain \(t\) or \(t'\), if\(\Gamma_1\vdash\phi\) we simply apply the same rule ((As) or (=I)) to\(\Gamma_2\) to get \(\Gamma_2\vdash\phi\). Assume that there are\(n>1\) steps in the proof of \(\phi\), and that Lemma 7 holds for anyproof with less than \(n\) steps. Suppose that the \(n^{th}\) ruleapplied to \(\Gamma_1\) was (\(\amp I\)). Then \(\phi\) is\(\psi\amp\chi\), and \(\Gamma_1\vdash\phi\amp\chi\). But then we knowthat previous steps in the proof include \(\Gamma_1\vdash\psi\) and\(\Gamma_1\vdash\chi\), and by induction, we have\(\Gamma_2\vdash\psi\) and \(\Gamma_2\vdash\chi\), since neither\(\psi\) nor \(\chi\) contain \(t\) or \(t'\). So, we simply apply(\(\amp I\)) to \(\Gamma_2\) to get \(\Gamma_2\vdash\psi\amp\chi\) asrequired. Suppose now that the last step applied in the proof of\(\Gamma_1\vdash\phi\) was (\(\amp E\)). Then, at a previous step inthe proof of \(\phi\), we know \(\Gamma_1\vdash\phi\amp\psi\) for somesentence \(\psi\). If \(\psi\) does not contain \(t\), then we simplyapply (\(\amp E\)) to \(\Gamma_2\) to obtain the desired result. Theonly complication is if \(\psi\) contains \(t\). Then we would havethat \(\Gamma_2\vdash (\phi\amp\psi)(t|t')\). But, since\((\phi\amp\psi)(t|t')\) is \(\phi(t|t')\amp\psi(t|t')\), and\(\phi(t|t')\) is just \(\phi\), we can just apply (\(\amp E\)) to get\(\Gamma_2\vdash\phi\) as required. The cases for the other rules aresimilar.
Theorem 8. The rule of Weakening. If \(\Gamma_1\vdash \phi\) and \(\Gamma_1 \subseteq \Gamma_2\), then \(\Gamma_2\vdash \phi\).
Proof: Again, we proceed by induction on the numberof rules that were used to arrive at \(\Gamma_1 \vdash \phi\). Supposethat \(n\gt 0\) is a natural number, and that the theorem holds forany argument that was derived using fewer than \(n\) rules. Supposethat \(\Gamma_1 \vdash \phi\) using exactly \(n\) rules. If \(n=1\),then the rule is either (As) or \((=\)I). In these cases, \(\Gamma_2\vdash \phi\) by the same rule. If the last rule applied was (&I),then \(\phi\) has the form \((\theta \amp \psi)\), and we have\(\Gamma_3 \vdash \theta\) and \(\Gamma_4 \vdash \psi\), with\(\Gamma_1 = \Gamma_3, \Gamma_4\). We apply the induction hypothesisto the deductions of \(\theta\) and \(\psi\), to get \(\Gamma_2 \vdash\theta\) and \(\Gamma_2 \vdash \psi\). and then apply (&I) to theresult to get \(\Gamma_2 \vdash \phi\). Most of the other cases areexactly like this. Slight complications arise only in the rules\((\forall\)I) and \((\exists\)E), because there we have to payattention to the conditions for the rules.
Suppose that the last rule applied to get \(\Gamma_1 \vdash \phi\) is\((\forall\)I). So \(\phi\) is a sentence of the form \(\forallv\theta\), and we have \(\Gamma_1 \vdash \theta (v|t)\) and \(t\) doesnot occur in any member of \(\Gamma_1\) or in \(\theta\). The problemis that \(t\) may occur in a member of \(\Gamma_2\), and so we cannotjust invoke the induction hypothesis and apply \((\forall\)I) to theresult. So, let \(t'\) be a term not occurring in any sentence in\(\Gamma_2\). Let \(\Gamma'\) be the result of substituting \(t'\) forall \(t\) in \(\Gamma_2\). Then, since \(t\) does not occur in\(\Gamma_1\), \(\Gamma_1\subseteq\Gamma'\). So, the inductionhypothesis gives us \(\Gamma'\vdash\theta (v|t)\), and we know that\(\Gamma'\) does not contain \(t\), so we can apply (\(\forall I\)) toget \(\Gamma'\vdash\forall v\theta\). But \(\forall v\theta\) does notcontain \(t\) or \(t'\), so \(\Gamma_2\vdash\forall v\theta\) by Lemma7.
Suppose that the last rule applied was \((\exists\)E), we have\(\Gamma_3 \vdash \exists v\theta\) and \(\Gamma_4, \theta (v|t)\vdash \phi\), with \(\Gamma_1\) being \(\Gamma_3, \Gamma_4\), and\(t\) not in \(\phi\), \(\Gamma_4\) or \(\theta\). If \(t\) does notoccur free in \(\Gamma_2\), we apply the induction hypothesis to get\(\Gamma_2 \vdash \exists v\theta\), and then \((\exists\)E) to end upwith \(\Gamma_2 \vdash \phi\). If \(t\) does occur free in\(\Gamma_2\), then we follow a similar proceedure to \(\forall I\),using Lemma 7.
Theorem 8 allows us to add on premises at will. It follows that\(\Gamma \vdash \phi\) if and only if there is a subset\(\Gamma'\subseteq \Gamma\) such that \(\Gamma'\vdash \phi\). Somesystems of relevant logic do not have weakening, nor doessubstructural logic (See the entries onrelevance logic,substructural logics, andlinear logic).
By clause (*), all derivations are established in a finite number ofsteps. So we have
Theorem 9. \(\Gamma \vdash \phi\) if and only ifthere is a finite \(\Gamma'\subseteq \Gamma\) such that\(\Gamma'\vdash \phi\).
Theorem 10. The rule ofex falso quodlibetis a “derived rule” of \(D\): if \(\Gamma_1 \vdash\theta\) and \(\Gamma_2 \vdash \neg \theta\), then \(\Gamma_1,\Gamma_2\vdash \psi\), for any sentence \(\psi\).
Proof: Suppose that \(\Gamma_1 \vdash \theta\) and\(\Gamma_2 \vdash \neg \theta\). Then by Theorem \(8, \Gamma_1,\neg\psi \vdash \theta\), and \(\Gamma_2,\neg \psi \vdash \neg \theta\).So by \((\neg\)I), \(\Gamma_1, \Gamma_2 \vdash \neg \neg \psi\). By(DNE), \(\Gamma_1, \Gamma_2 \vdash \psi\).
Theorem 11. The rule of Cut. If \(\Gamma_1 \vdash\psi\) and \(\Gamma_2, \psi \vdash \theta\), then \(\Gamma_1, \Gamma_2\vdash \theta\).
Proof: Suppose \(\Gamma_1 \vdash \psi\) and\(\Gamma_2, \psi \vdash \theta\). We proceed by induction on thenumber of rules used to establish \(\Gamma_2, \psi \vdash \theta\).Suppose that \(n\) is a natural number, and that the theorem holds forany argument that was derived using fewer than \(n\) rules. Supposethat \(\Gamma_2, \psi \vdash \theta\) was derived using exactly \(n\)rules. If the last rule used was \((=\)I), then \(\Gamma_1, \Gamma_2\vdash \theta\) is also an instance of \((=\)I). If \(\Gamma_2, \psi\vdash \theta\) is an instance of (As), then either \(\theta\) is\(\psi\), or \(\theta\) is a member of \(\Gamma_2\). In the formercase, we have \(\Gamma_1 \vdash \theta\) by supposition, and get\(\Gamma_1, \Gamma_2 \vdash \theta\) by Weakening (Theorem 8). In thelatter case, \(\Gamma_1, \Gamma_2 \vdash \theta\) is itself aninstance of (As). Suppose that \(\Gamma_2, \psi \vdash \theta\) wasobtained using (&E). Then we have \(\Gamma_2, \psi \vdash(\theta\amp \phi)\). The induction hypothesis gives us \(\Gamma_1, \Gamma_2\vdash(\theta \amp \phi)\), and (&E) produces \(\Gamma_1, \Gamma_2\vdash \theta\). The remaining cases are similar.
Theorem 11 allows us to chain together inferences. This fits thepractice of establishing theorems and lemmas and then using thosetheorems and lemmas later, at will. The cut principle is, some think,essential to reasoning. In some logical systems, the cut principle isa deep theorem; in others it is invalid. The system here was designed,in part, to make the proof of Theorem 11 straightforward.
If \(\Gamma \vdash_D \theta\), then we say that the sentence\(\theta\) is adeductive consequence of the set of sentences\(\Gamma\), and that the argument \(\langle \Gamma,\theta \rangle\) isdeductively valid. A sentence \(\theta\) is alogicaltheorem, or adeductive logical truth, if \(\vdash_D\theta\). That is, \(\theta\) is a logical theorem if it is adeductive consequence of the empty set. A set \(\Gamma\) of sentencesisconsistent if there is no sentence \(\theta\) such that\(\Gamma \vdash_D \theta\) and \(\Gamma \vdash_D \neg \theta\). Thatis, a set is consistent if it does not entail a pair of contradictoryopposite sentencess.
Theorem 12. A set \(\Gamma\) is consistent if andonly if there is a sentence \(\theta\) such that it is not the casethat \(\Gamma \vdash \theta\).
Proof: Suppose that \(\Gamma\) is consistent and let\(\theta\) be any sentence. Then either it is not the case that\(\Gamma \vdash \theta\) or it is not the case that \(\Gamma \vdash\neg \theta\). For the converse, suppose that \(\Gamma\) isinconsistent and let \(\psi\) be any sentence. We have that there is asentence such that both \(\Gamma \vdash \theta\) and \(\Gamma \vdash\neg \theta\). Byex falso quodlibet (Theorem 10), \(\Gamma\vdash \psi\).
Define a set \(\Gamma\) of sentences of the language \(\LKe\) to bemaximally consistent if \(\Gamma\) is consistent and forevery sentence \(\theta\) of \(\LKe\), if \(\theta\) is not in\(\Gamma\), then \(\Gamma,\theta\) is inconsistent. In other words,\(\Gamma\) is maximally consistent if \(\Gamma\) is consistent, andadding any sentence in the language not already in \(\Gamma\) rendersit inconsistent. Notice that if \(\Gamma\) is maximally consistentthen \(\Gamma \vdash \theta\) if and only if \(\theta\) is in\(\Gamma\).
Theorem 13. The Lindenbaum Lemma. Let \(\Gamma\) beany consistent set of sentences of \(\LKe .\) Then there is a set\(\Gamma'\) of sentences of \(\LKe\) such that \(\Gamma \subseteq\Gamma'\) and \(\Gamma'\) is maximally consistent.
Proof: Although this theorem holds in general, weassume here that the set \(K\) of non-logical terminology is eitherfinite or denumerably infinite (i.e., the size of the natural numbers,usually called \(\aleph_0)\). It follows that there is an enumeration\(\theta_0, \theta_1,\ldots\) of the sentences of \(\LKe\), such thatevery sentence of \(\LKe\) eventually occurs in the list. Define asequence of sets of sentences, by recursion, as follows: \(\Gamma_0\)is \(\Gamma\); for each natural number \(n\), if \(\Gamma_n,\theta_n\) is consistent, then let \(\Gamma_{n+1} = \Gamma_n,\theta_n\). Otherwise, let \(\Gamma_{n+1} = \Gamma_n\). Let\(\Gamma'\) be the union of all of the sets \(\Gamma_n\). Intuitively,the idea is to go through the sentences of \(\LKe\), throwing each oneinto \(\Gamma'\) if doing so produces a consistent set. Notice thateach \(\Gamma_n\) is consistent. Suppose that \(\Gamma'\) isinconsistent. Then there is a sentence \(\theta\) such that\(\Gamma'\vdash \theta\) and \(\Gamma'\vdash \neg \theta\). By Theorem9 and Weakening (Theorem 8), there is finite subset \(\Gamma''\) of\(\Gamma'\) such that \(\Gamma''\vdash \theta\) and \(\Gamma''\vdash\neg \theta\). Because \(\Gamma''\) is finite, there is a naturalnumber \(n\) such that every member of \(\Gamma''\) is in\(\Gamma_n\). So, by Weakening again, \(\Gamma_n \vdash \theta\) and\(\Gamma_n \vdash \neg \theta\). So \(\Gamma_n\) is inconsistent,which contradicts the construction. So \(\Gamma'\) is consistent. Nowsuppose that a sentence \(\theta\) is not in \(\Gamma'\). We have toshow that \(\Gamma', \theta\) is inconsistent. The sentence \(\theta\)must occur in the aforementioned list of sentences; say that\(\theta\) is \(\theta_m\). Since \(\theta_m\) is not in \(\Gamma'\),then it is not in \(\Gamma_{m+1}\). This happens only if \(\Gamma_m,\theta_m\) is inconsistent. So a pair of contradictory opposites canbe deduced from \(\Gamma_m,\theta_m\). By Weakening, a pair ofcontradictory opposites can be deduced from \(\Gamma', \theta_m\). So\(\Gamma', \theta_m\) is inconsistent. Thus, \(\Gamma'\) is maximallyconsistent.
Notice that this proof uses a principle corresponding to the law ofexcluded middle. In the construction of \(\Gamma'\), we assumed that,at each stage, either \(\Gamma_n\) is consistent or it is not.Intuitionists, who demur from excluded middle, do not accept theLindenbaum lemma.
Let \(K\) be a set of non-logical terminology. Aninterpretation for the language \(\LKe\) is a structure \(M =\langle d,I\rangle\), where \(d\) is a non-empty set, called thedomain-of-discourse, or simply thedomain, of theinterpretation, and \(I\) is aninterpretation function.Informally, the domain is what we interpret the language \(\LKe\) tobe about. It is what the variables range over. The interpretationfunction assigns appropriate extensions to the non-logical terms. Inparticular,
If \(c\) is a constant in \(K\), then \(I(c)\) is a member of thedomain \(d\).
Thus we assume that every constant denotes something. Systems wherethis is not assumed are calledfree logics (see the entry onfree logic). Continuing,
If \(P^0\) is a zero-place predicate letter in \(K\), then \(I(P)\) isa truth value, either truth or falsehood.
If \(Q^1\) is a one-place predicate letter in \(K\), then \(I(Q)\) isa subset of \(d\). Intuitively, \(I(Q)\) is the set of members of thedomain that the predicate \(Q\) holds of. For example, \(I(Q)\) mightbe the set of red members of the domain.
If \(R^2\) is a two-place predicate letter in \(K\), then \(I(R)\) isa set of ordered pairs of members of \(d\). Intuitively, \(I(R)\) isthe set of pairs of members of the domain that the relation \(R\)holds between. For example, \(I(R)\) might be the set of pairs\(\langle a,b\rangle\) such that \(a\) and \(b\) are the members ofthe domain for which \(a\) loves \(b\).
In general, ifS\(^n\) is an \(n\)-place predicate letter in\(K\), then \(I(S)\) is a set of ordered \(n\)-tuples of members of\(d\).
Define \(s\) to be avariable-assignment, or simply anassignment, on an interpretation \(M\), if \(s\) is afunction from the variables to the domain \(d\) of \(M\). The role ofvariable-assignments is to assign denotations to thefreevariables of open formulas. (In a sense, the quantifiers determine the“meaning” of the bound variables.)
Let \(t\) be a term of \(\LKe\). We define thedenotation of\(t\) in \(M\) under \(s\), in terms of the interpretation functionand variable-assignment:
If \(t\) is a constant, then \(D_{M,s}(t)\) is \(I(t)\), and if \(t\)is a variable, then \(D_{M,s}(t)\) is \(s(t)\).
That is, the interpretation \(M\) assigns denotations to theconstants, while the variable-assignment assigns denotations to the(free) variables. If the language contained function symbols, thedenotation function would be defined by recursion.
We now define a relation ofsatisfaction betweeninterpretations, variable-assignments, and formulas of \(\LKe\). If\(\phi\) is a formula of \(\LKe, M\) is an interpretation for\(\LKe\), and \(s\) is a variable-assignment on \(M\), then we write\(M,s\vDash \phi\) for \(M\)satisfies \(\phi\)under theassignment \(s\). The idea is that \(M,s\vDash \phi\) is ananalogue of “\(\phi\) comes out true when interpreted as in\(M\) via \(s\)”.
We proceed by recursion on the complexity of the formulas of\(\LKe\).
If \(t_1\) and \(t_2\) are terms, then \(M,s\vDash t_1 =t_2\) if andonly if \(D_{M,s}(t_1)\) is the same as \(D_{M,s}(t_2)\).
This is about as straightforward as it gets. An identity \(t_1 =t_2\)comes out true if and only if the terms \(t_1\) and \(t_2\) denote thesame thing.
If \(P^0\) is a zero-place predicate letter in \(K\), then \(M,s\vDashP\) if and only if \(I(P)\) is truth.
IfS\(^n\) is an \(n\)-place predicate letter in \(K\) and\(t_1, \ldots,t_n\) are terms, then \(M,s\vDash St_1 \ldots t_n\) ifand only if the \(n\)-tuple \(\langle D_{M,s}(t_1),\ldots,D_{M,s}(t_n)\rangle\) is in \(I(S)\).
This takes care of the atomic formulas. We now proceed to the compoundformulas of the language, more or less following the meanings of theEnglish counterparts of the logical terminology.
\(M,s\vDash \neg \theta\) if and only if it is not the case that\(M,s\vDash \theta\).
\(M,s\vDash(\theta \amp \psi)\) if and only if both \(M,s\vDash\theta\) and \(M,s\vDash \psi\).
\(M,s\vDash(\theta \vee \psi)\) if and only if either \(M,s\vDash\theta\) or \(M,s\vDash \psi\).
\(M,s\vDash(\theta \rightarrow \psi)\) if and only if either it is notthe case that \(M,s\vDash \theta\), or \(M,s\vDash \psi\).
\(M,s\vDash \forall v\theta\) if and only if \(M,s'\vDash \theta\),for every assignment \(s'\) that agrees with \(s\) except possibly atthe variable \(v\).
The idea here is that \(\forall v\theta\) comes out true if and onlyif \(\theta\) comes out true no matter what is assigned to thevariable \(v\). The final clause is similar.
\(M,s\vDash \exists v\theta\) if and only if \(M,s'\vDash \theta\),for some assignment \(s'\) that agrees with \(s\) except possibly atthe variable \(v\).
So \(\exists v\theta\) comes out true if there is an assignment to\(v\) that makes \(\theta\) true.
Theorem 6, unique readability, assures us that this definition iscoherent. At each stage in breaking down a formula, there is exactlyone clause to be applied, and so we never get contradictory verdictsconcerning satisfaction.
As indicated, the role of variable-assignments is to give denotationsto the free variables. We now show that variable-assignments play noother role.
Theorem 14. For any formula \(\theta\), if \(s_1\)and \(s_2\) agree on the free variables in \(\theta\), then \(M,s_1\vDash \theta\) if and only if \(M,s_2 \vDash \theta\).
Proof: We proceed by induction on the complexity ofthe formula \(\theta\). The theorem clearly holds if \(\theta\) isatomic, since in those cases only the values of thevariable-assignments at the variables in \(\theta\) figure in thedefinition. Assume, then, that the theorem holds for all formulas lesscomplex than \(\theta\). And suppose that \(s_1\) and \(s_2\) agree onthe free variables of \(\theta\). Assume, first, that \(\theta\) is anegation, \(\neg \psi\). Then, by the induction hypothesis, \(M,s_1\vDash \psi\) if and only if \(M,s_2 \vDash \psi\). So, by the clausefor negation, \(M,s_1 \vDash \neg \psi\) if and only if \(M,s_2 \vDash\neg \psi\). The cases where the main connective in \(\theta\) isbinary are also straightforward. Suppose that \(\theta\) is \(\existsv\psi\), and that \(M,s_1 \vDash \exists v\psi\). Then there is anassignment \(s_1'\) that agrees with \(s_1\) except possibly at \(v\)such that \(M,s_1'\vDash \psi\). Let \(s_2'\) be the assignment thatagrees with \(s_2\) on the free variables not in \(\psi\) and agreeswith \(s_1'\) on the others. Then, by the induction hypothesis,\(M,s_2'\vDash \psi\). Notice that \(s_2'\) agrees with \(s_2\) onevery variable except possibly \(v\). So \(M,s_2 \vDash \existsv\psi\). The converse is the same, and the case where \(\theta\)begins with a universal quantifier is similar.
By Theorem 14, if \(\theta\) is a sentence, and \(s_1, s_2\), are anytwo variable-assignments, then \(M,s_1 \vDash \theta\) if and only if\(M,s_2 \vDash \theta\). So we can just write \(M\vDash \theta\) if\(M,s\vDash \theta\) for some, or all, variable-assignments \(s\). Sowe define
\(M\vDash \theta\) where \(\theta\) is a sentence just in case\(M,s\vDash\theta\) for all variable assignments \(s\).
In this case, we call \(M\) amodel of \(\theta\).
Suppose that \(K'\subseteq K\) are two sets of non-logical terms. If\(M = \langle d,I\rangle\) is an interpretation of \(\LKe\), then wedefine therestriction of \(M\) to \(\mathcal{L}1K'{=}\) tobe the interpretation \(M'=\langle d,I'\rangle\) such that \(I'\) isthe restriction of \(I\) to \(K'\). That is, \(M\) and \(M'\) have thesame domain and agree on the non-logical terminology in \(K'\). Astraightforward induction establishes the following:
Theorem 15. If \(M'\) is the restriction of \(M\) to\(\mathcal{L}1K'{=}\), then for every sentence \(\theta\) of\(\mathcal{L}1K'\), \(M\vDash\theta\) if and only if \(M'\vDash\theta\).
Theorem 16. If two interpretations \(M_1\) and\(M_2\) have the same domain and agree on all of the non-logicalterminology of a sentence \(\theta\), then \(M_1\vDash\theta\) if andonly if \(M_2\vDash \theta\).
In short, the satisfaction of a sentence \(\theta\) only depends onthe domain of discourse and the interpretation of the non-logicalterminology in \(\theta\).
We say that an argument \(\langle \Gamma,\theta \rangle\) issemantically valid, or justvalid, written \(\Gamma\vDash \theta\), if for every interpretation \(M\) of the language, if\(M\vDash\psi\), for every member \(\psi\) of \(\Gamma\), then\(M\vDash\theta\). If \(\Gamma \vDash \theta\), we also say that\(\theta\) is alogical consequence, orsemanticconsequence, ormodel-theoretic consequence of\(\Gamma\). The definition corresponds to the informal idea that anargument is valid if it is not possible for its premises to all betrue and its conclusion false. Our definition of logical consequencealso sanctions the common thesis that a valid argument istruth-preserving – to the extent that satisfaction representstruth. Officially, an argument in \(\LKe\) is valid if its conclusioncomes out true under every interpretation of the language in which thepremises are true. Validity is the model-theoretic counterpart todeducibility.
A sentence \(\theta\) islogically true, orvalid,if \(M\vDash \theta\), for every interpretation \(M\). A sentence islogically true if and only if it is a consequence of the empty set. If\(\theta\) is logically true, then for any set \(\Gamma\) ofsentences, \(\Gamma \vDash \theta\). Logical truth is themodel-theoretic counterpart of theoremhood.
A sentence \(\theta\) issatisfiable if there is aninterpretation \(M\) such that \(M\vDash \theta\). That is, \(\theta\)is satisfiable if there is an interpretation that satisfies it. A set\(\Gamma\) of sentences is satisfiable if there is an interpretation\(M\) such that \(M\vDash\theta\), for every sentence \(\theta\) in\(\Gamma\). If \(\Gamma\) is a set of sentences and if \(M\vDash\theta\) for each sentence \(\theta\) in \(\Gamma\), then we say that\(M\) is amodel of \(\Gamma\). So a set of sentences issatisfiable if it has a model. Satisfiability is the model-theoreticcounterpart to consistency.
Notice that \(\Gamma \vDash \theta\) if and only if the set\(\Gamma,\neg \theta\) is not satisfiable. It follows that if a set\(\Gamma\) is not satisfiable, then if \(\theta\) is any sentence,\(\Gamma \vDash \theta\). This is a model-theoretic counterpart toex falso quodlibet (see Theorem 10). We have the following,as an analogue to Theorem 12:
Theorem 17. Let \(\Gamma\) be a set of sentences. Thefollowing are equivalent: (a) \(\Gamma\) is satisfiable; (b) there isno sentence \(\theta\) such that both \(\Gamma \vDash \theta\) and\(\Gamma \vDash \neg \theta\); (c) there is some sentence \(\psi\)such that it is not the case that \(\Gamma \vDash \psi\).
Proof: (a)\(\Rightarrow\)(b): Suppose that \(\Gamma\)is satisfiable and let \(\theta\) be any sentence. There is aninterpretation \(M\) such that \(M\vDash \psi\) for every member\(\psi\) of \(\Gamma\). By the clause for negations, we cannot haveboth \(M\vDash \theta\) and \(M\vDash \neg \theta\). So either\(\langle \Gamma,\theta \rangle\) is not valid or else \(\langle\Gamma,\neg \theta \rangle\) is not valid. (b)\(\Rightarrow\)(c): Thisis immediate. (c)\(\Rightarrow\)(a): Suppose that it is not the casethat \(\Gamma \vDash \psi\). Then there is an interpretation \(M\)such that \(M\vDash \theta\), for every sentence \(\theta\) in\(\Gamma\) and it is not the case that \(M\vDash \psi\). A fortiori,\(M\) satisfies every member of \(\Gamma\), and so \(\Gamma\) issatisfiable.
We now present some results that relate the deductive notions to theirmodel-theoretic counterparts. The first one is probably the moststraightforward. We motivated both the various rules of the deductivesystem \(D\) and the various clauses in the definition of satisfactionin terms of the meaning of the English counterparts to the logicalterminology (more or less, with the same simplifications in bothcases). So one would expect that an argument is deducible, ordeductively valid, only if it is semantically valid.
Theorem 18. Soundness. For any sentence \(\theta\)and set \(\Gamma\) of sentences, if \(\Gamma \vdash_D \theta\), then\(\Gamma \vDash \theta\).
Proof: We proceed by induction on the number ofclauses used to establish \(\Gamma \vdash \theta\). So let \(n\) be anatural number, and assume that the theorem holds for any argumentestablished as deductively valid with fewer than \(n\) steps. Andsuppose that \(\Gamma \vdash \theta\) was established using exactly\(n\) steps. If the last rule applied was \((=\)I) then \(\theta\) isa sentence in the form \(t=t\), and so \(\theta\) is logically true. Afortiori, \(\Gamma \vDash \theta\). If the last rule applied was (As),then \(\theta\) is a member of \(\Gamma\), and so of course anyinterpretation that satisfies every member of \(\Gamma\) alsosatisfies \(\theta\). Suppose the last rule applied is (&I). So\(\theta\) has the form \((\phi \amp \psi)\), and we have \(\Gamma_1\vdash \phi\) and \(\Gamma_2 \vdash \psi\), with \(\Gamma = \Gamma_1,\Gamma_2\). The induction hypothesis gives us \(\Gamma_1 \vDash \phi\)and \(\Gamma_2 \vDash \psi\). Suppose that \(M\) satisfies everymember of \(\Gamma\). Then \(M\) satisfies every member of\(\Gamma_1\), and so \(M\) satisfies \(\phi\). Similarly, \(M\)satisfies every member of \(\Gamma_2\), and so \(M\) satisfies\(\psi\). Thus, by the clause for “\(\amp\)” in thedefinition of satisfaction, \(M\) satisfies \(\theta\). So \(\Gamma\vDash \theta\).
Suppose the last clause applied was \((\exists\mathrm{E})\). So wehave \(\Gamma_1 \vdash \exists v\phi\) and \(\Gamma_2, \phi(v|t)\vdash \theta\), where \(\Gamma = \Gamma_1, \Gamma_2\), and \(t\) doesnot occur in \(\phi , \theta \), or in any member of \(\Gamma_2\).
We need to show that \(\Gamma\vDash\theta\). By the inductionhypothesis, we have that \(\Gamma_1\vDash\exists v\phi\) and\(\Gamma_2, \phi(v|t)\vDash\theta\). Let \(M\) be an interpretationsuch that \(M\) makes every member of \(\Gamma\) true. So, \(M\) makesevery member of \(\Gamma_1\) and \(\Gamma_2\) true. Then\(M,s\vDash\exists v\phi\) for all variable assignments \(s\), sothere is an \(s'\) such that \(M,s'\vDash\phi\). Let \(M'\) differfrom \(M\) only in that \(I_{M'}(t)=s'(v)\). Then,\(M',s'\vDash\phi(v|t)\) and \(M',s'\vDash\Gamma_2\) since \(t\) doesnot occur in \(\phi\) or \(\Gamma_2\). So, \(M',s'\vDash\theta\).Since \(t\) does not occur in \(\theta\) and \(M'\) differs from \(M\)only with respect to \(I_{M'}(t)\), \(M,s'\vDash\theta\). Since\(\theta\) is a sentence, \(s'\) doesn't matter, so \(M\vDash\theta\)as desired. Notice the role of the restrictions on \((\exists\)E)here. The other cases are about as straightforward.
Corollary 19. Let \(\Gamma\) be a set of sentences.If \(\Gamma\) is satisfiable, then \(\Gamma\) is consistent.
Proof: Suppose that \(\Gamma\) is satisfiable. So let\(M\) be an interpretation such that \(M\) satisfies every member of\(\Gamma\). Assume that \(\Gamma\) is inconsistent. Then there is asentence \(\theta\) such that \(\Gamma \vdash \theta\) and \(\Gamma\vdash \neg \theta\). By soundness (Theorem 18), \(\Gamma \vDash\theta\) and \(\Gamma \vDash \neg \theta\). So we have that \(M\vDash\theta\) and \(M\vDash \neg \theta\). But this is impossible, giventhe clause for negation in the definition of satisfaction.
Even though the deductive system \(D\) and the model-theoreticsemantics were developed with the meanings of the logical terminologyin mind, one should not automatically expect the converse to soundness(or Corollary 19) to hold. For all we know so far, we may not haveincluded enough rules of inference to deduce every valid argument. Theconverses to soundness and Corollary 19 are among the most importantand influential results in mathematical logic. We begin with thelatter.
Theorem 20. Completeness. Gödel [1930]. Let\(\Gamma\) be a set of sentences. If \(\Gamma\) is consistent, then\(\Gamma\) is satisfiable.
Proof: The proof of completeness is rather complex.We only sketch it here. Let \(\Gamma\) be a consistent set ofsentences of \(\LKe\). Again, we assume for simplicity that the set\(K\) of non-logical terminology is either finite or countablyinfinite (although the theorem holds even if \(K\) is uncountable).The task at hand is to find an interpretation \(M\) such that \(M\)satisfies every member of \(\Gamma\). Consider the language obtainedfrom \(\LKe\) by adding a denumerably infinite stock of new individualconstants \(c_0, c_1,\ldots\) We stipulate that the constants, \(c_0,c_1,\ldots\), are all different from each other and none of them occurin \(K\). One interesting feature of this construction, due to LeonHenkin, is that we build an interpretation of the language from thelanguage itself, using some of the constants as members of the domainof discourse. Let \(\theta_0 (x), \theta_1 (x),\ldots\) be anenumeration of the formulas of the expanded language with at most onefree variable, so that each formula with at most one free variableoccurs in the list eventually. Define a sequence \(\Gamma_0,\Gamma_1,\ldots\) of sets of sentences (of the expanded language) byrecursion as follows: \(\Gamma_0 = \Gamma\); and \(\Gamma_{n+1} =\Gamma_n,(\exists x\theta_n \rightarrow \theta_{n}(x|c_i))\), where\(c_i\) is the first constant in the above list that does not occur in\(\theta_n\) or in any member of \(\Gamma_n\). The underlying ideahere is that if \(\exists x\theta_n\)is true, then \(c_i\) is to beone such \(x\). Let \(\Gamma'\) be the union of the sets \(\Gamma_n\).
We sketch a proof that \(\Gamma'\) is consistent. Suppose that\(\Gamma'\) is inconsistent. By Theorem 9, there is a finite subset of\(\Gamma\) that is inconsistent, and so one of the sets \(\Gamma_m\)is inconsistent. By hypothesis, \(\Gamma_0 = \Gamma\) is consistent.Let \(n\) be the smallest number such that \(\Gamma_n\) is consistent,but \(\Gamma_{n+1} = \Gamma_n,(\exists x\theta_n \rightarrow\theta_{n}(x|c_i))\) is inconsistent. By \((\neg\)I), we have that
\[\tag{1}\Gamma_n \vdash \neg(\exists x\theta_n \rightarrow \theta_n(x|c_i)).\]Byex falso quodlibet (Theorem 10), \(\Gamma_n, \neg \existsx\theta_n, \exists x\theta_n \vdash \theta_n (x|c_i)\). So by\((\rightarrow\)I), \(\Gamma_n, \neg \exists x\theta_n \vdash(\existsx\theta_n \rightarrow \theta_n (x|c_i))\). From this and (1), we have\(\Gamma_n \vdash \neg \neg \exists x\theta_n\), by \((\neg\)I), andby (DNE) we have
\[\tag{2}\Gamma_n \vdash \exists x\theta_n .\]By (As), \(\Gamma_n, \theta_n (x|c_i), \exists x\theta_n \vdash\theta_n (x|c_i)\). So by \((\rightarrow\)I), \(\Gamma_n, \theta_n(x|c_i)\vdash(\exists x\theta_{n} \rightarrow \theta_{n}(x|c_i))\).From this and (1), we have \(\Gamma_n \vdash \neg \theta_n (x|c_i)\),by \((\neg\)I). Let \(t\) be a term that does not occur in\(\theta_n\) or in any member of \(\Gamma_n\). By uniform substitutionof \(t\) for \(c_i\), we can turn the derivation of \(\Gamma_n \vdash\neg \theta_n (x|c_i)\) into \(\Gamma_n \vdash \neg \theta_n (x|t)\).By \((\forall\)I), we have
\[\tag{3}\Gamma_n \vdash \forall v\neg \theta_n (x|v).\]By (As) we have \(\{\forall v\neg \theta_n (x|v),\theta_n\}\vdash\theta_n\) and by \((\forall\)E) we have \(\{\forall v\neg \theta_n(x|v), \theta_n\}\vdash \neg \theta_n\). So \(\{\forall v\neg \theta_n(x|v), \theta_n\}\) is inconsistent. Let \(\phi\) be any sentence ofthe language. Byex falso quodlibet (Theorem 10), we havethat \(\{\forall v\neg \theta_n (x|v),\theta_n\}\vdash \phi\) and\(\{\forall v\neg \theta_n (x|v), \theta_n\}\vdash \neg \phi\). Sowith (2), we have that \(\Gamma_n, \forall v\neg \theta_n (x|v)\vdash\phi\) and \(\Gamma_n, \forall v\neg \theta_n (x|v)\vdash \neg \phi\),by \((\exists\)E). By Cut (Theorem 11), \(\Gamma_n \vdash \phi\) and\(\Gamma_n \vdash \neg \phi\). So \(\Gamma_n\) is inconsistent,contradicting the assumption. So \(\Gamma'\) is consistent.
Applying the Lindenbaum Lemma (Theorem 13), let \(\Gamma''\) be amaximally consistent set of sentences (of the expanded language) thatcontains \(\Gamma'\). So, of course, \(\Gamma''\) contains \(\Gamma\).We can now define an interpretation \(M\) such that \(M\) satisfiesevery member of \(\Gamma''\).
If we did not have a sign for identity in the language, we would letthe domain of \(M\) be the collection of new constants \(\{c_0, c_1,\ldots \}\). But as it is, there may be a sentence in the form\(c_{i}=c_{j}\), with \(i\ne j\), in \(\Gamma''\). If so, we cannothave both \(c_i\) and \(c_j\) in the domain of the interpretation (asthey are distinct constants). So we define the domain \(d\) of \(M\)to be the set \(\{c_i\) | there is no \(j\lt i\) such that\(c_{i}=c_{j}\) is in \(\Gamma''\}\). In other words, a constant\(c_i\) is in the domain of \(M\) if \(\Gamma''\) does not declare itto be identical to an earlier constant in the list. Notice that foreach new constant \(c_i\), there is exactly one \(j\le i\) such that\(c_j\) is in \(d\) and the sentence \(c_{i}=c_{j}\) is in\(\Gamma''\).
We now define the interpretation function \(I\). Let \(a\) be anyconstant in the expanded language. By \((=\)I) and \((\exists\)I),\(\Gamma''\vdash \exists x x=a\), and so \(\exists x x=a \in\Gamma''\). By the construction of \(\Gamma'\), there is a sentence inthe form \((\exists x x=a \rightarrow c_i =a)\) in \(\Gamma''\). Wehave that \(c_i =a\) is in \(\Gamma''\). As above, there is exactlyone \(c_j\) in \(d\) such that \(c_{i}=c_{j}\) is in \(\Gamma''\). Let\(I(a)=c_j\). Notice that if \(c_i\) is a constant in the domain\(d\), then \(I\)(c\(_i)=c_i\). That is each \(c_i\) in \(d\) denotesitself.
Let \(P\) be a zero-place predicate letter in \(K\). Then \(I(P)\) istruth if \(P\) is in \(\Gamma''\) and \(I(P)\) is falsehood otherwise.Let \(Q\) be a one-place predicate letter in \(K\). Then \(I(Q)\) isthe set of constants \(\{\)c\(_i | c_i\) is in \(d\) and the sentence\(Qc\) is in \(\Gamma''\}\). Let \(R\) be a binary predicate letter in\(K\). Then \(I(R)\) is the set of pairs of constants \(\{\langlec_i,c_j\rangle | c_i\) is in \(d, c_j\) is in \(d\), and the sentence\(Rc_{i}c_{j}\) is in \(\Gamma''\}\). Three-place predicates, etc. areinterpreted similarly. In effect, \(I\) interprets the non-logicalterminology as they are in \(\Gamma''\).
The final item in this proof is a lemma that for every sentence\(\theta\) in the expanded language, \(M\vDash \theta\) if and only if\(\theta\) is in \(\Gamma''\). This proceeds by induction on thecomplexity of \(\theta\). The case where \(\theta\) is atomic followsfrom the definitions of \(M\) (i.e., the domain \(d\) and theinterpretation function \(I\)). The other cases follow from thevarious clauses in the definition of satisfaction.
Since \(\Gamma \subseteq \Gamma''\), we have that \(M\) satisfiesevery member of \(\Gamma\). By Theorem 15, the restriction of \(M\) tothe original language \(\LKe\) and \(s\) also satisfies every memberof \(\Gamma\). Thus \(\Gamma\) is satisfiable.
A converse to Soundness (Theorem 18) is a straightforwardcorollary:
Theorem 21. For any sentence \(\theta\) and set\(\Gamma\) of sentences, if \(\Gamma \vDash \theta\), then \(\Gamma\vdash_D \theta\).
Proof: Suppose that \(\Gamma \vDash \theta\). Thenthere is no interpretation \(M\) such thatM satisfies everymember of \(\Gamma\) but does not satisfy \(\theta\). So the set\(\Gamma,\neg \theta\) is not satisfiable. By Completeness (Theorem20), \(\Gamma,\neg \theta\) is inconsistent. So there is a sentence\(\phi\) such that \(\Gamma,\neg \theta \vdash \phi\) and\(\Gamma,\neg \theta \vdash \neg \phi\). By \((\neg\)I), \(\Gamma\vdash \neg \neg \theta\), and by (DNE) \(\Gamma \vdash \theta\).
Our next item is a corollary of Theorem 9, Soundness (Theorem 18), andCompleteness:
Corollary 22. Compactness. A set \(\Gamma\) ofsentences is satisfiable if and only if every finite subset of\(\Gamma\) is satisfiable.
Proof: If \(M\) satisfies every member of \(\Gamma\),then \(M\) satisfies every member of each finite subset of \(\Gamma\).For the converse, suppose that \(\Gamma\) is not satisfiable. Then weshow that some finite subset of \(\Gamma\) is not satisfiable. ByCompleteness (Theorem 20), \(\Gamma\) is inconsistent. By Theorem 9(and Weakening), there is a finite subset \(\Gamma'\subseteq \Gamma\)such that \(\Gamma'\) is inconsistent. By Corollary \(19, \Gamma'\) isnot satisfiable.
Soundness and completeness together entail that an argument isdeducible if and only if it is valid, and a set of sentences isconsistent if and only if it is satisfiable. So we can go back andforth between model-theoretic and proof-theoretic notions,transferring properties of one to the other. Compactness holds in themodel theory because all derivations use only a finite number ofpremises.
Recall that in the proof of Completeness (Theorem 20), we made thesimplifying assumption that the set \(K\) of non-logical constants iseither finite or denumerably infinite. The interpretation we producedwas itself either finite or denumerably infinite. Thus, we have thefollowing:
Corollary 23. Löwenheim-Skolem Theorem. Let\(\Gamma\) be a satisfiable set of sentences of the language \(\LKe\).If \(\Gamma\) is either finite or denumerably infinite, then\(\Gamma\) has a model whose domain is either finite or denumerablyinfinite.
In general, let \(\Gamma\) be a satisfiable set of sentences of\(\LKe\), and let \(\kappa\) be the larger of the size of \(\Gamma\)and denumerably infinite. Then \(\Gamma\) has a model whose domain isat most size \(\kappa\).
There is a stronger version of Corollary 23. Let \(M_1 =\langled_1,I_1\rangle\) and \(M_2 =\langle d_2,I_2\rangle\) beinterpretations of the language \(\LKe\). Define \(M_1\) to be asubmodel of \(M_2\) if \(d_1 \subseteq d_2, I_1 (c) = I_2(c)\) for each constant \(c\), and \(I_1\) is the restriction of\(I_2\) to \(d_1\). For example, if \(R\) is a binary relation letterin \(K\), then for all \(a,b\) in \(d_1\), the pair \(\langlea,b\rangle\) is in \(I_1 (R)\) if and only if \(\langle a,b\rangle\)is in \(I_2 (R)\). If we had included function letters among thenon-logical terminology, we would also require that \(d_1\) be closedunder their interpretations in \(M_2\). Notice that if \(M_1\) is asubmodel of \(M_2\), then any variable-assignment on \(M_1\) is also avariable-assignment on \(M_2\).
Say that two interpretations \(M_1 =\langle d_1,I_1\rangle, M_2=\langle d_2,I_2\rangle\) areequivalent if one of them is asubmodel of the other, and for any formula of the language and anyvariable-assignment \(s\) on the submodel, \(M_1,s\vDash \theta\) ifand only if \(M_2,s\vDash \theta\). Notice that if two interpretationsare equivalent, then they satisfy the same sentences.
Theorem 25. Downward Löwenheim-Skolem Theorem.Let \(M = \langle d,I\rangle\) be an interpretation of the language\(\LKe\). Let \(d_1\) be any subset of \(d\), and let \(\kappa\) bethe maximum of the size of \(K\), the size of \(d_1\), and denumerablyinfinite. Then there is a submodel \(M' = \langle d',I'\rangle\) of\(M\) such that (1) \(d'\) is not larger than \(\kappa\), and (2)\(M\) and \(M'\) are equivalent. In particular, if the set \(K\) ofnon-logical terminology is either finite or denumerably infinite, thenany interpretation has an equivalent submodel whose domain is eitherfinite or denumerably infinite.
Proof: Like completeness, this proof is complex, andwe rest content with a sketch. The downward Löwenheim-Skolemtheorem invokes the axiom of choice, and indeed, is equivalent to theaxiom of choice (see the entry onthe axiom of choice). So let \(C\) be a choice function on the powerset of \(d\), so thatfor each non-empty subset \(e\subseteq d, C(e)\) is a member of \(e\).We stipulate that if \(e\) is the empty set, then \(C(e)\) is\(C(d)\).
Let \(s\) be a variable-assignment on \(M\), let \(\theta\) be aformula of \(\LKe\), and let \(v\) be a variable. Define the\(v\)-witness of \(\theta\)over s, written \(w_v(\theta,s)\), as follows: Let \(q\) be the set of all elements \(c\ind\) such that there is a variable-assignment \(s'\) on \(M\) thatagrees with \(s\) on every variable except possibly \(v\), such that\(M,s'\vDash \theta\), and \(s'(v)=c\). Then \(w_v (\theta,s) =C(q)\). Notice that if \(M,s\vDash \exists v\theta\), then \(q\) isthe set of elements of the domain that can go for \(v\) in \(\theta\).Indeed, \(M,s\vDash \exists v\theta\) if and only if \(q\) isnon-empty. So if \(M,s\vDash \exists v\theta\), then \(w_v(\theta,s)\) (i.e., \(C(q))\) is a chosen element of the domain thatcan go for \(v\) in \(\theta\). In a sense, it is a“witness” that verifies \(M,s\vDash \exists v\theta\).
If \(e\) is a non-empty subset of the domain \(d\), then define avariable-assignment \(s\) to be an \(e\)-assignment if forall variables \(u, s(u)\) is in \(e\). That is, \(s\) is an\(e\)-assignment if \(s\) assigns an element of \(e\) to eachvariable. Define \(sk(e)\), theSkolem-hull of \(e\), to bethe set:
\[\begin{align*}e \cup \{w_v (\theta,s)|& \theta \text{ is a formula in } \LKe, \\ & v \text{ is a variable, and } \\ & s \text{ is an } e\text{-assignment} \}.\end{align*}\]That is, the Skolem-Hull of \(e\) is the set \(e\) together with every\(v\)-witness of every formula over every \(e\)-assignment. Roughly,the idea is to start with \(e\) and then throw in enough elements tomake each existentially quantified formula true. But we cannot restcontent with the Skolem-hull, however. Once we throw the“witnesses” into the domain, we need to deal with\(sk(e)\) assignments. In effect, we need a set which is its ownSkolem-hull, and also contains the given subset \(d_1\).
We define a sequence of non-empty sets \(e_0, e_1,\ldots\) as follows:if the given subset \(d_1\) of \(d\) is empty and there are noconstants in \(K\), then let \(e_0\) be \(C(d)\), the choice functionapplied to the entire domain; otherwise let \(e_0\) be the union of\(d_1\) and the denotations under \(I\) of the constants in \(K\). Foreach natural number \(n, e_{n+1}\) is \(sk(e_n)\). Finally, let \(d'\)be the union of the sets \(e_n\), and let \(I'\) be the restriction of\(I\) to \(d'\). Our interpretation is \(M' = \langled',I'\rangle\).
Clearly, \(d_1\) is a subset of \(d'\), and so \(M'\) is a submodel of\(M\). Let \(\kappa\) be the maximum of the size of \(K\), the size of\(d_1\), and denumerably infinite. A calculation reveals that the sizeof \(d'\) is at most \(\kappa\), based on the fact that there are atmost \(\kappa\)-many formulas, and thus, at most \(\kappa\)-manywitnesses at each stage. Notice, incidentally, that this calculationrelies on the fact that a denumerable union of sets of size at most\(\kappa\) is itself at most \(\kappa\). This also relies on the axiomof choice.
The final item is to show that \(M'\) is equivalent to \(M\): Forevery formula \(\theta\) and every variable-assignment \(s\) on\(M'\),
\[M,s\vDash \theta \text{ if and only if } M',s\vDash \theta.\]The proof proceeds by induction on the complexity of \(\theta\).Unfortunately, space constraints require that we leave this step as anexercise.
Another corollary to Compactness (Corollary 22) is the opposite of theLöwenheim-Skolem theorem:
Theorem 26. Upward Löwenheim-Skolem Theorem. Let\(\Gamma\) be any set of sentences of \(\LKe,\) such that for eachnatural number \(n\), there is an interpretation \(M_n = \langled_n,I_n\rangle\), such that \(d_n\) has at least \(n\) elements, and\(M_n\) satisfies every member of \(\Gamma\). In other words,\(\Gamma\) is satisfiable and there is no finite upper bound to thesize of the interpretations that satisfy every member of \(\Gamma\).Then for any infinite cardinal \(\kappa\), there is an interpretation\(M=\langle d,I\rangle\), such that the size of \(d\) isatleast \(\kappa\) and \(M\) satisfies every member of\(\Gamma\).
Proof: Add a collection of new constants\(\{c_{\alpha} | \alpha \lt \kappa \}\), of size \(\kappa\), to thelanguage, so that if \(c\) is a constant in \(K\), then \(c_{\alpha}\)is different from \(c\), and if \(\alpha \lt \beta \lt \kappa\), then\(c_{\alpha}\) is a different constant than \(c_{\beta}\). Considerthe set of sentences \(\Gamma'\) consisting of \(\Gamma\) togetherwith the set \(\{\neg c_{\alpha}=c_{\beta} | \alpha \ne \beta \}\).That is, \(\Gamma'\) consists of \(\Gamma\) together with statementsto the effect that any two different new constants denote differentobjects. Let \(\Gamma''\) be any finite subset of \(\Gamma'\), and let\(m\) be the number of new constants that occur in \(\Gamma''\). Thenexpand the interpretation \(M_m\) to an interpretation \(M_m'\) of thenew language, by interpreting each of the new constants in\(\Gamma''\) as a different member of the domain \(d_m\). Byhypothesis, there are enough members of \(d_m\) to do this. One caninterpret the other new constants at will. So \(M_m\) is a restrictionof \(M_m'\). By hypothesis (and Theorem 15), \(M'_m\) satisfies everymember of \(\Gamma\). Also \(M'_m\) satisfies the members of \(\{\negc_{\alpha}=c_{\beta} | \alpha \ne \beta \}\) that are in \(\Gamma''\).So \(M'_m\) satisfies every member of \(\Gamma''\). By compactness,there is an interpretation \(M = \langle d,I\rangle\) such that \(M\)satisfies every member of \(\Gamma'\). Since \(\Gamma'\) containsevery member of \(\{\neg c_{\alpha}=c_{\beta} | \alpha \ne \beta \}\),the domain \(d\) of \(M\) must be of size at least \(\kappa\), sinceeach of the new constants must have a different denotation. By Theorem15, the restriction of \(M\) to the original language \(\LKe\)satisfies every member of \(\Gamma\).
Combined, the proofs of the downward and upward Löwenheim-Skolemtheorems show that for any satisfiable set \(\Gamma\) of sentences, ifthere is no finite bound on the models of \(\Gamma\), then for anyinfinite cardinal \(\kappa\), there is a model of \(\Gamma\) whosedomain has sizeexactly \(\kappa\). Moreover, if \(M\) is anyinterpretation whose domain is infinite, then for any infinitecardinal \(\kappa\), there is an interpretation \(M'\) whose domainhas size exactly \(\kappa\) such that \(M\) and \(M'\) areequivalent.
These results indicate a weakness in the expressive resources offirst-order languages like \(\LKe\). No satisfiable set of sentencescan guarantee that its models are all denumerably infinite, nor canany satisfiable set of sentences guarantee that its models areuncountable. So in a sense, first-order languages cannot express thenotion of “denumerably infinite”, at least not in themodel theory. (See the entry onsecond-order and higher-order logic.)
Let \(A\) be any set of sentences in a first-order language \(\LKe\),where \(K\) includes terminology for arithmetic, and assume that everymember of \(A\) is true of the natural numbers. We can even let \(A\)be the set of all sentences in \(\LKe\) that are true of the naturalnumbers. Then \(A\) has uncountable models, indeed models of anyinfinite cardinality. Such interpretations are among those that aresometimes calledunintended, ornon-standard modelsof arithmetic. Let \(B\) be any set of first-order sentences that aretrue of the real numbers, and let \(C\) be any first-orderaxiomatization of set theory. Then if \(B\) and \(C\) are satisfiable(in infinite interpretations), then each of them has denumerablyinfinite models. That is, any first-order, satisfiable set theory ortheory of the real numbers, has (unintended) models the size of thenatural numbers. This is despite the fact that a sentence (seemingly)stating that the universe is uncountable is provable in mostset-theories. This situation, known as theSkolem paradox,has generated much discussion, but we must refer the reader elsewherefor a sample of it (see the entry onSkolem’s paradox and Shapiro 1996).
Surely, logic has something to do with correct reasoning, or at leastcorrect deductive reasoning. The details of the connection are subtle,and controversial – see Harman [1984] for an influential study.It is common to say that someone has reasoned poorly if they have notreasoned logically, or that a given (deductive) argument is bad, andmust be retracted, if it is shown to be invalid.
Some philosophers and logicians have maintained that there is a singlelogical system that is uniquely correct, in its role of characterizingvalidity. Among those, some, perhaps most, favor classical,first-order logic as uniquely correct, as the One True Logic. See, forexample, Quine [1986], Resnik [1996], Rumfitt [2015], Williamson[2017], and a host of others.
That classical, first-order logic should be given this role is perhapsnot surprising. It has rules which are more or less intuitive, and issimple for how strong it is. As we have seen in section 5, classical,first-order logic has interesting and important meta-theoreticproperties, such as soundness and completeness, that have lead to manyimportant mathematical and logical studies.
However, as noted, the main meta-theoretic properties of classical,first-order logic lead to expressivelimitations of theformal languages and model-theoretic semantics. Key notions, likefinitude, countability, minimal closure, natural number, and the likecannot be expressed.
Barwise [1985, 5] once remarked:
As logicians, we do our subject a disservice by convincing others thatlogic is first-order and then convincing them that almost none of theconcepts of modern mathematics can really be captured in first-orderlogic.
And Wang [1974, 154]:
When we are interested in set theory or classical analysis, theLöwenheim-Skolem theorem is usually taken as a sort of defect...of the first-order logic... [W]hat is established [by these theorems]is not that first-order logic is the only possible logic but ratherthat it is the only possible logic when we in a sense deny reality tothe concept of [the] uncountable...
Other criticisms of classical, first-order logic have also beenlodged. There are issues with its ability to deal with certainparadoxes (see, for example, the entry onRussel’s paradox), its apparent overgeneration of beliefs (see the entry on (the normative status of logic), and some argue that it has some arguments that do not match with theway we normally think we think (see for example, the entry onrelevance logic).
There are two main options available to those who are critical ofclassical, first-order logic, as the One True Logic. One is to proposesome other logic as the One True Logic. Priest [2006a] describes themethodology one might use to settle in the One True Logic.
The other main option is to simply deny that there is a single logicthat qualifies as the One True Logic. One instance of this is a kindoflogical nihilism, a thesis that there is no correct logic.Another is alogical pluralism, the thesis that a variety ofdifferent logical all qualify as correct, or best, or even the truelogic, at least in various contexts.
Of course, this is not the place to pursue this matter in detail. SeeBeall and Restall [2006] and Shapiro [2014] for examples of pluralism,and the entry onlogical pluralism for an overview of the terrain for both logical pluralism and logicalnihilism.
We close with brief sketches of some of the main alternatives toclassical, first-order logic, providing references to other work andentries to this Encyclopedia. See also the second half of Shapiro andKouri Kissel [2022].
In recent years, some work has been done to "approximate" classicallogic. The idea is to get as close to classical logic as possible, inorder to preserve some of the benefits, while at the same timeremoving some limitations of classical logic, like being closer tointuitive inference or applying to things like vagueness andparadoxes.
For example, Barrio, Pailos and Szmuc [2020] show that we canapproximate classical logic in something called the ST-hierarchy (STfor strict-tolerant, from Cobreros, Egre, Ripley and van Rooij[2012a,b]). This allows them to avoid certain classical problems ateach level of the hierarchy, like some of the paradoxes, while at thesame time maintaining many of the benefits of the strength ofclassical logic when considering the full hierarchy.
Dave Ripley [2013] provides a multi-sequent calculus version of“classical logic” that he argues solves some of theparadoxes. Notably, he claims it solves at least the Sorites and LiarParadoxes (see the entries on thesorites paradox andliar Paradox). The system conservatively extends classical logic. Ripley claims thatthis is what makes it classical. However, the system is nottransitive, and does not have a Cut rule.
There are, of course, some questions about whether these new logicsarereally classical, but it is informative worknonetheless.
One way to extend classical, first-order logic is to add additionaloperators to the underlying formal language.Modal logic addsoperators which designate necessity and possibility. So, we can saythat a proposition is possibly true, or necessarily true, rather thanjust true.
W. V. O Quine [1953] once argued that it is not coherent forquantifiers to bind variables inside modal operators, but opinion onthis matter has since changed considerably (see, for example, Barcan[1990]). There is now a thriving industry of developing modal logicsto capture various kinds of modality and temporal operators. See theentry onmodal logic.
All of the formal languages sketched above have only one sort ofvariable. These are sometimes calledfirst-order variables.Each interpretation of the language has a domain, which is the rangeof these first-order variables. It is what the language is about,according to the given interpretation.Second-order variablesrange over properties, sets, classes, relations, or functions of theitems in that domain.Third-order variables range overproperties, classes, relations of whatever is in the range of thesecond-order variables. And it goes on from there.
A formal language is calledsecond-order if it hassecond-order variables and first-order variables, and no others;Third-order if it has third-order, second-order, andfirst-order variables and no others, etc. A formal language ishigher-order if it is at least second-order.
As noted, it is not an exaggeration to say that classical, first-orderlogic is the paradigm of contemporary logical theory. Most textbooksdo not mention higher-order languages at all, and most of the restgive it scant treatment.
A number of different deductive systems and model-theoretic semanticshave been proposed for second- and higher-order languages. For thesemantics, the main additional feature of the model-theory is tospecify a range of the higher-order variables.
InHenkin semantics, each interpretation specifies a specificrange of the higher-order variables. For monadic second-ordervariables, each interpretation specifies a non-empty subset of thepowerset of the domain, for two-place second-order variables, anon-empty set of ordered pairs of members of the domain, etc. Thesystem has all of the above limitative meta-theoretic results. Thereis a deductive system that is sound and complete for Henkin semantics;the logic is compact; and the downward and upwardLöwenheim-Skolem theorems all hold.
In so-calledstandard semantics, sometimes calledfullsemantics, monadic second-order variables range over the entirepowerset of the domain; two-place second-order variables range overthe entire class of ordered pairs of members of the domain, etc. Itcan be shown that second-order languages, with standard semantics, cancharacterize many mathematical notions and structures, up toisomorphism. Examples include the notions of finitude, countability,well-foundedness, minimal closure, and structures like the naturalnumbers, the real numbers, and the complex numbers. As a result, noneof the limitative theorems of classical, first-order logic hold: thereis no effective deductive system is both sound and complete, the logicis not compact, and both Löwenheim-Skolem theorems fail. Some,such as Quine [1986], argue that second-order logic, with standardsemantics is not really logic, but is a form of mathematics, settheory in particular. For more on this, see Shapiro [1991] and theentry onhigher-order logic, along with the many references cited there.
One might also consider generalized quantifiers as an expansion ofclassical first-order logic (see the entry ongeneralized quantifiers). These quantifiers allow from an expansion between the classical“all” and “some” , and can accommodatequantifiers like “most” , “less than half” ,“usually” , etc. They are useful from both a logical andlinguistic perspective. For example, Kennedy andVäänänen [2021] use generalized quantifiers to arguethat “ uncountable” is a logical notion.
Some philosophers and logicians argue that classical, first-orderlogic is too strong: it declares that some argument-forms are validwhich are not. Here we sketch two kinds of proposals.
Advocates ofintuitionistic logic reject the validity of the(so-called) Law of Excluded Middle:
\[\Phi \vee \neg \Phi,\]and other inferences related to this, such as Double NegationElimination (DNE):
\[{\rm If}\ \Gamma \vdash \neg\neg\Phi \ {\rm then}\ \Gamma \vdash \Phi\]Roughly speaking, there are two main motivations for theserestrictions. The traditional intuitionists L. E. J. Brouwer (e.g.,[1964a], [1964b]) and Arend Heyting (e.g. [1956]) held that theessence of mathematics is idealized mental construction. Consider, forexample, the proposition that for every natural number \(n\), there isa prime number \(m \gt n\) such that \(m \lt n!+2\). For Brouwer, thisproposition invokes aprocedure that, given any naturalnumber \(n\), produces a prime number \(m\) that is greater than \(n\)but less than \(n!+2\). The proposition expresses the existence ofsuch a procedure. Given this orientation, we have no reason to holdthat for any mathematical proposition \(\Phi\), we can establisheither the procedure associated with \(\Phi\) or the procedureassociated with \(\neg \Phi\).
Michael Dummett (e.g., [1978]) provides general arguments concerninghow language functions, as a vehicle of communication, to argue thatintuitionistic logic is uniquely correct, the One True Logic, not justfor mathematics.
For an overview of intuitionistic logic, and its philosophicalmotivation, see the entry onintuitionistic logic.
This time the target inference, to be declared invalid, is the one weabove callex falso quodlibet, abbreviated (EFQ):\[{\rm If} \ \Gamma_1 \vdash \Theta \ {\rm and} \ \Gamma_2 \vdash \neg\Theta \ {\rm then} \ \Gamma_1, \Gamma_2 \vdash \Psi\] We can focus attention one kind of instance of this:\[\Phi, \neg\Phi \vdash \Psi,\] sometimes colorfully called “explosion”. Itsays that anything at all follows from a contradiction.
Logics that regard (EFQ) as invalid are calledparaconsistent. Broadly speaking, there are two camps oflogicians advocating for paraconsistent systems, either as candidatesfor the One True Logic or as instances of pluralism. One camp consistsof logicians who insist that in a valid argument, the premises must berelevant to the conclusion. Typically, relevance logiciansalso demur from certain classical logical truths calledparadoxesof material implication, such as \((\Phi \rightarrow (\Psi\rightarrow \Phi))\) and \((\Phi \rightarrow (\Psi \rightarrow\Psi))\).
For more, see the entry onrelevance logic, or Kerr [2019]. Classic works include Anderson and Belnap [1975],Anderson Belnap and Dunn [1992], and Read [1988]. Neil Tennant’s[2017] core logic is both relevant and intuitionistic.
The other main camp of logicians who prefer a paraconsistent logic (orparaconsistent logics) are advocates ofdialetheism, the viewthat some contradictions, some sentences in the form \[(\Phi \wedge \neg \Phi),\] aretrue. One supposed example is when \(\Phi\) is a statement of asemantic paradoxes, such as the Liar. Consider, for example, asentence \(\Phi\) that says that \(\Phi\) is not true.
In a system in which (EFQ) holds, any true contradiction would entailevery sentence of the formal language, thus rendering the language andtheory trivial. So, clearly, any logic for dialetheism would have tobe paraconsistent. See the entry ondialetheism. The classic work here is Priest [2006a].
Of course, the small sample presented here does not include everylogical system proposed as a rival to classical, first-order logic,again either as a candidate for the One True Logic, or as a furtherinstance of logical pluralism. See, for example, the entries onsubstructural logics,fuzzy logic, and many others.
There are many fine textbooks on mathematical logic. A samplefollows.
How to cite this entry. Preview the PDF version of this entry at theFriends of the SEP Society. Look up topics and thinkers related to this entry at the Internet Philosophy Ontology Project (InPhO). Enhanced bibliography for this entryatPhilPapers, with links to its database.
[Please contact the author with suggestions.]
logic: free |logic: infinitary |logic: intuitionistic |logic: linear |logic: modal |logic: paraconsistent |logic: relevance |logic: second-order and higher-order |logic: substructural |logic: temporal |logical consequence |logical form |logical truth |model theory |model theory: first-order |paradox: Skolem’s |proof theory: development of
View this site from another server:
The Stanford Encyclopedia of Philosophy iscopyright © 2025 byThe Metaphysics Research Lab, Department of Philosophy, Stanford University
Library of Congress Catalog Data: ISSN 1095-5054