Hilbert logics (2): Quine, quantifiers, and modalities

Thesis: all our most familiar logics, both sentential and quantificational, are Hilbert logics.

In Henkin’s paper (cited in the preceding post) his main target was not sentential but predicate logic. Given that the proof of the Deduction Theorem hinges on Modus Ponens being the sole rule of inference, there is an obvious question. Isn’t

Universal Generalization. from ⊢Fa to infer ⊢(x)Fx

an indispensable rule as well?

Moreover, Henkin indicates that his topic here is not just classical and intuitionistic quantification theory, but applies to quantified modal logic. At that time the concept of a normal modal logic had not yet been introduced. The main modal logics were Lewis and Langford’s S1 through S5, and in retrospect we can say that of these only S4 and S5 were normal. But as typically formulated now, the normal modal logics have a rule that is similar to Universal Generalization, namely

Necessitation. from ⊢A to infer ⊢□A

It is important, and easy to see, that these rules are quite different from Modus Ponens, which says

from A and A ⊃ B to infer B

or, using the deducibility symbol,

A, A ⊃ B ⊢ B

Universal Generalization and Necessitation do not provide inferences from premises to conclusions. Instead, what they do, in effect, is to add lots to the stock of axioms, that is, to the lines in a deduction that do not need to be inferred from previous lines.

It is in fact possible in both cases to have a formulation that is a Hilbert logic, and here I would like to explore one way in which that can be shown.

Quantifier logic a la Quine

For quantificational logic this was made this beautifully clear by Quine in his Mathematical Logic (revised edition 1951). His system of quantifier logic with identity, called ML, does not have the rule of Universal Generalization, and Modus Ponens is its only rule of inference. But he introduced a more nuanced notion of axiom schemata.

First he defines the notion of closure: if A is a well-formed formula which has in it free variables x, y, z, … then the closure of A is the statement which results when universal quantifiers are prefixed to A to bind those variables. So the closure of A looks like this: (x)(y)(z) …A.

Each axiom schema then defines a large class of statements, which are the closures of instances of the displayed form. For example for the class he calls C he displays the form

A ⊃ (x)A, where A is a formula in which x is not a free variable,

and C is the class of all closures of well-formed formulas that have this form. Examples would be (y)(Fy ⊃ (x)Fy), (z)(y)(Rzy ⊃ (x)Rzy), and so forth.

It is clear then that no axiom, and hence no theorem, has free variables in it. Quine’s language did not have individual constants. So it is not even possible to formulate a non-trivial version of Universal Generalization for ML itself. But more to the point, if individual constants are added, we have a choice, and for each choice we arrive at a complete quantificational logic without adding the Universal Generalization rule. If we add the axiom schema

(Ey)(y = a)

the result is the familiar classical predicate logic. Intuitively, each individual constant is the name of a real thing. If we do not add anything then the result is free logic. In that logic, again intuitively, the only existence assumption built it (as it was for Quine) is that there is something rather than nothing. But the individual constants can be non-referring terms, like our “Pegasus”.

In both cases it is provable as a meta-theorem that if ⊢Fa then ⊢(x)Fx. The need for the rule of Universal Generalization has been neatly by-passed by Quine’s tactic.

Modal logic

So how does it stand with the modal logics? Ruth Barcan Marcus wrote the seminal papers on quantified modal logic in the years 1946-1953. She proved the usual Deduction Theorem for material implication for S4 and S5, as formulated by Lewis and Langford, and showed that it could not be proved for S1 or S2.

In our present context, where the family of normal modal logics has a standard formulation, the rule Necessitation plays a salient role. Can Quine’s tactic be implemented here, mutatis mutandis?

The weakest normal logic is system K, formulated as follows with one axiom schema and three rules:

R0. If p is a theorem of classical propositional logic then ⊢ p

R1. p, p ⊃ q ⊢ q

R2. If ⊢ p then ⊢ p

A1. ⊢ (p ⊃q) ⊃ (□p ⊃ □q)

Only R1, modus ponens, is an ordinary rule of inference. Rule R0 can be replaced by the axiom schemata of a Hilbert logic formulation of sentential logic. But R2 is Necessitation.

Let’s adapt Quine’s tactic as follows. If A is a formula then call the necessitations of A all formulas that consist of A preceded by zero or more symbols □. And introduce the more nuanced notion of the use of schemata as follows: each axiom schema introduces a large class of axioms, namely all the necessitations of instances of the displayed form. So for example the schemata that can replace R0 include the form (p ⊃ p), and this means that the axioms include all sentences of form (p ⊃ p), □(p ⊃ p), □□(p⊃p), …, and do forth.

Lets say that K* is the modal logic which is like K except for lacking rule R2, but with the axiom schemata understood in this new way. Then it is easy to prove that K* and K have the same theorems (see Appendix). And K* is a Hilbert logic, that is, a Hilbert logic formulation of K.

The same can be done, in the same way, for the other normal modal logics.So, it follows from this that Henkin was quite right, if we read him anachronistically as referring to the normal modal logics: the Deduction Theorem is provable for all of them.

Conclusion Both the normal modal logics and quantificational logics (classical, normal modal, and intuitionistic) are Hilbert logics.

Historical note. Ruth Barcan Marcus who pioneered quantified modal logic, proved the Deduction Theorem for S4 and S5, and showed that it could not be proved for S1 or S2 (Journal of Symbolic Logic 18 (1953): 234-236). For the completeness of ML with names added as free logic, van Fraassen (Zeitschr. Math. Logic und Grundl. der Math. 12 (1966): 219-234).

APPENDIX

K* and K have precisely the same theorems. The proof is by induction.

Suppose A is a theorem of K; then there is a proof S = S(1), …., S(n) in K which has A as its last line. Change this proof into the sequence T = T(1), …, T(k), … , T(n) by moving the axioms that appear in S to the beginning, though kept in the same order they had in S: these are the lines T(1) …, T(k). The lines T(k+1), … , T(n) are the remaining lines of S in the same order in which they appeared in S.

T is also a proof of A in K, for the axioms need no justification, and if another line of S followed from preceding lines by modus ponens, that same line appearing in T follows from lines that precede it by modus ponens as well.

Claim: There is a proof in K* of □A.

Proof.

Each of □T(i), for in = 1, …, k is an axiom, for T(i) is a necessitation of axiom schema A1, and therefore so is □T(i).

Hypothesis of induction: for k < m ≤ n, and for all j < m, there is a proof of □T(j) in K*.

In S the line S(c) which appears as T(j) in T came from preceding lines S(a) and S(b) by modus ponens, so S(b) was [S(a) ⊃ S(c)]. Thus by hypothesis there are also proofs of □S(a) and □[S(a) ⊃ S(c)]. From these, by an instance of axiom A1. and two applications of modus ponens, □S(c) follows. That conclusion is the same as □T(j).

Therefore, by induction, there is a proof in K* of □T(i), for 1 ≤ i ≤ n; hence of □T(n), which is □A.

Hilbert logics (1): Hilbert languages

Preface. Many things that can be proved about classical logic are provable on the basis of just a few important characteristics, and so generalize to large classes of logics.  I am interested at the moment especially in characteristics that play a role in discussions of the Curry Paradox.  While I will not discuss that paradox here, it is the reason for why I will focus on just features that pertain to the conditional.

Hilbert logics

While it is customary today to teach logic in a natural deduction formulation, that was a relatively new development.  Looking back to Principia Mathematica and other sources in the first half of the twentieth century what we find instead are axiom systems.  These can in general be read as consisting of a series of axiom schemata, with modus ponens as single rule of inference.  That sort of formulation was initiated by Hilbert, and of course it did not disappear.  For example, Kleene’s seminal Introduction to Metamathematics (1952) presents classical logic in this form (Ch. IV, sect. 19).  

Hilbert style deductive system is one in which a finite sequence of formulas is (a) a proof if and only if every line is either an axiom or follows from preceding lines by a rule of inference, and (b) a deduction (also, derivation) from certain premises if and only if every line is either a premise or an axiom or follows from preceding lines by a rule of inference. The last line of a proof is a theorem and the last line of a deduction is called the conclusion deducible from those premises.

Note well that this definition of the notion of a derivation from premises is correct for a deductive system in which modus ponens is the only rule of inference, it is not correct if certain other kinds of rules of inference (which only preserve theorem-hood) are present. In such cases we have to raise the question whether the logic has an equivalent formulation (in the sense that it has the same theorems) which does not have those other kinds of rules.

In subsequent literature the term Hilbert algebras  came to refer to algebraic models of a fragment of classical (as well as of intuitionistic) logic, that Henkin studied in 1950 as the system of Positive Implicational Logic.  Its only axiom schemata are:

H1. A ⊃ (B ⊃ A)

H2. [A ⊃ (B ⊃ C)] ⊃ [(A ⊃ B) ⊃ ( A ⊃C)]

and its only rule of inference is modus ponens:  

from A and A ⊃ B to infer B.

Henkin then noted that the Deduction Theorem (Church’s term, now in common use) can be proved for any system that includes Positive Implicational Logic.  The Deduction Theorem states, in effect, that the now familiar rule of Conditional Proof is an admissible rule for Positive Implicational Logic:

            if B is derivable from set of premises X and A, then A ⊃ B is derivable from X

if X, A ⊢B then X ⊢ A ⊃ B.

We are therefore staying close the historical concepts and terminology with the following definition:

Hilbert logic  is a logic which is a Hilbert style deductive system, with modus ponens as sole inference rule, of which Positive Implicational Logic is a fragment.

Classical logic and intuitionistic logic both have formulations which are Hilbert logics.  (Henkin offers other examples; I will postpone those to the next post).  

We often refer to those logics as understood in a formulation-independent way. It will not be a great abuse of nomenclatures to just say then that they are Hilbert logics, meaning that they have equivalent formulations that are Hilbert logics.

Does the Deduction Theorem carry over to logics of which Positive Implicational Logic is a fragment? Not always, it depends on what that larger logic is like. We can at least say this:

If L is a Hilbert Logic and is a fragment of logic L* then: (a) if L* is a Hilbert logic then the Deduction Theorem holds for L* as well; (b) if L* is not a Hilbert logic then the Deduction Theorem will hold for those deductions in L* in which modus ponens is the only inference rule used.

The addition of other axioms will not affect the proof of the Deduction Theorem, but the addition of other inference rules can certainly change the game.

Hilbert languages

Classical and intuitionistic logic have something else in common:  for each there is a strong completeness proof, showing that they capture the valid inferences  — the semantic entailment relation – in a language with the pertinent syntax.

So let us have a corresponding definition for that as well:

Hilbert language  is a language with formally defined syntax and semantics, in which the semantic entailment relation is finitary and there exists a Hilbert logic whose deducibility relation coincides with that semantic entailment relation.

So the generalization of classical logic and classical sentential language that I am interested in exploring is Hilbert logics and Hilbert languages.  

NOTE. Specifically, my previous posts about the Curry Paradox showed how it was possible to build a supervaluational language on top of a classical language in which Curry’s desideratum of expressive completeness was honored.  The salient question for me, at the moment, is whether my construction there and its accompanying argument go through for Hilbert languages and not just for the language of classical logic.

APPENDIX.

L. Henkin, “An Algebraic Characterization of Quantifiers”, Fundamenta Mathematicae, vol. 37, 1950, 63-74.   The name “Positive Implicational Logic” Henkin took from Hilbert and Bernays.  Alonzo Church’s proof  of the Deduction Theorem, which Henkin cites, is in Church’s An Introduction to Mathematical Logic (1944).  

NB: Hilbert’s definition of what counts as a deduction is such that the normal structural rules hold for the deducibility (consequence) relation.  So-called “sub-structural” logics are therefore not Hilbert logics.

Finitude. The restriction of proofs to finite sequences is not arbitrary. Imagine a logic whose only rule of inference (schema) is ” (A or A), therefore A”, which is surely something anyone would accept. But now suppose that its theorems are the last lines of proofs whether finite or infinite. Then think of a sequence whose lines are numbered with the negative integers, so the last line has number (-1). Let that last line be “God exists”, and if the line (-k) was B then the line (-(k+1)) was (B or B). It looks like this toward the end:

(-3) …………………………………………………….

(-2) God exists or God exists

(-1) God exists

Thus every line was either an axiom (that is vacuously true) or followed from a preceding line by the rule of inference.

So here we have it, a proof that God exists — since “God exists” is the last line of a proof. (I think I heard this example from Rich Thomason.) This is one paradox that is now universally blocked by the limitation of proofs to finite length, without ever even being mentioned.