Propositional calculus

This is a technical mathematical article about the area of mathematical logic variously known as "propositional calculus" or "propositional logic".

In logic and mathematics, a propositional calculus (or a sentential calculus) is a formal system in which formulae representing propositions can be formed by combining atomic propositions using logical connectives, and a system of formal proof rules allows certain formulæ to be established as "theorems".

In logic and philosophy, propositional calculus is often intended to symbolize rational deduction.

Contents

Terminology

In general terms, a calculus is a formal system that consists of a set of syntactic expressions (well-formed formulæ or wffs), a distinguished subset of these expressions (axioms), plus a set of formal rules that define a specific binary relation, intended to be interpreted as logical equivalence, on the space of expressions.

When the formal system is intended to be a logical system, the expressions are meant to be interpreted as statements, and the rules, known as inference rules, are typically intended to be truth-preserving. In this setting, the rules (which may include axioms) can then be used to derive ("infer") formulæ representing true statements from given formulæ representing true statements.

The set of axioms may be empty, a nonempty finite set, a countably infinite set, or be given by axiom schemata. A formal grammar recursively defines the expressions and well-formed formulæ (wffs) of the language. In addition a semantics may be given which defines truth and valuations (or interpretations).

The language of a propositional calculus consists of (1) a set of primitive symbols, variously referred to as atomic formulae, placeholders, proposition letters, or variables, and (2) a set of operator symbols, variously interpreted as logical operators or logical connectives. A well-formed formula (wff) is any atomic formula or any formula that can be built up from atomic formulæ by means of operator symbols according to the rules of the grammar.

Outline

The following outlines a standard propositional calculus. Many different formulations exist which are all more or less equivalent but differ in the details of (1) their language, that is, the particular collection of primitive symbols and operator symbols, (2) the set of axioms, or distinguished formulæ, and (3) the set of inference rules.

Generic description of a propositional calculus

A propositional calculus is a formal system \mathcal{L} = \mathcal{L} \left( \Alpha,\ \Omega,\ \Zeta,\ \Iota \right), whose formulæ are constructed in the following manner:

\Omega = \Omega_0 \cup \Omega_1 \cup \ldots \cup \Omega_j \cup \ldots \cup \Omega_m.
In this partition, \Omega_j\,\! is the set of operator symbols of arity j\,\!.
In the more familiar propositional calculi, \Omega\ ,\! is typically partitioned as follows:
\Omega_1 = \left\{ \lnot \right\},
\Omega_2 \subseteq \left\{ \land, \lor, \rightarrow, \leftrightarrow \right\}.
A frequently adopted option treats the constant logical values as operators of arity zero, thus:
\Omega_0 = \left\{ 0, 1 \right\}.
Some writers use the tilde (~) instead of (\neg); and some use the ampersand (&) or \cdot instead of \wedge. Notation varies even more for the set of logical values, with symbols like {false, true}, {F, T}, or \left\{ \bot,\ \top \right\} all being seen in various contexts instead of {0, 1}.
Depending on the precise formal grammar or the grammar formalism that is being used, syntactic auxiliaries like the left parenthesis, "(", and the right parentheses, ")", may be necessary to complete the construction of formulæ.
The language of \mathcal{L}, also known as its set of formulæ, well-formed formulas or wffs, is inductively or recursively defined by the following rules:
  1. Base. Any element of the alpha set \Alpha\ ,\! is a formula of \mathcal{L}.
  2. Step (a). If p\,\! is a formula, then \neg p is a formula.
  3. Step (b). If p\,\! and q\,\! are formulæ, then \left(p \land q \right), \left(p \lor q \right), \left(p \rightarrow q \right), and \left(p \leftrightarrow q \right) are formulæ.
  4. Close. Nothing else is a formula of \mathcal{L}.
Repeated applications of these rules permits the construction of complex formulæ. For example:
  1. By rule 1, p\,\! is a formula.
  2. By rule 2, \neg p is a formula.
  3. By rule 1, q\,\! is a formula.
  4. By rule 3, \left( \neg p \lor q \right) is a formula.

Example 1. Simple axiom system

Let \mathcal{L}_1 = \mathcal{L}\ (\Alpha,\ \Omega,\ \Zeta,\ \Iota), where \Alpha,\ \Omega,\ \Zeta,\ \Iota are defined as follows:

\Alpha = \{p, q, r, s, t, u \} \,.
\Omega_1 = \{ \lnot \}
\Omega_2 = \{ \rightarrow \}
  • (p \to (q \to p))
  • ((p \to (q \to r)) \to ((p \to q) \to (p \to r)))
  • ((\neg p \to \neg q) \to (q \to p))

Example 2. Natural deduction system

Let \mathcal{L}_2 = \mathcal{L}\ (\Alpha,\ \Omega,\ \Zeta,\ \Iota), where \Alpha,\ \Omega,\ \Zeta,\ \Iota are defined as follows:

\Alpha = \{p, q, r, s, t, u \}\,\!
\Omega_1 = \{ \lnot \}
\Omega_2 = \{ \land, \lor, \rightarrow, \leftrightarrow \}

In the following example of a propositional calculus, the transformation rules are intended to be interpreted as the inference rules of a so-called natural deduction system. The particular system presented here has no initial points, which means that its interpretation for logical applications derives its theorems from an empty axiom set.

Our propositional calculus has ten inference rules. These rules allow us to derive other true formulae given a set of formulae that are assumed to be true. The first nine simply state that we can infer certain wffs from other wffs. The last rule however uses hypothetical reasoning in the sense that in the premise of the rule we temporarily assume an (unproven) hypothesis to be part of the set of inferred formulae to see if we can infer a certain other formula. Since the first nine rules don't do this they are usually described as non-hypothetical rules, and the last one as a hypothetical rule.

Reductio ad absurdum (negation introduction) 
From  (p \to q), if accepting (q) leads to a proof that  \neg(p \to \ q), infer \neg q.
Double negative elimination 
From \neg \neg p, infer p\,\!.
Conjunction introduction 
From p\,\! and q\,\!, infer (p \land q).
From p\,\! and q\,\!, infer (q \land p).
Conjunction elimination 
From (p \land q), infer p\,\!
From (p \land q), infer q\,\!.
Disjunction introduction 
From p\,\!, infer (p \lor q)
From p\,\!, infer (q \lor p).
Disjunction elimination 
From (p \lor q), (p \to r), (q \to r), infer r\,\!.
Biconditional introduction 
From (p \to q), (q \to p), infer (p \leftrightarrow q).
Biconditional elimination 
From (p \leftrightarrow q), infer (p \to q)\,\!;
From (p \leftrightarrow q), infer (q \to p).
Modus ponens (conditional elimination) 
From p\,\!, (p \to q), infer q\,\!.
Conditional proof (conditional introduction) 
If accepting p\,\! allows a proof of q\,\!, infer (p \to q).

Basic and Derived Argument Forms
Name Sequent Description
Modus Ponens ((p \to q) \land p) \vdash q If p then q; p; therefore q
Modus Tollens ((p \to q) \land \neg q) \vdash \neg p If p then q; not q; therefore not p
Hypothetical Syllogism ((p \to q) \land (q \to r)) \vdash (p \to r) If p then q; if q then r; therefore, if p then r
Disjunctive Syllogism ((p \lor q) \land \neg p) \vdash q Either p or q, or both; not p; therefore, q
Constructive Dilemma ((p \to q) \land (r \to s) \land (p \lor r)) \vdash (q \lor s) If p then q; and if r then s; but p or r; therefore q or s
Destructive Dilemma ((p \to q) \land (r \to s) \land(\neg q \lor \neg s)) \vdash (\neg p \lor \neg r) If p then q; and if r then s; but not q or not s; therefore not p or not r
Bidirectional Dilemma ((p \to q) \land (r \to s) \land(p \lor \neg s)) \vdash (q \lor \neg r) If p then q; and if r then s; but p or not s; therefore q or not r
Simplification (p \land q) \vdash p p and q are true; therefore p is true
Conjunction p, q \vdash (p \land q) p and q are true separately; therefore they are true conjointly
Addition p \vdash (p \lor q) p is true; therefore the disjunction (p or q) is true
Composition ((p \to q) \land (p \to r)) \vdash (p \to (q \land r)) If p then q; and if p then r; therefore if p is true then q and r are true
De Morgan's Theorem (1) \neg (p \land q) \vdash (\neg p \lor \neg q) The negation of (p and q) is equiv. to (not p or not q)
De Morgan's Theorem (2) \neg (p \lor q) \vdash (\neg p \land \neg q) The negation of (p or q) is equiv. to (not p and not q)
Commutation (1) (p \lor q) \vdash (q \lor p) (p or q) is equiv. to (q or p)
Commutation (2) (p \land q) \vdash (q \land p) (p and q) is equiv. to (q and p)
Commutation (3) (p \leftrightarrow q) \vdash (q \leftrightarrow p) (p is equiv. to q) is equiv. to (q is equiv. to p)
Association (1) (p \lor (q \lor r)) \vdash ((p \lor q) \lor r) p or (q or r) is equiv. to (p or q) or r
Association (2) (p \land (q \land r)) \vdash ((p \land q) \land r) p and (q and r) is equiv. to (p and q) and r
Distribution (1) (p \land (q \lor r)) \vdash ((p \land q) \lor (p \land r)) p and (q or r) is equiv. to (p and q) or (p and r)
Distribution (2) (p \lor (q \land r)) \vdash ((p \lor q) \land (p \lor r)) p or (q and r) is equiv. to (p or q) and (p or r)
Double Negation p \vdash \neg \neg p p is equivalent to the negation of not p
Transposition (p \to q) \vdash (\neg q \to \neg p) If p then q is equiv. to if not q then not p
Material Implication (p \to q) \vdash (\neg p \lor q) If p then q is equiv. to not p or q
Material Equivalence (1) (p \leftrightarrow q) \vdash ((p \to q) \land (q \to p)) (p is equiv. to q) means (if p is true then q is true) and (if q is true then p is true)
Material Equivalence (2) (p \leftrightarrow q) \vdash ((p \land q) \lor (\neg p \land \neg q)) (p is equiv. to q) means either (p and q are true) or (both p and q are false)
Material Equivalence (3) (p \leftrightarrow q) \vdash ((p \lor \neg q) \land (\neg p \lor q)) (p is equiv. to q) means, both (p or not q is true) and (not p or q is true)
Exportation ((p \land q) \to r) \vdash (p \to (q \to r)) from (if p and q are true then r is true) we can prove (if q is true then r is true, if p is true)
Importation (p \to (q \to r)) \vdash ((p \land q) \to r)  
Tautology (1) p \vdash (p \lor p) p is true is equiv. to p is true or p is true
Tautology (2) p \vdash (p \land p) p is true is equiv. to p is true and p is true
Tertium non datur (Law of Excluded Middle) \vdash (p \lor \neg p) p or not p is true
Law of Non-Contradiction \vdash \neg (p \land \neg p) p and not p is false, is a true statement

Proofs in propositional calculus

One of the main uses of a propositional calculus, when interpreted for logical applications, is to determine relations of logical equivalence between propositional formulæ. These relationships are determined by means of the available transformation rules, sequences of which are called derivations or proofs.

In the discussion to follow, a proof is presented as a sequence of numbered lines, with each line consisting of a single formula followed by a reason or justification for introducing that formula. Each premise of the argument, that is, an assumption introduced as an hypothesis of the argument, is listed at the beginning of the sequence and is marked as a "premise" in lieu of other justification. The conclusion is listed on the last line. A proof is complete if every line follows from the previous ones by the correct application of a transformation rule. (For a contrasting approach, see proof-trees).

Example of a proof

Example of a Proof
Number Formula Reason
1 A \, premise
2 A \lor A From (1) by disjunction introduction
3 (A \lor A) \land A From (1) and (2) by conjunction introduction
4 A\, From (3) by conjunction elimination
5 A \vdash A Summary of (1) through (4)
6 \vdash A \to A From (5) by conditional proof

Interpret A \vdash A as "Assuming A \,, infer A \,". Read \vdash A \to A as "Assuming nothing, infer that A\, implies A\,", or "It is a tautology that A\, implies A\,", or "It is always true that A\, implies A\,".

Soundness and completeness of the rules

The crucial properties of this set of rules are that they are sound and complete. Informally this means that the rules are correct and that no other rules are required. These claims can be made more formal as follows.

We define a truth assignment as a function that maps propositional variables to true or false. Informally such a truth assignment can be understood as the description of a possible state of affairs (or possible world) where certain statements are true and others are not. The semantics of formulae can then be formalized by defining for which "state of affairs" they are considered to be true, which is what is done by the following definition.

We define when such a truth assignment A\, satisfies a certain wff with the following rules:

With this definition we can now formalize what it means for a formula \phi to be implied by a certain set S\, of formulae. Informally this is true if in all worlds that are possible given the set of formulae S\, the formula \phi also holds. This leads to the following formal definition: We say that a set S\, of wffs semantically entails (or implies) a certain wff \phi if all truth assignments that satisfy all the formulae in S\, also satisfy \phi.

Finally we define syntactical entailment such that \phi\,\! is syntactically entailed by S\, if and only if we can derive it with the inference rules that were presented above in a finite number of steps. This allows us to formulate exactly what it means for the set of inference rules to be sound and complete:

Soundness 
If the set of wffs S\, syntactically entails wff \phi\,\! then S\, semantically entails \phi
Completeness 
If the set of wffs S\, semantically entails wff \phi\,\! then S\, syntactically entails \phi

For the above set of rules this is indeed the case.

Sketch of a soundness proof

(For most logical systems, this is the comparatively "simple" direction of proof)

Notational conventions: Let "G" be a variable ranging over sets of sentences. Let "A", "B", and "C" range over sentences. For "G syntactically entails A" we write "G proves A". For "G semantically entails A" we write "G implies A".

We want to show: (A)(G)(if G proves A, then G implies A).

We note that "G proves A" has an inductive definition, and that gives us the immediate resources for demonstrating claims of the form "If G proves A, then ...". So our proof proceeds by induction.

(a) Assume for arbitrary G and A that if G proves A in n or fewer steps, then G implies A.
(b) For each possible application of a rule of inference at step n+1, leading to a new theorem B, show that G implies B.

Notice that Basis Step II can be omitted for natural deduction systems because they have no axioms. When used, Step II involves showing that each of the axioms is a (semantic) logical truth.

The Basis step(s) demonstrate(s) that the simplest provable sentences from G are also implied by G, for any G. (The is simple, since the semantic fact that a set implies any of its members, is also trivial.) The Inductive step will systematically cover all the further sentences that might be provable—by considering each case where we might reach a logical conclusion using an inference rule—and shows that if a new sentence is provable, it is also logically implied. (For example, we might have a rule telling us that from "A" we can derive "A or B". In III.(a) We assume that if A is provable it is implied. We also know that if A is provable then "A or B" is provable. We have to show that then "A or B" too is implied. We do so by appeal to the semantic definition and the assumption we just made. A is provable from G, we assume. So it is also implied by G. So any semantic valuation making all of G true makes A true. But any valuation making A true makes "A or B" true, by the defined semantics for "or". So any valuation which makes all of G true makes "A or B" true. So "A or B" is implied.) Generally, the Inductive step will consist of a lengthy but simple case-by-case analysis of all the rules of inference, showing that each "preserves" semantic implication.

By the definition of provability, there are no sentences provable other than by being a member of G, an axiom, or following by a rule; so if all of those are semantically implied, the deduction calculus is sound.

Sketch of completeness proof

(This is usually the much harder direction of proof.)

We adopt the same notational conventions as above.

We want to show: If G implies A, then G proves A. We proceed by contraposition: We show instead that If G does not prove A then G does not imply A.

QED

Another outline for a completeness proof

If a formula is a tautology, then there is a truth table for it which shows that each valuation yields the value true for the formula. Consider such a valuation. By mathematical induction on the length of the subformulae, show that the truth or falsity of the subformula follows from the truth or falsity (as appropriate for the valuation) of each propositional variable in the subformula. Then combine the lines of the truth table together two at a time by using "(P is true implies S) implies ((P is false implies S) implies S)". Keep repeating this until all dependencies on propositional variables have been eliminated. The result is that we have proved the given tautology. Since every tautology is provable, the logic is complete.

Alternative calculus

It is possible to define another version of propositional calculus, which defines most of the syntax of the logical operators by means of axioms, and which uses only one inference rule.

Axioms

Let \phi\,\!, \chi\,\! and \psi\,\! stand for well-formed formulæ. (The wffs themselves would not contain any Greek letters, but only capital Roman letters, connective operators, and parentheses.) Then the axioms are as follows:

Axioms
Name Axiom Schema Description
THEN-1 \phi \to (\chi \to \phi) Add hypothesis \chi\,\!, implication introduction
THEN-2 (\phi \to (\chi \to \psi)) \to ((\phi \to \chi) \to (\phi \to \psi)) Distribute hypothesis \phi\,\! over implication
AND-1 \phi \land \chi \to \phi Eliminate conjunction
AND-2 \phi \land \chi \to \chi  
AND-3 \phi \to (\chi \to (\phi \land \chi)) Introduce conjunction
OR-1 \phi \to \phi \lor \chi Introduce disjunction
OR-2 \chi \to \phi \lor \chi  
OR-3 (\phi \to \psi) \to ((\chi \to \psi) \to (\phi \lor \chi \to \psi)) Eliminate disjunction
NOT-1 (\phi \to \chi) \to ((\phi \to \neg \chi) \to \neg \phi) Introduce negation
NOT-2 \phi \to (\neg \phi \to \chi) Eliminate negation
NOT-3 \phi \lor \neg \phi Excluded middle, classical logic
IFF-1 (\phi \leftrightarrow \chi) \to (\phi \to \chi) Eliminate equivalence
IFF-2 (\phi \leftrightarrow \chi) \to (\chi \to \phi)  
IFF-3 (\phi \to \chi) \to ((\chi \to \phi) \to (\phi \leftrightarrow \chi)) Introduce equivalence

Axiom THEN-2 may be considered to be a "distributive property of implication with respect to implication."
Axioms AND-1 and AND-2 correspond to "conjunction elimination". The relation between AND-1 and AND-2 reflects the commutativity of the conjunction operator.
Axiom AND-3 corresponds to "conjunction introduction."
Axioms OR-1 and OR-2 correspond to "disjunction introduction." The relation between OR-1 and OR-2 reflects the commutativity of the disjunction operator.
Axiom NOT-1 corresponds to "reductio ad absurdum."
Axiom NOT-2 says that "anything can be deduced from a contradiction."
Axiom NOT-3 is called "tertium non datur" (Latin: "a third is not given") and reflects the semantic valuation of propositional formulae: a formula can have a truth-value of either true or false. There is no third truth-value, at least not in classical logic. Intuitionistic logicians do not accept the axiom NOT-3.

Inference rule

The inference rule is modus ponens:

Meta-inference rule

Let a demonstration be represented by a sequence, with hypotheses to the left of the turnstile and the conclusion to the right of the turnstile. Then the deduction theorem can be stated as follows:

If the sequence
 \phi_1, \ \phi_2, \ ... , \ \phi_n, \ \chi \vdash \psi
has been demonstrated, then it is also possible to demonstrate the sequence
 \phi_1, \ \phi_2, \ ..., \ \phi_n \vdash \chi \rightarrow \psi .

This deduction theorem (DT) is not itself formulated with propositional calculus: it is not a theorem of propositional calculus, but a theorem about propositional calculus. In this sense, it is a meta-theorem, comparable to theorems about the soundness or completeness of propositional calculus.

On the other hand, DT is so useful for simplifying the syntactical proof process that it can be considered and used as another inference rule, accompanying modus ponens. In this sense, DT corresponds to the natural conditional proof inference rule which is part of the first version of propositional calculus introduced in this article.

The converse of DT is also valid:

If the sequence
 \phi_1, \ \phi_2, \ ..., \ \phi_n \vdash \chi \rightarrow \psi
has been demonstrated, then it is also possible to demonstrate the sequence
 \phi_1, \ \phi_2, \ ... , \ \phi_n, \ \chi \vdash \psi

in fact, the validity of the converse of DT is almost trivial compared to that of DT:

If
 \phi_1, \ ... , \ \phi_n \vdash \chi \rightarrow \psi
then
1:  \phi_1, \ ... , \ \phi_n, \ \chi \vdash \chi \rightarrow \psi
2:  \phi_1, \ ... , \ \phi_n, \ \chi \vdash \chi
and from (1) and (2) can be deduced
3:  \phi_1, \ ... , \ \phi_n, \ \chi \vdash \psi
by means of modus ponens, Q.E.D.

The converse of DT has powerful implications: it can be used to convert an axiom into an inference rule. For example, the axiom AND-1,

 \vdash \phi \wedge \chi \rightarrow \phi

can be transformed by means of the converse of the deduction theorem into the inference rule

 \phi \wedge \chi \vdash \phi

which is conjunction elimination, one of the ten inference rules used in the first version (in this article) of the propositional calculus.

Example of a proof

The following is an example of a (syntactical) demonstration, involving only axioms THEN-1 and THEN-2:
Prove: AA (Reflexivity of implication).
Proof:

1. (A → ((BA) → A)) → ((A → (BA)) → (AA))
Axiom THEN-2 with φ = A, χ = BA, ψ = A
2. A → ((BA) → A)
Axiom THEN-1 with φ = A, χ = BA
3. (A → (BA)) → (AA)
From (1) and (2) by modus ponens.
4. A → (BA)
Axiom THEN-1 with φ = A, χ = B
5. AA
From (3) and (4) by modus ponens.

Equivalence to equational logics

The preceding alternative calculus is an example of a Hilbert-style deduction system. In the case of propositional systems the axioms are terms built with logical connectives and the only inference rule is modus ponens. Equational logic as standardly used informally in high school algebra is a different kind of calculus from Hilbert systems. Its theorems are equations and its inference rules express the properties of equality, namely that it is a congruence on terms that admits substitution.

Classical propositional calculus as described above is equivalent to Boolean algebra, while intuitionistic propositional calculus is equivalent to Heyting algebra. The equivalence is shown by translation in each direction of the theorems of the respective systems. Theorems Φ of classical or intuitionistic propositional calculus are translated as equations Φ = 1 of Boolean or Heyting algebra respectively. Conversely theorems x = y\,\! of Boolean or Heyting algebra are translated as theorems (x \to y) \land (y \to x) of classical or propositional calculus respectively, for which x \equiv y is a standard abbreviation. In the case of Boolean algebra x = y\,\! can also be translated as (x \land y) \lor (\neg x \land \neg y), but this translation is incorrect intuitionistically.

In both Boolean and Heyting algebra, inequality x \le y can be used in place of equality. The equality x = y\,\! is expressible as a pair of inequalities x \le y and y \le x. Conversely the inequality x \le y is expressible as the equality x \land y = x, or as x \lor y = y. The significance of inequality for Hilbert-style systems is that it corresponds to the latter's deduction or entailment symbol \vdash. An entailment

 \phi_1, \ \phi_2, \ldots, \ \phi_n \vdash \psi

is translated in the inequality version of the algebraic framework as

 \phi_1\ \land\ \phi_2\ \land\ \ldots\ \land \ \phi_n\ \ \le\ \ \psi

Conversely the algebraic inequality x \le y is translated as the entailment

x\ \vdash\ y

The difference between implication x \to y and inequality or entailment x \le y or x\ \vdash\ y is that the former is internal to the logic while the latter is external. Internal implication between two terms is another term of the same kind. Entailment as external implication between two terms expresses a metatruth outside the language of the logic, and is considered part of the metalanguage. Even when the logic under study is intuitionistic, entailment is ordinarily understood classically as two-valued: either the left side entails, or is less-or-equal to, the right side, or it is not.

Similar but more complex translations to and from algebraic logics are possible for natural deduction systems as described above and for the sequent calculus. The entailments of the latter can be interpreted as two-valued, but a more insightful interpretation is as a set, the elements of which can be understood as abstract proofs organized as the morphisms of a category. In this interpretation the cut rule of the sequent calculus corresponds to composition in the category. Boolean and Heyting algebras enter this picture as special categories having at most one morphism per homset, i.e. one proof per entailment, corresponding to the idea that existence of proofs is all that matters: any proof will do and there is no point in distinguishing them.

Graphical calculi

Main article: Logical graph

It is possible to generalize the definition of a formal language from a set of finite sequences over a finite basis to include many other sets of mathematical structures, so long as they are built up by finitary means from finite materials. What's more, many of these families of formal structures are especially well-suited for use in logic.

For example, there are many families of graphs that are close enough analogues of formal languages that the concept of a calculus is quite easily and naturally extended to them. Indeed, many species of graphs arise as parse graphs in the syntactic analysis of the corresponding families of text structures. The exigencies of practical computation on formal languages frequently demand that text strings be converted into pointer structure renditions of parse graphs, simply as a matter of checking whether strings are wffs or not. Once this is done, there are many advantages to be gained from developing the graphical analogue of the calculus on strings. The mapping from strings to parse graphs is called parsing and the inverse mapping from parse graphs to strings is achieved by an operation that is called traversing the graph.

Other logical calculi

Propositional calculus is about the simplest kind of logical calculus in any current use. (Aristotelian "syllogistic" calculus, which is largely supplanted in modern logic, is in some ways simpler — but in other ways more complex — than propositional calculus.) It can be extended in several ways.

The most immediate way to develop a more complex logical calculus is to introduce rules that are sensitive to more fine-grained details of the sentences being used. When the "atomic sentences" of propositional logic are broken up into terms, variables, predicates, and quantifiers, they yield first-order logic, or first-order predicate logic, which keeps all the rules of propositional logic and adds some new ones. (For example, from "All dogs are mammals" we may infer "If Rover is a dog then Rover is a mammal".) It makes sense to refer to propositional logic as "zeroth-order logic", when comparing it with first-order logic and second-order logic.

With the tools of first-order logic it is possible to formulate a number of theories, either with explicit axioms or by rules of inference, that can themselves be treated as logical calculi. Arithmetic is the best known of these; others include set theory and mereology.

Modal logic also offers a variety of inferences that cannot be captured in propositional calculus. For example, from "Necessarily p" we may infer that p. From p we may infer "It is possible that p". The translation between modal logics and algebraic logics is as for classical and intuitionistic logics but with the introduction of a unary operator on Boolean or Heyting algebras, different from the Boolean operations, interpreting the possibility modality, and in the case of Heyting algebra a second operator interpreting necessity (for Boolean algebra this is redundant since necessity is the De Morgan dual of possibility). The first operator preserves 0 and disjunction while the second preserves 1 and conjunction.

Many-valued logics are those allowing sentences to have values other than true and false. (For example, neither and both are standard "extra values"; "continuum logic" allows each sentence to have any of an infinite number of "degrees of truth" between true and false.) These logics often require calculational devices quite distinct from propositional calculus. When the values form a Boolean algebra (which may have more than two or even infinitely many values), many-valued logic reduces to classical logic; many-valued logics are therefore only of independent interest when the values form an algebra that is not Boolean.

Solvers

Finding solutions to propositional logic formulas is an NP-complete problem. However, recent breakthroughs (Chaff algorithm, 2001) have led to the development of small, efficient SAT solvers, which are very fast for most cases. Recent work has extended the SAT solver algorithms to work with propositions containing arithmetic expressions; these are the SMT solvers.

References

See also

Higher logical levels

Related topics

  • Combinatory logic
  • Conceptual graph
  • Disjunctive syllogism
  • Entitative graph
  • Existential graph
  • Frege's propositional calculus
  • Implicational propositional calculus
  • Intuitionistic propositional calculus
  • Laws of Form
  • Logical graph
  • Logical value
  • Minimal negation operator
  • Multigrade operator
  • Operation
  • Parametric operator
  • Peirce's law
  • Propositional formula
  • Symmetric difference
  • Truth table

Related works

External links