Arguments pro material implication

So far I've understood that not every "If" in natural English can be symbolized by the material implication ⊃ for truth-functionally

But then, I find in several books an argument "for ⊃" stating that:

"If A then B" is true iff A ⊃ B is true.

Then it develops proving that statement. I agree that they are logically equivalent but I do not see how this might be an argument for ⊃ being a good representative of the "if" in English.

As I understand it, in order to develop the proof of the statement you need to assume "If A then B" has the nature of the material implication. So I think of it as an attempt to define a word with the same word.

Why would it be a good argument?


Solution 1:

My personal but ever more strongly held view is that relating formal logic to natural language is something an introductory book on logic (and most books on logic) should not do. Many books do do this, but, in my opinion, this just creates unnecessary confusion, a misrepresentation of how logic is used by most mathematicians, and in some cases is just meaningless. I do agree that linguistics is an application area of logic, but the picture of this presented in most (all?) introductory logic texts that take this approach is utterly facile.

To start, if this is the "definition" of $A \supset B$ you were given, then it is completely inadequate. Even when we are applying logic to lingistics, the goal is to understand natural language utterances by translating them to logic formulas. This is of no use if the logical formulas are "defined" by natural language utterances as you point out. When we're mathematicians and logicians and not linguists, then "If $A$, then $B$" is simply not unambiguous nor something that can be mathematically manipulated.

Now, what you can do is talk about how to read logical formulas. You can say $A\supset B$ is read as "if $A$, then $B$". This is a unidirectional relationship and generally not meant as a definition, but merely as a communication aid. For example, I can say $x+y$ can be read as "add $x$ to $y$" but 1) this clearly doesn't explain anything about addition, and 2) this doesn't mean "add Tom to the invitation" means $\text{Tom}+\text{the invitation}$.

There are two ways of defining the logical connectives. One approach is syntactic where we give rules and axioms that show how a connective can be used. For example, one way of defining $\supset$ is via the following two rules: $$\cfrac{\Gamma, A\vdash B}{\Gamma\vdash A\supset B}\supset\!\!I\qquad\cfrac{\Gamma\vdash A \supset B \quad \Gamma\vdash A}{\Gamma\vdash B}\supset\!\!E$$ What's actually happening here is these two rules form part of a definition of a binary relation we've notated as $\vdash$. The second rule, $\supset\!\!E$, states (in the informal metalogic) that if $\Gamma \vdash A\supset B$ and $\Gamma \vdash A$ hold, then $\Gamma\vdash B$ holds. The collection of all such rules gives an inductive definition of the $\vdash$ relation. Informally, $\Gamma\vdash A$ means "$A$ is provable given assumptions $\Gamma$". This is, again, just a way to read the notation or at best a guide to the desired interpretation. The definition is the collection of rules.

The second approach of defining logical connectives is semantic. Here we systematically map logical formulas into mathematical objects. The most well-known such semantics is one for classical propositional logic where formulas are mapped into Boolean functions which are usually presented as "truth tables". Here the interpretation of a formula $A\supset B$, written $[\![A\supset B]\!]$ must be of the form $[\![\supset]\!]([\![A]\!],[\![B]\!])$ and the interpretation of $\supset$ is $$[\![\supset]\!](x,y)=\begin{cases}0, &\text{if }x=1\text{ and }y=0\\1, &\text{otherwise}\end{cases}$$ Many people take issue with this definition when they first see it. Part of this is due to the use of loaded terms like "true" and "false". Another part is that this semantics is only adequate for classical propositional logic. More complicated semantics are needed for more complicated logics like classical first-order logic. In general, different logics will require different notions of semantics, e.g. truth tables are not suitable for intuitionistic propositional logic. What this definition does accomplish is reducing the meaning of logical formulas to a clear and unambiguous notion that can be manipulated mathematically. For those who do want to translate natural language to logical formulas, they will need to disambiguate and spell out nuances in the translation, but they will at least have a solid target whose consequences can be calculated to cross-check.

In some cases these two approaches are connected by soundness and completeness (meta)theorems. These theorems together imply that a given syntactic notion of proof/derivability coincides with a given semantic notion of validity.

So, to answer your question, there is no good argument because 1) this is not how implication is defined, 2) matching the natural language "if" is not the goal, and 3) $\supset$ is indeed not a good representation for the natural language "if". The issue is natural language is messy, ambiguous, and inconsistent. We clearly don't want those properties to apply to our formalization, so the primary goal of mathematical clarity overrides any secondary goal of corresponding to natural language. This pushes the complexity of mapping a natural language statement into a logical formula to the translation process where it won't impact consumers of logic that aren't interested in linguistics. Even for linguists, it makes sense to have a precise language where different interpretation of superficially similar or even the same statement look different.

In practice, classical propositional logic is way too simple to capture, in any way, the meaning of many natural language statements even ignoring nuance. Different and more powerful logics are typically used. It is a completely reasonable goal to want to build a logic that more closely captures natural language utterances, but this is a goal at the intersection of linguistics and logic, not the goal of logic as a whole. There are plenty of reasons why the definitions given above are useful mathematically that are completely unrelated to natural language.

Solution 2:

The usual arguments for assigning simple indicative conditionals the same truth-conditional content as the material conditional are variants of the following.

Argument 1:

(i) Suppose if A then C. So we either have not-A, or we have A and hence C. So if A then C implies either not-A or C.

(ii) Conversely, suppose we are given either not-A or C. Then if not the first, then the second. So we can infer if A then C.

Hence if A then C is equivalent to the truth-functional not-A or C.

Argument 2

(i) The claim if A then C rules out having A true and C false. So if A then C implies it isn’t the case that both A and not-C.

(ii) Conversely, suppose we are given that it isn’t the case that both A and not-C. Then we can infer that if A is actually true we can’t have not-C as well: in other words if A then C.

Hence if A then C is equivalent to the truth-functional not(A and not-C).

But each of not-A or C and not(A and not-C) is (uncontroversially) equivalent to the material conditional we symbolize A $\supset$ C. Hence if A then C is true if and only if the corresponding A $\supset$ C is true.


Now, I'm not claiming that these arguments are irresistible. But we can make three points

  1. We do have some substantial arguments here: these are not e.g. bogus "attempt(s) to define a word with the same word".
  2. If you are going to reject the identification of the basic, truth-relevant content of 'if A then C' with the material conditional, you need to explain where Arguments (1) and (2) go wrong....
  3. ... and you need to explain too why Arguments (1) and (2) have looked as compelling as they do to many logicians.

But equally, of course, if you do accept these Arguments (1) and (2), then you have other tasks, basically to explain away why the identication of 'if' and '$\supset$' leads to such counter-intuitive results in certain cases!

The literature on both sides of the debate is enormous, and inconclusive! My view, for what it is worth, is that Frege's attitude in his Begriffsschrift, at the birth of modern logic, is the right one. Treat (what we now call) the material conditional as a workable substitute for the ordinary conditional for mathematical purposes, and don't worry too much about its exact relation to messy ordinary usage.