How does one show a set of axioms is independent? (of each other?)

I am being asked to show that the groups axioms of existence of an identity, of inverses and of associativity are independent. Does this mean that none of two of them imply the third? That is: do I have to exhibit a "bad" group which has two but fails to have the third? (for each case) Or do I have to find examples also where one succeeds and the other two fail?

ADD The axiom of inverses uses the axiom of the identity, so, how can the axiom of identity be independent of that of the inverses? Ans One can state the axiom of inverses without resorting to the notion of a unity.

What does it mean, more formally, that a set of axioms is independent?


Solution 1:

The set $\mathcal{A}$ of axioms is independent if none of the elements of $\mathcal{A}$ is a consequence of the others. More formally still, the condition is that for every $\varphi\in \mathcal{A}$, the sentence $\varphi$ is not a theorem of the theory with axiom set $\mathcal{A}\setminus\{\varphi\}$.

That means your first interpretation is correct. To show independence, it is enough to produce for every $\varphi$ a structure $M_{\varphi}$ in which all the axioms in $\mathcal{A}\setminus\{\varphi\}$ are true , but $\varphi$ fails. This is a semantic argument. One can also imagine syntactic arguments, but semantic arguments are the mathematically natural ones.

Solution 2:

Let's try to summarize the already given answers but starting from the beginning.

Syntax

Usually we express things by using finite strings of symbols taken from a particular set. The way to formalize this idea is to define language as the set from which you pick these symbols.

Let ${\cal V}$ be a nonempty and large enough set. This will be the set of variables of the language.

Let $$\mathrm{BIN}=\{\vee,\wedge,\implies,\iff\}$$ the set of binary connectives.

Then we define the language $L$ as the set $$L:={\cal V}\cup \mathrm{BIN}\cup \{\neg\}\cup \{),(\}$$ where we agree that all the sets involved are pairwise disjoint. I think all this symbols are self explanatory. Its intended meaning is the usual.

From this language we can form words, for example if you take $A,B$ variables then $$A,\quad )B\neg,\quad (\neg A\vee B)$$ are words.

Intuitively, the words you would use to express something "useful" are called formulas. For example, perhaps we all agree that the second word in the example above is not a formula. To be precise, let ${\cal W}(L)$ the set of words of $L$. The set of formulas of $L$ is defined as the smallest subset $X$ of ${\cal W}(L)$ which satisfies:

  • it contains ${\cal V}$
  • if $F\in X$ then $\neg F\in X$
  • if $F\in X$ and $G\in X$ then for all $\alpha\in\mathrm{BIN}$, $(F\mathbin{\alpha} G)\in X$

The set of formulas of the language $L$ is called ${\cal F}(L)$. There is an equivalent inductive way of define ${\cal F}(L)$.

Then you can try to decide when a given formula is "true" or "false". To do this you consider maps $\delta:{\cal V}\to \{0,1\}$ and then you extend it to a map $\bar{\delta}$ defined on ${\cal F}(L)$ in an inductive fashion. For example, once $\bar{\delta}$ is defined for some formulas $A,B$ then $$\bar{\delta}((A\wedge B))=\delta{A}\cdot\delta{B}$$ which intuitively says that the formula $(A\wedge B)$ is false whenever $A$ or $B$ is false.

In this way you can define tautologies among other things, and study relations between formulas and theories (here theory is just a set of formulas), for example you might want to say when two different formulas are essentially the same.

The study of all this stuff is called propositional calculus.

But math is quite more complicated so we need more to abstract it.

Let $\cal C$ a nonempty set, a first order language $L$ is a set $$\begin{align*} L &:={\cal V}\cup \{\forall,\exists\}\cup \mathrm{BIN}\cup \{\neg\}\cup \{),(\}\\ &\phantom{{}:=}\cup{\cal C}\cup \bigcup_{i=1}^\infty \cal{F}_i \cup \bigcup_{i=1}^\infty \cal{R}_i \end{align*}$$ where all the sets involved are pairwise disjoint and:

  • for each $n\geq 1$ the elements of ${\cal F}_n$/${\cal R}_n$ are called $n$-ary function/relation symbols.
  • the elements of $\cal C$ are called constant symbols.
  • Notice that (if present) the symbol for equality $\simeq$ is an element of $\cal{R}_2$.

The elements in the first row above are called logical symbols since they are common to all first order languages. When you specify a first order language you must only say what the no-logical symbols are. For example, for the language of groups you need a constant symbol, say $e$, and a binary function symbol, say $\ast$. This language would be then $$L_G=\{e,\ast\}.$$

As before you can define "useful" words from this language, and you can define formulas. But here we must be more careful. First we must define terms which are variables, constants or "valid" combinations of those with function symbols. For example, with $L_G$, $$v_0*v_1$$ would be a term.

Then we must define atomic formulas which are terms or "valid" combinations of those with relation symbols. Again with $L_G$, $$v_0*v_1\simeq e$$ would be an atomic formula (when not stated otherwise, all languages have the equality sign).

And from the atomic formulas we can define the formulas. For example $$\forall v_0(v_0*v_1\simeq e),\quad v_1\simeq e\vee \exists v_0(v_0\simeq v_1).$$ This formulas are called predicate formulas.

We say that a variable occur in a formula if it appear in the formula. Intuitively occurrence of a variable is bounded if "it is in the scope of some quantifier" and it is free otherwise. You can see examples in the above formulas. A formula is closed if it have no free occurrences of variables.

A theory $T$ of a first order language $L$ is a set of closed formulas of $L$.

Semantics

Okay, but this time how do you say which formulas are true or false? How do you give meaning to the formulas?

This is done with the structures. Given a first order language an $L$-structure is a nonempty set plus some functions and relations defined on it to give interpretations to functions, relations and constants symbols.

For example consider the language: $$L=\{R,s\}$$ where $R$ is a binary relation and $s$ is a binary function. Then $${\cal M}=\langle\Bbb R,\leq,+\rangle$$ is an $L$-structure.

In the structures you can interpret the atomic formulas, but this interpretation depends of the interpretation of the variables (if present) that you give. For example with the above language and $L$-structure the atomic formula $s(v_0,v_0)Rv_1$ is $$1+1\leq 1$$ when $v_0$ and $v_1$ are interpreted as $1$ and is $$0+0\leq 1$$ when $v_0$ is interpreted as $0$ and $v_1$ as 1.

Let's concentrate in closed formulas now. Given a first order language the structures decide whether or not a closed formula is true or false. Intuitively an $L$-structure $\cal M$ satisfies a closed formula $F$ if what is stated by the formula is indeed true in the structure. For example consider the language $$L=\{R,c\}$$ where $R$ is a binary relation and $c$ is a constant symbol, and the closed formula $$F=\forall v(vRc).$$ Consider the $L$-structures $${\cal M}=\langle \Bbb R,\geq,0\rangle,\qquad {\cal N}=\langle \Bbb N, \geq, 1\rangle.$$ Then notice that $F$ is not satisfied in $\cal{M}$ since for example $-1\not\geq 0$, but $\cal{N}$ does satisfies $F$ as you can see. In this case we write $${\cal N}\models F$$ and we say that $\cal{N}$ is a model for $F$.

Now, fix a first order language $L$.

  • Given a theory $T$ of a language $L$ we say that an $L$-structure $\cal{M}$ is a model for $T$ if for each $F\in T$, ${\cal M}\models F$.
  • We say that a formula $F$ is semantic consequence of a theory $T$, and write $T\vdash^\ast F$ if for each $L$-structure $\cal{M}$ $${\cal M}\models T\quad\text{implies}\quad {\cal M}\models F.$$
  • A theory $T$ is universally valid if any $L$-structure is a model for $T$.

In the other hand the concept of proof can be formalized as well. Given a theory $T$ of $L$ a derivation of a formula $F$ is a finite sequence of formulas $(F_0,\ldots,F_n)$ with $F_n=F$ and where each formula $F_i$ satisfies:

  • $F_i\in T$ or
  • $F_i$ is a tautology (obtained by substituting the variables in the tautologies of propositional calculus by predicate formulas) or
  • $F_i$ is in one of the axiom schemas (which are some special sets of universally valid formulas) or
  • $F_i$ is deduced from previous formulas in the sequence by some of the rules of inference (for example a rule of inference is modus ponens: from $A$ and $A\implies B$ you can deduce $B$).

We say that a closed formula $F$ is syntactic consequence of a theory $T$ if $F$ can be derived from $T$. In that case we write $T\vdash F$.

One way to state the completeness theorem of first order logic is to say that being syntactic or being semantic consequence is the same thing. That is:

Given a theory $T$ and a formula $F$ of $L$, $$T\vdash^\ast F\quad\text{if and only if}\quad T\vdash F.$$

That is, if you want to proof that a formula $F$ can not be derived from a theory $T$ it is enough to exhibit a model of $T$ which is not a model of $F$.

The language of groups and the theory of groups.

Let $$L_G'=\{\ast\}\quad\text{and}\quad L_G=\{e,\ast\}$$ where $*$ is a binary function and $e$ is a constant symbol.

Notice that $L_G'\subseteq L_G$ and that the $L_G'$-structures are of the form $${\cal M'}=\langle M,+\rangle$$ where $+$ is a binary function defined on $M$ and is of course the interpretation of $\ast$ in $\cal M'$. But since $M$ must be nonempty there is a element $m$ that could be used as the interpretation of $c$. That is $${\cal M}=\langle M,m,+\rangle$$ is an $L_G$-structure.

When you have two languages $L'\subseteq L$ and an $L'$-structure $\cal M'$ and you keep the same underlying set, the functions, relations and constants that you already have, and add some new constants, functions and relations in order to obtain an $L$-structure $\cal M$, $\cal M$ is called an enrichment of $\cal M'$.

It must be intuitively clear, that if $F$ is a closed formula of $L'$ then $${\cal M'}\models F\qquad\text{iff}\qquad {\cal M}\models F.$$

Now, let $A$ be the closed formula of $L_G$ and $L_G'$ given by $$A=\forall v\forall w\forall x(v\ast(w\ast x)\simeq (v\ast w)\ast x).$$ Let $$T_G=\{A,\forall x(x\ast e\simeq e\ast x\simeq x),\forall x\exists y(x\ast y\simeq y\ast x\simeq e) \}.$$ You immediately recognize this as the axioms for of groups (notice the abuse of the language with $\simeq$, there must be some $\wedge$s).

But what if you try to write some reasonable axioms for groups with the language $L_G'$? Here you do not have the constant symbol $e$, so you must add something like: $$\begin{gather*} Id=\exists u \forall x(x\ast u\simeq u\ast x\simeq x)\\ In=\forall x \exists y\forall w(w\ast (x\ast y)\simeq w\ast (y\ast x)\simeq w) \end{gather*}$$ Let $$T_G'=\{A,Id,In\}.$$ If you start with $L_G'$ and $T_G'$, it turns out that the formula that says that the identity element is unique can be deduced from $T_G'$. So you can enrich the language to $L_G$ and then the models of $T_G'$ will be (with the right interpretation of $e$) exactly the models of $T_G$.

And, as you can see the formulas of $T_G$ are simpler. In fact with $L_G$ it is possible to have a theory of groups with a single elegant axiom (This is always possible, since the theory is finite and you can take the conjunction of the axioms, but that's not the axiom I'm talking about: look).

So it is easier to say things if your language is richer.


For your exercise, for sure it's supposed you must use $L_G$ and $T_G$. So you must follow what is given in the other answers.

Hope this helps.

Solution 3:

To say that a set of axioms $T$ is independent is to say that for every $\varphi\in T$, there is a model of $T\setminus\{\varphi\}$ in which $\varphi$ is false, and another in which it is true. This is, of course, a consequence of the completeness theorem of first-order logic.

For example, in the theory of abelian groups the commutative axiom is independent of the others. $S_6$ is a non-abelian group, and the trivial group is an abelian group.

Solution 4:

I will preface the following by saying that like you I don't currently understand how the axiom of inverses can exist independently of the axioms of identity for a group. I can see how associativity can get shown as independent.

The concept of independence of axioms may get fruitfully compared with what we could call a non-independent axiom. A non-independent axiom of a system consists of an axiom which comes as a necessary consequence of the other axioms of the system. In the theory of groups, statements like the cancellation law and the involutive law --x=x or equivalently x--=x in reverse polish notation, can get called non-independent axioms in the sense that given the 3 standard axioms used for group theory, they necessarily follow. Within the theory of groups, those axioms cannot "stand alone" from the 3 axioms most commonly used to start group theory.

As perhaps a better example, you can find authors like Nguyen and Walker in "A First Course in Fuzzy Logic" who will axiomitize lattices as commutative, idempotent, absorptive, and associative. But, idempotence can get called as non-independent, since you can derive it from the other axioms. In contrast, an axiom x qualifies as independent of a set of axioms S if it can "stand alone" from all the axioms of S, in the sense that we can't derive x from the axioms of S.

So, suppose we have a finite set of axioms {a, ..., n}. If I want to show axiom n independent of the other axioms, then I need to show all axioms belonging to {a, ..., m} can hold true, while axiom n does not hold true. Thus, we can't start with axiom set {a, ..., m} and derive axiom n, because then we could have a true statement that could lead to a false one. If I want to show the set of axioms {a, b, c} independent, then I want to show that for all members of {a, b, c}, independence holds for each of the axioms, or that each axiom is independent, or can stand alone without the others. So, I'll want to show that {a, b} can hold true while c does not hold true, {a, c} can hold true while b does not hold true, and {b, c} can hold true while "a" does not hold true. Sometimes one does this by exhibiting concrete algebraic systems (similar to what you called ""bad" groups") which make the independence of each axiom (hopefully) clear. These systems give us a model of all of the axioms except one of them.

As an example of how to demonstrate independence of a set of axioms, let's try and show the axioms for semilattices independent of each other. A semilattice as an algebraic structure consists of a set S with a binary operation S which satisfies commutativity, associativity, and idempotence. Or more symbolically for (S, S) it holds that for all x, y, z belonging to S we have:

1-xyzSS=xySzS associativity

2-xxS=x idempotence

3-xyS=yxS commutativity.

So, let's first consider algebraic structure ({0, 1}, E) with E defined by this table:

 E  0  1
 0  1  0
 1  0  1

Idempotence fails since 00E=1, while commutativity can get seen to hold by inspection, and associativity holds since this can get understood as the table for logical equivalence, which associates. Thus, axiom 2 is independent of axiom set {1, 3}. If I understand logical terminology correctly, we can say that ({0, 1}, E) models commutativity and associativity but not idempotency.

Next let's consider ({0, 1, 2}, A) with A defined by this table:

A  0  1  2
0  0  1  1
1  1  1  2
2  1  2  2

Commutativity and idempotence hold, while 012AA=02A=1 and 01A2A=12A=2 so associativity fails. Thus, axiom 1 is independent of {2, 3}.

Finally consider structure ({0, 1}, B) with B defined by this table:

B  0  1
0  0  0
1  1  1

Idempotence holds, and for this structure for all x, xyB=x. Thus, xyzBB=x, and xyBzB=xzB=x, and thus associativity holds. Commutativity fails, and thus axiom 3 is independent of {1, 2}.

Consequently, it follows that all axioms for a semilattice are independent in the sense that for all of the axioms in the axiom set you can have all of the axioms hold true but one of them hold false, let's call the true axiom o and the set of axioms with without axiom o set T, and not have the ability to validly derive axiom o from set T.