Doubt about the definition of free vector spaces.

Solution 1:

We know that a map out of a vector space is determined by its action on a basis, so if we want a free vector space on a set $X$, that means we want to have a basis element for each element of $X$. Let's write $\delta_x$ for the basis element associated to $x \in X$.

What does it mean that the family $\{ \delta_x \}_{x \in X}$ forms a basis? It means that every vector in our vector space can be written as a (finite!) linear combination of the $\delta_x$. That is, every vector looks like

$$a_1 \delta_{x_1} + a_2 \delta_{x_2} + \ldots + a_n \delta_{x_n}$$

for some $x_1, \ldots, x_n \in X$ and some $a_1, \ldots, a_n \in K$.

Now, of course, this is too easy. As mathematicians it's our job to find slick presentations of objects at the risk of confusing new students. Indeed, depending on the mathematicians you talk to, confusing new students is actually a feature rather than a bug (I'm joking, but only a little bit).

Really what happens is we would like to write our free vector space in terms of another, more concrete, vector space so that it's easy to check that it really is a vector space. We need to put some complexity somewhere, and by making the definition a bit more opaque it can make the coming proofs much cleaner. This is great for the person writing the textbook (who knows why the opaque definition works, but doesn't want to check that addition is associative, etc.) but is less great for the person reading the textbook (who is still building intuition for these objects).

So how does the idea of "linear combinations of the $\delta_x$" evolve over time?

Well, notice we need precisely the information of a coefficient $a_x$ for each $\delta_x$. Then we can look at the vector $\sum_{x \in X} a_x \delta_x$. Of course, since we're only allowed finite linear combinations of the basis vectors, we need to know that each of these sums is finite. That is, that all but finitely many of the $a_x$ are $0$.

So we can actually get away with a function $a : X \to K$ so that $a_x = 0$ for all but finitely many $K$. Of course, once we have these functions it's easy to see that we don't need to write down the $\delta_x$ anymore. These functions themselves form a vector space, and the functions $\chi_x : X \to K$ with $\chi_x(y) = \begin{cases} 1 & y=x \\ 0 & y \neq x \end{cases}$ form a basis.

(As a quick exercise, do you see how the function $\chi_x$ corresponds to the vector $\delta_x$ in a natural way?)

So we see that the vector space can be thought of as functions $X \to K$ of finite support. Exactly the definition that was presented to you.


I hope this helps ^_^

Solution 2:

I don't think we choose this particular representation. It comes naturally by the fact that every element $V\in F(X)$ can be represented in a unique way as a combination of the elements of $X$ \begin{equation} V = \sum_{x\in X} v_x x \end{equation} and only a finite number of the $v_x$s are non zero.

It is natural to define the mapping that associates to $V$ and $x$ the coefficient of $V$ on the element $x$ \begin{align} &C: F(X)\times X \to K\\ &(V, x)\mapsto v_x \end{align} Now if one considers $f_V(x) = C(V, x)$, then $f_V$ is a function $X\to K$ with finite support and the mapping $V\mapsto f_V$ is a vector space isomorphism, thus we can identify $V$ and $f_V$.