Why do we need (the abstract concept of) random variables (in discrete probability models)?

What we defined: Suppose we have a (discrete) probability model $\left(\Omega,P\right)$, where $P$ is the probability function (at least, that was the way it was introduced in a course I took; that means only that $\Omega$ is at most countable, that $P\left(\bigcup_{i}A_{i}\right)=\sum P\left(A_{i}\right)$, for all (at most countable and disjoint) events $A_{i}$ and that of course $P\left(\Omega\right)=1$). We defined a random variable (rd from now on) $X$ to be a mapping $X:\Omega\rightarrow\mathbb{R}$ and then discussed some aspects of $P\left(X=k\right)$ for some $k\in\mathbb{R}$ .

What bothers me: To me, this definition seems rather artificial: Why define a mapping like that, if there is no apparent need for it (at least in this probability model) ?

Since if we have an event $A\subseteq\Omega$, that depends on some parameter $k\in\mathbb{R}$, (for example the sum of the faces of dice be $k$) then we could just as easily define a collection of events $A_{k}$ - one for each parameter - and discuss aspects of $P\left(A_{k}\right)$, instead of the above way by using $X$. Of course one could now argue, that $A$ does not always need to depend on some parameter $k$, so one in some cases really has to "convert" probabilities to numbers via $X$, but of all examples that I have seen until now, even in $\Omega$ isn't made up out of numbers, somewhere a parameter $k$ does sneak in, so we could equivalently work with $P\left(A_{k}\right)$ instead of $P\left(X=k\right)$, since defining the subsets of $\Omega$ whose elements we want to count (to establish the probability of the subset) always amounts to using some $k$ in the definition of those subsets (this reasoning extends of course also to other cases like when we consider $P\left(X\leq k\right)$, since this can also be circumvented by considering an appropriate $P\left(\cup_{j\leq k}A_{j}\right)$).

Thus, introducing rd's seems to me to be a superfluous definition of events without specifying $k$"


In short, the abstract concepts are for blurring out the details, i.e. the abstract concept of random variables is useful because it allows us to work at higher level than ordinary atoms.

To give some more motivation:

  • You can have more than one random variable on single probability space, but doing it with atoms you would need to care for every possible combination.
  • The notion of conditional expected value $\mathbb{E}(X | Y)$ is very useful, but cumbersome to define in terms of atoms.
  • You can nicely characterize the random variable in terms of characteristic function.

Finally, there's one more thing to add: humans have a very strong intuition about randomness, but in mathematics there is no randomness at all. All you have is just functions (I will skip other entities for simplicity), and every time you apply the same arguments to a function you get the same result. I will repeat: in mathematics there is no randomness at all.

How one deal with that? By introducing something like "the state of the world" or what we often denote by $\Omega$, the universe. So if the world happens to be in some state $\omega_1 \in \Omega$, then the outcome of every experiment is precisely determined, we can predict perfectly every outcome of every action (tossing a coin or whatever it is). However, we don't know in which state the world is, and this way we hacked the randomness into math. And this interpretation gives natural rise to the definition of random variables -- they are only random, because we do not know in which state of the world we live in.

To conclude, it does not matter whether $\Omega$ is finite or not, it does not matter if we prefer to deal with raw sets and atoms, or theorems about abstract concepts, the notion of a random variable is very useful, and I would even say that is the intuitive connection between pure mathematics and the imperfect (and therefore beautiful) world we live in.

Edit: To answer one of the comments of OP (to long for another comment), random variables aren't theoretically necessary, as we could inline their definitions into our proofs, etc., but in practice any such proof would be unreadable and unmanageable, so random variables are necessary. Consider real numbers: one could always work with sequences of rational numbers or continued fractions, or whatever, so the question is, are they essential? I think yes, and there are many similar examples.


If everything was about computing probabilities of events $A_k$, you might be able to get away with avoiding random variables. However, there's a lot more to probability than that. When, for example, you want to calculate means and variances, or talk about the relations between many "parametrized families", it's very useful to have random variables.