Theorems true "with probability 1"?
Solution 1:
One sometimes hears "The Riemann Hypothesis is true with probability 1." What's meant is that the Riemann Hypothesis is true if a certain statement $S$ about the Möbius function $\mu(n)$ is true; $\mu(n)$ is a member of a very natural family of functions, a family on which there is a natural probability measure; with probability 1, a function chosen from this family satisfies $S$.
Solution 2:
A somewhat interesting development in partial differential equations is the study of the wellposedness of the initial value problem for randomised initial data. An example is this recent paper of Burq and Tzvetkov. Let me briefly and vaguely describe what they showed.
Wellposedness of Initial Value Problems
The roughest statement of the question of wellposedness for initial value problems in partial differential equations is to ask "if we prescribe the initial value at time $t = 0$, does there exist a unique solution for $t > 0$". However, to make mathematically precise, the first problem is to specify in what sense and in what sets you allow the solution to exist and require it to be unique (and also in what sense and in what sets do you prescribe the initial data). Well, since functions defined on any manifold form a vector space, we have that our set will at least have a vector space structure. Furthermore, it is realized that a physically natural requirement for wellposedness is an approximation property: that two initial data not differing by too much should generate solutions not differing too much. To quantify "not differing too much" we are led to require our sets to have topology (so we get a notion of continuity). This is in fact the usual notion of wellposedness:
An initial value problem is said to be wellposed if there exists some topological vector space (of functions defined at $t=0$) $X$ of initial data, and some topological vector space (of functions defined on $T > t\geq 0$) $Y$ of solutions such that there exists a continuous function $f:X\to Y$ mapping initial data to solutions.
The main work is then finding a suitable pair of $(X,Y)$ such that $f$ can exist as a function (so the domain is whole of $X$ and for each element in $X$ there is only one image in $Y$) and is continuous.
Regularity issues
Now, it is quite clear that if you make $X$ smaller, then it is "easier" to find the mapping $f$: there are "fewer" elements (less pathological behaviour) to worry about, and so it is more likely you will find that the solution map exists for all initial data in $X$. It is also quite clear that if you make the topology finer continuity becomes easier to prove. If your topology on the vector space is given by a family of semi-norms, then adding more semi-norms will make the space "smaller", and generate a finer topology.
In reality, this is in fact the case: letting $X$ be in the $L^2$-Sobolev scale of function spaces $H^s$, where $s$ measures the number of derivatives you control, the larger the $s$, the more likely you can prove a wellposedness statement.
Burq and Tzvetkov studied a particular equation--the cubic semilinear wave equation, henceforth 3SLW--in their paper. It is known from previous results that in fact there is a cut-off for wellposedness. In the Sobolev scale, 3SLW is wellposed with the time of existence $T$ possibly only finite for $\frac12<s<\frac34$ (some of the inequality signs may actually be non-strict, but I don't remember which way the inclusions go for sure), and with $T$ infinite when $s > \frac34$. For $0 < s < \frac12$, however, it is known that wellposedness fails because the breakdown of continuity (the topology of $X$ becomes too coarse).
Probabilistic wellposedness
However, it is reasoned that the breakdown of continuity for 3SLW is a very special situation, and that "most" rough data will still give good behaviour. In this spirit, fixing $s$ to be any number between 0 and 1, Burq and Tzvetkov constructed a probability measure $\mu$ (see below) on $H^s$ such that there exists a full measure set $\Sigma\subset H^s$ such that the solution map is well-defined (existence and uniqueness) on $\Sigma$ and $\Sigma$ is invariant under the solution flow (which implies the time of existence is infinite). Furthermore, with respect to this measure $\mu$, they showed a "probabilistic continuity" of the solution map: that is, defining the set
$$ G(\epsilon,\delta):= \left\{ (V,U)\in X\times X \mid \|f(V)-f(U)\|_Y \geq \epsilon, \|V-U\|_X\leq \delta\right\} $$
of pairs of points where the solution differ by a large amount at least $\epsilon$ while the data differs by a small amount at most $\delta$, Burq and Tzvetkov could show that for any fixed $\epsilon$,
$$ \lim_{\delta\to 0} \left|G(\epsilon,\delta)\right|_{\mu\times\mu} = 0$$
in the product measure.
The measure
You may ask, well, what if you just define the measure to be supported precisely on $H^1 \subset H^s$? The main point of the construction is that given any fixed $H^s$ function, one can construct the measure $\mu$ by "randomising" the given function, in such a way that the regularity doesn't increase. (More precisely, they showed that if the given function is in $H^s$ but not in $H^r$ for $r > s$, then the measure $\mu$ they constructed will have the property that $H^r$ has $\mu$-measure zero. So the measure $\mu$ is essentially an object living in the low regularity space.)
Solution 3:
Two examples, of perhaps opposite kinds:
almost every real number is normal base 10, therefore (most likely) $\pi$ is normal base 10
Almost every real number is transcendental, therefore there exists a transcendental number. (Much easier proof than exhibiting with proof an individual transcendental number.)