Chebyshev's versus Markov's inequality
Markov's inequality is a "large deviation bound". It states that the probability that a non-negative random variable gets values much larger than its expectation is small.
Chebyshev's inequality is a "concentration bound". It states that a random variable with finite variance is concentrated around its expectation. The smaller the variance, the stronger the concentration.
Both inequalities are used to claim that most of the time, random variables don't get "unexpected" values.
A typical application of concentration bounds is polling - if you poll 200 people, then the standard deviation is so-and-so, and so the probability that the results are off by some amount $x$ is only $p$ which is very small. You can bound $p$ using Chebyshev's inequality, although there are better bounds.
Another application is in the statistics of patients in a hospital. Suppose they conduct a survey and find that the average number of patients per day is 10. Suppose they want to be able all patients 90% of the time. Then they need to handle at most 100 patients.
The list of probability results goes on. The law of large numbers states that the average of identical random variables tends (almost certainly) to the expectation. The central limit theorem describes the behavior of the noise, but only "close" to the expectation. Chernoff bounds are large deviation bounds useful for understanding the behavior of the noise away from the expectation.
As you can see in the wikipedia links, Markov's inequality can be used to prove Chebyshev's Inequality.
$t\mu\{x\in X : |f(x)|\geq t\}\leq\int_Xf$ is simple to remember. take a value $t$ use it for a lower bound for $f$ on $\{x\in X : |f(x)|\geq t\}$. This is definitely less or equal $\int_X f$. take the graph of a nice function $f$, draw a horizontal line at height $t$, and take the area underneath the line and under the graph.