Why "bother" with a null hypothesis at all?

(note: this is a very basic probability question, so it is highly probable (heh) that it is a duplicate)

Every time I am trying to get into statistics (again), I am always lost at hypothesis testing.

My basic question is - why do we form a null hypothesis as a negation of what we want to prove in the first place, and only then do we prove or disprove the null hypothesis?

Why do we do it at all, instead of just proving the original hypothesis?


Solution 1:

You seem confused. Statistical methods do not set out to prove; one either rejects or fails to reject a hypothesis. This wording is Very Importantâ„¢. (See also this and this.)

I'll use the classical judicial analogy. The accused (hypothesis) standing before the judge can be taken as "guilty" or "not guilty" by the judge (hypothesis test). Even with this, we can't totally eliminate the possibility of committing a Type I (innocent goes to jail) or Type II (guilty goes free) error. For all we know, even with all the evidence considered by the prosecution, defense, and jury, there might be a few confounding factors that weren't seen at the time. (Think of all the cases whose verdicts got changed when DNA tests became vogue.)

Put another way, using the word "accept" misleads some people. Here, it means that we're accepting the possibility that it's true, not that it is certainly true.

Solution 2:

I realise this has already been answered very well, but actually the point you raise, whether out of confusion or not, is very valid. In fact there has been controversy in the past raised by the "obsessive" focus on the rejection of the null hypothesis.

For example, in "The Fallacy of the Null-Hypothesis Significance Test" by Rozeboom (1960), the following conclusion is drawn:

The traditional null-hypothesis significance-test method ... is here vigorously excoriated for its inappropriateness as a method of inference. While a number of serious objections to the method are raised, its most basic error lies in mistaking the aim of a scientific investigation to be a decision, rather than a cognitive evaluation of propositions. It is further argued that the proper application of statistics to scientific inference is irrevocably committed to extensive consideration of inverse probabilities, and to further this end, certain suggestions are offered, both for the development of statistical theory and for more illuminating application of statistical analysis to empirical data.

Furthermore, in "Consequences of Prejudice Against the Null Hypothesis" by Greenwald (1975), the following conclusion is given,

Accordingly, it is concluded that research traditions and customs of discrimination against accepting the null hypothesis may be very detrimental to research progress.

More recently an alternative method was given in "An Alternative to Null-Hypothesis Significance Tests" (Killeen 2005)

The statistic $P_{rep}$ estimates the probability of replicating an effect. It captures traditional publication criteria for signal-to-noise ratio, while avoiding parametric inference and the resulting Bayesian dilemma. In concert with effect size and replication intervals, $P_{rep}$ provides all of the information now used in evaluating research, while avoiding many of the pitfalls of traditional statistical inference.

Solution 3:

A very basic explanation is that having only one hypothesis around will tell you nothing. You can get pretty numbers out of statistical analysis, but without something to compare them to, you'll be none the wiser.

As soon as you have two (or more) hypotheses to pit against each other, the analysis can begin to tell you something about how large a leap of faith it will require to conclude that the observed data are caused by this hypothesis rather than that. Thus, the null hypothesis: something neutral and boring that you can compare other hypotheses to.

In order to find out which answer is the best, you need to have more than one possible answer you consider. Otherwise you can't even speak of whether this answer is better or worse than that one.

Solution 4:

The part where you get lost is probably where you are thinking in Bayesian terms but you are trying to learn frequentist methods.

Solution 5:

To prove the original hypothesis, you want to be sure the results have been obtained because of your alternative hypothesis, rather than by chance. The burden of proof has to be on the alternative hypothesis. If I believe fairies are real, but you cannot see them, then this is consistent with the world we see. The null hypothesis is that they are not real. I need to therefore come up with some seriously good evidence to reject that null hypothesis, before we can scientifically say that fairies are probably real.