Page 99 - 4660
P. 99

2.3. Hypothesis testing




                   So far we have concentrated on using a data sample to obtain a number or a set of numbers.
               These numbers may be estimated values for the moments or central moments of the population
               from which the sample was drawn or, more generally, the values of some parameters a in an
               assumed model for the data. Sometimes, however, one wishes to use the data to give a ’yes’ or
               ’no’ answer to a particular question. For example, one might wish to know whether some
               assumed model does, in fact, provide a good fit to the data, or whether two parameters have the
               same value.




                     Simple and composite hypotheses

               In order to use data to answer questions of this sort, the question must be posed precisely. This is
               done by first asserting that some hypothesis is true. The hypothesis under consideration is
               traditionally called the null hypothesis and is denoted by H 0 . In particular, this usually specifies
               some form P(x|H 0 ) for the probability density function from which the data x are drawn. If the
               hypothesis determines the PDF uniquely, then it is said to be a simple hypothesis. If, however,
               the hypothesis determines the functional form of the PDF but not the values of certain
               parameters a on which it depends then it is called a composite hypothesis.
                   One decides whether to accept or reject the null hypothesis H0 by performing some
               statistical test, as described below in subsection. In fact, formally one uses a statistical test to
               decide between the null hypothesis H 0 and the alternative hypothesis H 1 . We define the latter to
                                    ¯
               be the complement H 0 of the null hypothesis within some restricted hypothesis space known (or
               assumed) in advance. Hence, rejection of H 0 implies acceptance of H 1 , and vice versa.
                   As an example, let us consider the case in which a sample x is drawn from a Gaussian
                                                      2
               distribution with a known variance σ but with an unknown mean µ. If one adopts the null
               hypothesis H 0 that µ = 0 which we write as H 0 : µ = 0, then the corresponding alternative
               hypothesis must be H 1 : µ ̸= 0. Note that, in this case, H 0 is a simple hypothesis whereas H 1 is a
               composite hypothesis. If, however, one adopted the null hypothesis H 0 :        µ < 0 then the
               alternative hypothesis would be H 1 : µ ≥ 0, so that both H 0 and H 1 would be composite
               hypotheses. Very occasionally both H 0 and H 1 will be simple hypotheses. In our illustration, this
               would occur, for example, if one knew in advance that the mean µ of the Gaussian distribution
               were equal to either zero or unity. In this case, if one adopted the null hypothesis H 0 : µ = 0 then
               the alternative hypothesis would be H 1 : µ = 1.




                     Statistical tests

               In our discussion of hypothesis testing we will restrict our attention to cases in which the null
               hypothesis H 0 is simple (see above). We begin by constructing a test statistic t(x) from the data
               sample. Although, in general, the test statistic need not be just a (scalar) number, and could be a
               multi-dimensional (vector) quantity, we will restrict our attention to the former case. Like any
               statistic, t(x) will be a random variable.     Moreover, given the simple null hypothesis H 0
               concerning the PDF from which the sample was drawn, we may determine (in principle) the
               sampling distribution P(t|H 0 ) of the test statistic.  A typical example of such a sampling
               distribution is shown in Fig. 3.1.
                   One defines for t a rejection region containing some fraction α of the total probability. For
               example, the (one-tailed) rejection region could consist of values of t greater than some value t crit ,


                                                              99
   94   95   96   97   98   99   100   101   102   103   104