<< Chapter < Page | Chapter >> Page > |
Usually, the likelihood is manipulated to derive a sufficient statistic. The resulting sequential decision rule is Note that the thresholds and , derived from the thresholds and , usually depend on the number of observations used in the decision rule.
Let be a Gaussian random vector as in our previous examples withstatistically independent components. The mean vector is assumed for simplicity to consist of equal positive values: , . Using the previous derivations, our sequential test becomes Starting with , we gather the data and compute the sum. The sufficient statistic will lie in the middle range betweenthe two thresholds until one of them is exceeded as shown in . The model evaluation procedure then terminates and the chosen model announced. Note how the thresholds depend onthe amount of data available (as expressed by ). This variation typifies the sequential hypothesis tests.
The awake reader might wonder whether that the sequential likelihood ratio test just derived has the disturbing propertythat it may never terminate: can the likelihood ratio wander between the two thresholds forever? Fortunately, thesequential likelihood ratio test has been shown to terminate with probability one ( Wald ). Confident of eventual termination, we need to explore how manyobservations are required to meet performance specifications. The number of observations is variable, depending on theobserved data and the stringency of the specifications. The average number of observations required can be determined in the interesting case when theobservations are statistically independent.
Assuming that the observations are statistically independent and identically distributed, the likelihood ratio is equal tothe product of the likelihood ratios evaluated at each observation. Considering , the logarithm of the likelihood ratio when a decision is made on observation , we have where is the likelihood ratio corresponding to the observation. We seek an expression for , the expected value of the number of observations required to make the decision. To derive this quantity, weevaluate the expected value of the likelihood ratio when the decision is made. This value will usually vary with whichmodel is actually valid; we must consider both models separately. Using the laws of conditional expectation (see Joint Distributions ), we find that the expected value of , assuming that model was true, is given by The outer expected value is evaluated with respect to the probability distribution of ; the inner expected value is average value of the log-likelihood assuming that observations were required to choose model . In the latter case, the log-likelihood is the sum of component log-likelihood ratios Noting that the expected value on the right is a constant with respect to the outer expected value, we find that The average number of observations required to make a decision, correct or incorrect, assuming that is true is thus expressed by Assuming that the other model was true, we have the complementary result
The numerator is difficult to calculate exactly but easily approximated; assuming that the likelihood ratio equals itsthreshold value when the decision is made, Note these expressions are not problem dependent; they depend only on the specified probabilities.The denominator cannot be approximated in a similar way with such generality; it must be evaluated for each problem.
In the Gaussian example we have been exploring, the log-likelihood of each component observation is given by The conditional expected values required to evaluate the expression for the average number of required observationsare For simplicity, let's assume that the false-alarm and detection probabilities are symmetric (i.e. ). The expressions for the average number of observations are equal foreach model and we have where is a function equal to . Thus, the number of observations decreases with increasing signal-to-noise ratio and increases as the false-alarm probability is reduced.
Suppose we used a likelihood ratio test where all data were considered once and a decision made; how many observationswould be required to achieve a specified level of performance and how would this fixed number compare with theaverage number of observations in a sequential test? In this example, we find from our earlier calculations (see equation ) that so that The duration of the sequential and block tests depend on the signal-to-noise ratio in the same way;however, the dependence on the false-alarm probability is quite different. As depicted in the , the disparity between these quantities increases rapidly as the false alarm probabilitydecreases, with the sequential test requiring correspondingly fewer observations on the average .
We must not forget that these results apply to the average number of observations required to make a decision. Expressions for the distribution of thenumber of observations are complicated and depend heavily on the problem. When an extremely large number of observationare required to resolve a difficult case to the required accuracy, we are forced to truncate the sequential test, stopping when a specified number ofobservations have been used. A decision would then be made by dividing the region between the boundaries in half andselecting the model corresponding to the boundary nearest to the sufficient statistic. If this truncation point is largerthan the expected number, the performance probabilities will change little. "Larger" is again problem dependent;analytic results are few, leaving the option of computer simulations to estimate the distribution of the number ofobservations required for a decision.
Notification Switch
Would you like to follow the 'Statistical signal processing' conversation and receive update notifications?