<< Chapter < Page | Chapter >> Page > |
Statistical analysis is fundamentally an inversion process. The objective is to the "causes"--parameters of the probabilistic data generationmodel--from the "effects"--observations. This can be seen in our interpretation of the likelihood function.
Given a parameter , observations are generated according to The likelihood function has the same form as the conditional density function above except now is given (we take measurements) and is the variable. The likelihood function essentially inverts the role of observation(effect) and parameter (cause).
Unfortunately, the likelihood function does not provide a formal framework for the desired inversion.
One problem is that the parameter is supposed to be a fixed and deterministic quantity while the observation is the realization of a random process. So their role aren't really interchangeable in thissetting.
Moreover, while it is tempting to interpret the likelihood as a density function for , this is not always possible; for example, often
Another problematic issue is the mathematical formalization of statements like: "Based on the measurements , I am 95% confident that falls in a certain range."
Suppose you toss a coin 10 times and each time it comes up "heads." It might be reasonable to say that we are99% sure that the coin is unfair, biased towards heads.
Formally: which is the binomial likelihood. The problem with this is that implies that is a random , not deterministic, quantity. So, while "confidence" statements are very reasonable and in fact a normal part of "everyday thinking," this idea can not besupported from the classical perspective.
All of these "deficiencies" can be circumvented by a change in how we view the parameter .
If we view as the realization of a random variable with density , then Bayes Rule (Bayes, 1763) shows that Thus, from this perspective we obtain a well-defined inversion: Given , the parameter is distributing according to .
From here, confidence measures such as are perfectly legitimate quantities to ask for.
The prior distriubtion (or prior for short) models the uncertainty in the parameter. More specifically, models our knowledge--or lack thereof--prior to collecting data.
Notice that since the data are known , is just a constant. Hence, is proportional to the likelihood function multiplied by the prior.
Bayesian analysis has some significant advantages over classical statistical analysis:
Notification Switch
Would you like to follow the 'Statistical signal processing' conversation and receive update notifications?