<< Chapter < Page | Chapter >> Page > |
In this set of notes, we begin our foray into learning theory. Apart from being interesting and enlightening in its own right, this discussion will also help ushone our intuitions and derive rules of thumb about how to best apply learning algorithms in different settings. We will also seek to answer a few questions: First, can wemake formal the bias/variance tradeoff that was just discussed? The will also eventually lead us to talk about model selection methods, which can,for instance, automatically decide what order polynomial to fit to a training set. Second, in machine learning it's really generalization errorthat we care about, but most learning algorithms fit their models to the training set. Why should doing well on the training set tell us anythingabout generalization error? Specifically, can we relate error on the training set to generalization error? Third and finally, are there conditions underwhich we can actually prove that learning algorithms will work well?
We start with two simple but very useful lemmas.
Lemma. (The union bound). Let be different events (that may not be independent). Then
In probability theory, the union bound is usually stated as an axiom (and thus we won't try to prove it), but it also makes intuitive sense: The probability ofany one of events happening is at most the sums of the probabilities of the different events.
Lemma. (Hoeffding inequality) Let be independent and identically distributed (iid) random variables drawn from a Bernoulli( ) distribution. I.e., , and . Let be the mean of these random variables, and let any be fixed. Then
This lemma (which in learning theory is also called the Chernoff bound ) says that if we take —the average of Bernoulli( ) random variables—to be our estimate of , then the probability of our being far from the true value is small, so long as is large. Another way of saying this is that if you have a biased coin whosechance of landing on heads is , then if you toss it times and calculate the fraction of times that it came up heads, that will be agood estimate of with high probability (if is large).
Using just these two lemmas, we will be able to prove some of the deepest and most important results in learning theory.
To simplify our exposition, let's restrict our attention to binary classification in which the labels are . Everything we'll say here generalizes to other, including regression and multi-classclassification, problems.
We assume we are given a training set of size , where the training examples are drawn iid from some probability distribution . For a hypothesis , we define the training error (also called the empirical risk or empirical error in learning theory) to be
This is just the fraction of training examples that misclassifies. When we want to make explicit the dependence of on the training set , we may also write this a . We also define the generalization error to be
Notification Switch
Would you like to follow the 'Machine learning' conversation and receive update notifications?