<< Chapter < Page | Chapter >> Page > |
And it can have other, most esoteric conditions like that because again, this is a condition that you can solve for the optimal margin, and then just by scaling, you have w up and down. You can – you can then ensure you meet this condition as well.
So again, [inaudible] one of these conditions right now, not all of them. And so our ability to choose any scaling condition on w that's convenient to us will be useful again in a second.
All right. So let's go ahead and break down the optimization problem. And again, my goal is to choose parameters w and b so as to maximize the geometric margin.
Here's my first attempt at writing down the optimization problem. Actually wrote this one down right at the end of the previous lecture. Begin to solve the parameters gamma w and b such that – that [inaudible] i for in training examples.
Let's say I choose to add this normalization condition. So the norm condition that w – the normal w is equal to 1 just makes the geometric and the functional margin the same. And so I'm saying I want to find a value – I want to find a value for gamma as big as possible so that all of my training examples have functional margin greater than or equals gamma, and with the constraint that normal w equals 1, functional margin and geometric margin are the same. So it's the same. Find the value for gamma so that all the values – all the geometric margins are greater or equal to gamma.
So you solve this optimization problem, then you have derived the optimal margin classifier – that there's not a very nice optimization problem because this is a nasty, nonconvex constraints. And [inaudible] is asking that you solve for parameters w that lie on the surface of a unisphere, lie on his [inaudible]. It lies on a unicircle – a unisphere.
And so if we can come up with a convex optimization problem, then we'd be guaranteed that our [inaudible] descend to other local [inaudible]will not have local optimal. And it turns out this is an example of a nonconvex constraint. This is a nasty constraint that I would like to get rid of.
So let's change the optimization problem one more time. Now, let me pose a slightly different optimization problem. Let me maximize the functional margin divided by the normal w subject to yi w transpose xi.
So in other words, once you find a number, gamma hat, so that every one of my training examples has functional margin greater than the gamma hat, and my optimization objective is I want to maximize gamma hat divided by the normal w. And so I wanna maximize the function margin divided by the normal w.
And we saw previously the function margin divided by the normal w is just a geometric margin, and so this is a different way of posing the same optimization problem.
[Inaudible] confused, though. Are there questions about this?
Student: [Inaudible] the second statement has to be made of the functional margin y divided by – why don't you just have it the geometric margin? Why do you [inaudible]?
Instructor (Andrew Ng) :[Inaudible] say it again?
Notification Switch
Would you like to follow the 'Machine learning' conversation and receive update notifications?