# Exploiting Directions of Negative Curvature

1. Sep 24, 2011

### brydustin

The title of an old paper..... It mentions that in order to use the full information of a hessian in 2nd order optimization that you should make a part of your iterative step to include v (eigenvector corresponding to smallest eigenvalue, assuming that the eigenvalue is negative).
By doing the following: p = -sign(g'*v)*v : where g is the gradient. So here is the question, what is the geometrical meaning of the dot product of {g,v}? Because the idea is to find a local minimum but I'm trying to find a local maximum and would like to use similar information. Another condition for a local minimum would be that all the eigenvalues are positive, so in my case I would want all of them to be negative. So in my case would I set
p = + or - sign(g'*w)*w, where w is the eigenvalue corresponding to the largest eigenvalue (assuming that its also greater than 0 -- obviously if max(eigenvalue) < 0 then hessian is sufficiently conditioned to find a maximizer. Anyway, I appreciate any help on this.... which sign do I pick and why (what's the geometry behind it?)
Thanks

2. Sep 24, 2011

### AlephZero

Finding the minimum of x is the same problem as finding the maximum of -x.

That should be all you need to answer your questions about signs.

3. Sep 24, 2011

### brydustin

Okay.... but that doesn't actually answer my question.
What is the geometrical meaning behind dot(gradient, eigenvector of smallest eigenvalue), simply saying to flip the signs always makes no sense. Maybe in my case, -sgn(dot(g,eigenvector))*eigenvector STILL makes sense because of the sign of the eigenvector, but I don't know. The crux of the question is about geometry and not a naive change of sign. You don't change the sign mindlessly, for example, when solving g +Hd=0 you don't suddenly say d = inv(H)*g. My question is one of geometry.