Q about 2nd derivative test for multivariable functions

Click For Summary

Homework Help Overview

The discussion revolves around the second derivative test for multivariable functions, specifically focusing on the Hessian matrix and its determinant (D). Participants explore the implications of the conditions D > 0, D < 0, and the behavior of the second partial derivative fxx, particularly in relation to identifying local minima, maxima, and saddle points.

Discussion Character

  • Conceptual clarification, Assumption checking

Approaches and Questions Raised

  • Participants question the validity of the original poster's statements regarding the implications of D > 0 and fxx = 0. There is exploration of how these conditions apply differently in the context of functions of two variables versus multivariable functions.

Discussion Status

The discussion is ongoing, with participants providing clarifications and corrections regarding the conditions under which the second derivative test is applicable. Some participants have offered insights into the differences between two-variable and multivariable cases, indicating a productive exchange of ideas without reaching a consensus.

Contextual Notes

There is a noted distinction between the treatment of functions of two variables and those of three or more variables, with some participants emphasizing the need for precise terminology and definitions in the context of the Hessian matrix and its properties.

kostoglotov
Messages
231
Reaction score
6

Homework Statement



So the test is to take the determinant (D) of the Hessian matrix of your multivar function.

Then if D>0 & fxx>0 it's a min point, if D>0 & fxx<0 it's a max point.

For D<0 it's a saddle point, and D=0 gives no information.

My question is, what happens if fxx=0? Is that equivalent to D=0?
 
Physics news on Phys.org
Your first statement isn't true. D>0 doesn't tell much of anything about the type of extremum in the general multivariate case. What is the complete statement of your problem ?
 
Last edited:
wabbit said:
Your first statement isn't true. D>0 doesn't tell much of anything about the type of extremum in the general multivariate case. What is the complete statement of your problem ?

It's not my statement, it's the textbooks...there's no specific problem to which I can refer. I was just wondering what it tells us if fxx=0.

In the textbook (Stewart's Calculus 6 edt. p960 ) it says

Suppose the second partial derivatives of f are continuous on a disk with center (a,b) and suppose that fx(a,b) = 0 and fy(a,b) = 0 [that is, (a,b) is a critical point of f], Let

D = D(a,b) = fxx(a,b)fyy(a,b) - [fxy(a,b)]^2

(a) If D > 0 and fxx > 0, then f(a,b) is a local minimum

(b) If D > 0 and fxx < 0, then f(a,b) is a local maximum

(c) If D < 0 then f(a,b) is not a local min or max, it is a saddle point
 
kostoglotov said:
It's not my statement, it's the textbooks...there's no specific problem to which I can refer. I was just wondering what it tells us if fxx=0.

In the textbook (Stewart's Calculus 6 edt. p960 ) it says

Suppose the second partial derivatives of f are continuous on a disk with center (a,b) and suppose that fx(a,b) = 0 and fy(a,b) = 0 [that is, (a,b) is a critical point of f], Let

D = D(a,b) = fxx(a,b)fyy(a,b) - [fxy(a,b)]^2

(a) If D > 0 and fxx > 0, then f(a,b) is a local minimum

(b) If D > 0 and fxx < 0, then f(a,b) is a local maximum

(c) If D < 0 then f(a,b) is not a local min or max, it is a saddle point

You and Wabbit are both right. For the special case of 2 variables, ##f_{xx} > 0## and ##D > 0## guarantee a strict local min. However, for 3 or more variables, having ##D > 0## really does not give much useful information at all. For example, for ##f(x,y,z) = x^2-y^2-z^2##, the point (0,0,0) is stationary, and the Hessian is
H = \left[ \begin{array}{ccc}2 &amp;0 &amp; 0 \\ 0 &amp; -2 &amp; 0 \\ 0 &amp; 0 &amp; -2\end{array} \right]
which has ##f_{xx} > 0## and ##D > 0##. But (0,0,0) is a saddle point.

In general, it is necessary (not sufficient) that all ##f_{x_i x_i} > 0## in order to have a strict local min. For the 2-variable case, having ##f_{xx} > 0## and ##D > 0## automatically guarantees ##f_{yy} > 0## as well.
 
  • Like
Likes   Reactions: kostoglotov
@kostoglotov. Yes as explained by Ray your statement is correct in the bivariate case and then fxx=0 leads to D<0 or D=0, it only excludes the case D>0.

And no, it's not the textbook that is to blame.

Your statement is false because it says "multivar", The textbook is correct because it does no such thing and talks about functions of two variables.
 
  • Like
Likes   Reactions: kostoglotov
wabbit said:
@kostoglotov. Yes as explained by Ray your statement is correct in the bivariate case and then fxx=0 leads to D<0 or D=0, it only excludes the case D>0.

And no, it's not the textbook that is to blame.

Your statement is false because it says "multivar", The textbook is correct because it does no such thing and talks about functions of two variables.

Ah, ok, I'm being loose with my terminology. I'll take note of that! Thanks.
 
A little bit more detail: strictly speaking, "the derivative" of a multi-variable function is the gradient vector- the vector whose components, in a given coordinate system, are the partial derivatives of the function and the second derivative is the Hessian- the matrix having all second partial derivatives as components. Since f_{xy}= f_{yx}, this is a symmetric matrix and so is diagonalizable. That is, there exist a coordinate system in which the "mixed" second partials are 0.

Now, if the second derivatives at a critical point, in this coordinate system, are all positive the point is "relative minimum". If they are all negative a "relative maximum". If there are some positive and some negative second derivatives we have a "saddle point".

For two variables, it is fairly easy to see which we have without having to change coordinate system. If the matrix of second derivatives is
\begin{bmatrix} f_{xx} &amp; f_{xy} \\ f_{xy} &amp; f_{yy} \end{bmatrix}
then its determinant is f_{xx}f_{yy}- f_{xy}^2. Since the determinant is independent of coordinate changes, that would be the same as f_{x&#039;x&#039;}f_{y&#039;y&#039;} in a coordinate system in which the mixed partials were 0. So if that determinant were positive, f_{x&#039;x&#039;} and f_{y&#039;y&#039;} must have the same sign so the critical point is either a maximum or a minimum. We can determine which by looking at either f_{x&#039;x&#039;} or f_{y&#039;y&#039;} separately. If that determinant were negative then we would have a saddle point.

But with more than two coordinates, it is no longer that simple. Now, the determinant might be positive if there are an even number of negative second derivatives in that diagonalized matrix. For example, in three dimensions, if the determinant is positive, we might have all three second derivatives positive, a minimum, or we might have two negative and one positive, a saddle point.

The only way to be sure is to find the specific eigenvalues of the Hessian matrix since they will be the numbers on the diagonal matrix. If all eigenvalues are positive, we have a minimum, if all are negative, a maximum, and, if some are positive and some negative, a saddle point.
 
  • Like
Likes   Reactions: kostoglotov
kostoglotov said:
Ah, ok, I'm being loose with my terminology. I'll take note of that! Thanks.
In the n-variable case, the stationary point ##\vec{x}_0 = (x_{10}, x_{20}, \ldots, x_{n0})## is a strict local minimizing point of ##f(\vec{x})## if the Hessian matrix ##H_0 = H(\vec{x}_0)## is positive-definite. One test for this is that all the principle minors of ##H_0## be positive; that is, the ##1\times 1## upper-left determinant should be > 0---and of course, that is just the statement ##f_{x_1 x_1} > 0##; as well, the upper left ##2 \times 2## determinant should be >0, the upper-left ##3 \times 3## determinant should be >0, etc. In other words, the upper left submatrices having rows ##1\ldots i## and columns ##1\ldots i## should all have positive determinants for ##i = 1, 2, \ldots, n##. Note that for the case ##n = 2## this is exactly what your textbook says.

If the Hessian is "indefinite", the point ##\vec{x}_0## is a saddle point; if the Hessian is not indefinite but is only positive semi-definite, we cannot say: the point could be a minimum or a saddle point (but for sure it is not a maximum).
 
  • Like
Likes   Reactions: kostoglotov
That's fascinating. I'm doing the MIT linear algebra course in the second half of this year, I'll have to bookmark your post and come back to it when I understand changing of coordinate systems and eigen-stuff better! Thanks.
 

Similar threads

  • · Replies 12 ·
Replies
12
Views
2K
Replies
4
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
Replies
3
Views
14K
Replies
5
Views
3K
Replies
5
Views
3K
Replies
5
Views
7K
Replies
1
Views
2K
  • · Replies 11 ·
Replies
11
Views
1K
  • · Replies 9 ·
Replies
9
Views
3K