Fermat's theorem applied to multivariate functions

Click For Summary
Fermat's theorem can be extended to multivariate functions, where local maxima or minima occur at points where the partial derivatives equal zero. For a function f(x,y), if it has a local extremum, both partial derivatives must vanish at that point. However, the existence of these partial derivatives does not guarantee the function is differentiable at that point. An example illustrates that a function can have partial derivatives yet remain non-differentiable due to the nature of how limits are approached in higher dimensions. Understanding continuity in multivariate functions involves ensuring that values remain close as inputs approach a point from any direction, which is a stricter requirement than for single-variable functions.
elementbrdr
Messages
43
Reaction score
0
Fermat's theorem provides that, if a function f(x) has a local max or min at a, and if f'(a) exists, then f'(a)=0. I was wondering whether a similar theory exists for a function f(x,y) or f(x,y,z) etc.
 
Physics news on Phys.org
Yep. You can even work it out yourself, using the fact that, e.g., if (0,0) is a local minimum of f(x,y), then 0 is a local minimum of g(t), where g(t) is defined by
g(t) = f(at, bt)​
 
Thanks for the response, Hurkyl. I'm not following you as to how I could use that to solve for a pair of (x,y), though. Do you have an example?
 
If f is a differentiable function of two variables, x and y, then at any max or min we must have
\frac{\partial f}{\partial x}= 0 and \frac{\partial f}{\partial y}= 0
at that point.

By the way, the existence of the partial derivatives at a given point does not always imply that f itself is differetiable there. Better is the statement
\text{\grad }f= \nabla f= \vec{0}[/itex]
 
HallsofIvy, your last statement is interesting. I'm not sure if I'm at the point where I can understand it quite yet, unfortunately.

I have not yet begun to study multivariate calculus. I am currently reviewing single variable calculus in preparation for linear algebra. I reviewed Fermat's Theorem yesterday and recalled that I had encountered a problem where applying it in the multivariate context would have been helpful. So I was primarily interested in whether my intuition that such application was possible was conceptually sound.

I'm curious, though, why the existence of a partial derivative with respect to each variable does not imply that the function is differentiable. I thought a whole derivative was either (a) an ordered set of partial derivative values or (b) the vectors sum of the partial derivatives. So if you can calculate partial derivatives, how could the function not be differentiable? (I could be way off here, but figured it wouldn't hurt to ask)
 
elementbrdr said:
HallsofIvy, your last statement is interesting. I'm not sure if I'm at the point where I can understand it quite yet, unfortunately.

I have not yet begun to study multivariate calculus. I am currently reviewing single variable calculus in preparation for linear algebra. I reviewed Fermat's Theorem yesterday and recalled that I had encountered a problem where applying it in the multivariate context would have been helpful. So I was primarily interested in whether my intuition that such application was possible was conceptually sound.

I'm curious, though, why the existence of a partial derivative with respect to each variable does not imply that the function is differentiable. I thought a whole derivative was either (a) an ordered set of partial derivative values or (b) the vectors sum of the partial derivatives. So if you can calculate partial derivatives, how could the function not be differentiable? (I could be way off here, but figured it wouldn't hurt to ask)

Think of the following one-dimensional analogy:

A function f(x) is 1 at all rational points (set Q), but f(x)=x for all irrational points (set I).

Now, if you constrain your limiting procedure onto (set Q), you will find that the "partial" derivative on THAT set equals 0, at all points in Q.

But on (set I), you'll get the "partial" derivative equal to 1, at all points in I.Thus, at every point, you may say that a "partial" derivative exists, but the function is NOT differentiable on the reals as such...What is required of a differentiable function is that along ALL and EVERY path converging to some point, the differentiation procedure must give the same answer.
That is a much stricter requirement than the definition of the partial derivative requires, and thus, all partial derivatives may exist, even though the function isn't differentiable.
 
That makes sense. But f(x) is not continuous, so it not be differentiable, right? If you expand your analogy to two dimensions using a function f(x,y), with x and y taking the place of Q and I, respectively, then wouldn't you have f'(x) = (d/dx, d/dy) = (0,1) at all points? I'm having trouble visualizing how you could obtain a complete set of partial derivatives for all variables of a function but fail to have a differentiable function. Again, I haven't taken multivariate, so I'm just trying to apply single variable calc concepts here...
 
Arildno, I received your response by email. I don't see it on the forum yet, though. What you say makes a lot of sense. If I understand you correctly, you are saying that, for a function f(x,y), the existence of the partial derivatives d/dx and d/dy only represent 4 possible approaches to a given point f(a,b) out of an infinite number of possible approaches (not sure if I'm using proper terminology). But if that is correct, then how can one test whether a function f(x,y) is actually differentiable?
 
elementbrdr said:
Arildno, I received your response by email. I don't see it on the forum yet, though. What you say makes a lot of sense. If I understand you correctly, you are saying that, for a function f(x,y), the existence of the partial derivatives d/dx and d/dy only represent 4 possible approaches to a given point f(a,b) out of an infinite number of possible approaches (not sure if I'm using proper terminology). But if that is correct, then how can one test whether a function f(x,y) is actually differentiable?
Strange.
Don't know how the e-mail was activated??
 
  • #10
Did you post and then delete? I have instant email notification activated for responses to my threads. Here's what I received in my inbox:

1. As to continuity:
f is continuous, when restricting our limiting process upon the subset specified.
That is why it can be "partially" differentiable there!

2. However, neither of the two subsets I specify are connected sets (they consist, so to speak, of isolated points), something that is peculiar for the one-variable case, but not for higher dimensional cases (the x-axis is a connected set, and so is the y-axis).


3. Well, \frac{ \partial f }{ \partial x } \frac{ \partial f}{ \partial x} constrains us to only look at how f behaves along the x-axis, (or along an axis parallell to that).
It does NOT take into account how we may approach a point further along the x-axis by LEAVING the x-axis, and then rejoin it at some other point.
And it is precisely this restricted perspective of the partial derivative(s) that make it possible that all partial derivatives exists, even though the function remains non-differentiable at some point.
 
Last edited:
  • #11
That's probably it.

Anyhow.

What is the truly essential idea about "continuity"?

It is that if you are "close enough" in your argument space of values, you'll be "close enough" in your space of function values.

But, if your argument space is not a line, but a plane, being "close enough" to some point is to be within some circle of sufficiently tiny radius to that point.
Is that clear?

Take the function f(x,y)=x^{2}+y^{2}
How can we prove that this is continuous everywhere?

Well, if we pick a point (x_[0},y_{0}, any other point in the plane can be represented as x_{0}+r\cos\theta,y_{0}+r\sin\theta.

Now, we look at the difference between the assumed limit L=x_{0}^{2}+y_{0}^{2} (i.e, what it needs to be if continuous), and the general expression of f:

|(x_{0}+r\cos\theta)^{2}+(y_{0}+r\sin\theta)^{2}-x_{0}^{2}-y_{0}^{2}|=r|2x_{0}\cos\theta+2y_{0}\sin\theta+r|, an expression that:

Strictly vanishes when r goes to zero, wholly independent of the angle (which will be different for different paths towards our point)!

Thus, we have proven continuity of the function.Now, tell me if this is OK, and ikf it is, we can go on with the derivatives for higher-dimensional functions.
 
  • #12
I think I follow your explanation, though I doubt I could personally prove the continuity of other functions. But the basic idea in this case is that lim r->0 = x^2 + y^2. Conceptually this means that, for any point within the function's range is continuous with every other point within an arbitrarily large radius within a certain tolerance for error. And you could formalize this in the same way that a limit is precisely defined for single variable functions by showing a necessary relationship between the variance of (x,y) and f(x,y).
 

Similar threads

  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 6 ·
Replies
6
Views
2K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 4 ·
Replies
4
Views
1K
  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 3 ·
Replies
3
Views
2K
  • · Replies 21 ·
Replies
21
Views
4K