Fermat's theorem applied to multivariate functions

Click For Summary

Discussion Overview

The discussion revolves around the application of Fermat's theorem to multivariate functions, specifically exploring whether a similar principle exists for functions of two or more variables. Participants delve into the conditions for local maxima and minima in multivariable calculus, the implications of partial derivatives, and the relationship between differentiability and continuity.

Discussion Character

  • Exploratory
  • Technical explanation
  • Conceptual clarification
  • Debate/contested
  • Mathematical reasoning

Main Points Raised

  • Some participants propose that if a function f(x,y) has a local minimum at (0,0), then a related function g(t) defined by g(t) = f(at, bt) can be used to analyze the local behavior.
  • It is noted that for a differentiable function of two variables, the conditions for local extrema require both partial derivatives to equal zero at that point.
  • One participant expresses confusion about the implications of the existence of partial derivatives and questions why this does not guarantee differentiability, suggesting that a function can have partial derivatives without being differentiable.
  • A later reply introduces an analogy involving a function that is 1 at rational points and follows a different rule at irrational points, illustrating how partial derivatives can exist while the function is not differentiable overall.
  • Another participant seeks clarification on how to test for differentiability in multivariable functions, given the limited perspective of partial derivatives.
  • Discussion includes an example of a continuous function f(x,y) = x² + y², with participants exploring the proof of its continuity and its implications for differentiability.

Areas of Agreement / Disagreement

Participants express varying levels of understanding regarding the relationship between partial derivatives, continuity, and differentiability. There is no consensus on the implications of these concepts, and multiple viewpoints are presented without resolution.

Contextual Notes

Limitations include the dependence on definitions of differentiability and continuity, as well as the potential for misunderstanding the implications of partial derivatives in higher dimensions. The discussion does not resolve the mathematical nuances involved.

elementbrdr
Messages
43
Reaction score
0
Fermat's theorem provides that, if a function f(x) has a local max or min at a, and if f'(a) exists, then f'(a)=0. I was wondering whether a similar theory exists for a function f(x,y) or f(x,y,z) etc.
 
Physics news on Phys.org
Yep. You can even work it out yourself, using the fact that, e.g., if (0,0) is a local minimum of f(x,y), then 0 is a local minimum of g(t), where g(t) is defined by
g(t) = f(at, bt)​
 
Thanks for the response, Hurkyl. I'm not following you as to how I could use that to solve for a pair of (x,y), though. Do you have an example?
 
If f is a differentiable function of two variables, x and y, then at any max or min we must have
[itex]\frac{\partial f}{\partial x}= 0[/itex] and [itex]\frac{\partial f}{\partial y}= 0[/itex]
at that point.

By the way, the existence of the partial derivatives at a given point does not always imply that f itself is differetiable there. Better is the statement
[tex]\text{\grad }f= \nabla f= \vec{0}[/itex][/tex]
 
HallsofIvy, your last statement is interesting. I'm not sure if I'm at the point where I can understand it quite yet, unfortunately.

I have not yet begun to study multivariate calculus. I am currently reviewing single variable calculus in preparation for linear algebra. I reviewed Fermat's Theorem yesterday and recalled that I had encountered a problem where applying it in the multivariate context would have been helpful. So I was primarily interested in whether my intuition that such application was possible was conceptually sound.

I'm curious, though, why the existence of a partial derivative with respect to each variable does not imply that the function is differentiable. I thought a whole derivative was either (a) an ordered set of partial derivative values or (b) the vectors sum of the partial derivatives. So if you can calculate partial derivatives, how could the function not be differentiable? (I could be way off here, but figured it wouldn't hurt to ask)
 
elementbrdr said:
HallsofIvy, your last statement is interesting. I'm not sure if I'm at the point where I can understand it quite yet, unfortunately.

I have not yet begun to study multivariate calculus. I am currently reviewing single variable calculus in preparation for linear algebra. I reviewed Fermat's Theorem yesterday and recalled that I had encountered a problem where applying it in the multivariate context would have been helpful. So I was primarily interested in whether my intuition that such application was possible was conceptually sound.

I'm curious, though, why the existence of a partial derivative with respect to each variable does not imply that the function is differentiable. I thought a whole derivative was either (a) an ordered set of partial derivative values or (b) the vectors sum of the partial derivatives. So if you can calculate partial derivatives, how could the function not be differentiable? (I could be way off here, but figured it wouldn't hurt to ask)

Think of the following one-dimensional analogy:

A function f(x) is 1 at all rational points (set Q), but f(x)=x for all irrational points (set I).

Now, if you constrain your limiting procedure onto (set Q), you will find that the "partial" derivative on THAT set equals 0, at all points in Q.

But on (set I), you'll get the "partial" derivative equal to 1, at all points in I.Thus, at every point, you may say that a "partial" derivative exists, but the function is NOT differentiable on the reals as such...What is required of a differentiable function is that along ALL and EVERY path converging to some point, the differentiation procedure must give the same answer.
That is a much stricter requirement than the definition of the partial derivative requires, and thus, all partial derivatives may exist, even though the function isn't differentiable.
 
That makes sense. But f(x) is not continuous, so it not be differentiable, right? If you expand your analogy to two dimensions using a function f(x,y), with x and y taking the place of Q and I, respectively, then wouldn't you have f'(x) = (d/dx, d/dy) = (0,1) at all points? I'm having trouble visualizing how you could obtain a complete set of partial derivatives for all variables of a function but fail to have a differentiable function. Again, I haven't taken multivariate, so I'm just trying to apply single variable calc concepts here...
 
Arildno, I received your response by email. I don't see it on the forum yet, though. What you say makes a lot of sense. If I understand you correctly, you are saying that, for a function f(x,y), the existence of the partial derivatives d/dx and d/dy only represent 4 possible approaches to a given point f(a,b) out of an infinite number of possible approaches (not sure if I'm using proper terminology). But if that is correct, then how can one test whether a function f(x,y) is actually differentiable?
 
elementbrdr said:
Arildno, I received your response by email. I don't see it on the forum yet, though. What you say makes a lot of sense. If I understand you correctly, you are saying that, for a function f(x,y), the existence of the partial derivatives d/dx and d/dy only represent 4 possible approaches to a given point f(a,b) out of an infinite number of possible approaches (not sure if I'm using proper terminology). But if that is correct, then how can one test whether a function f(x,y) is actually differentiable?
Strange.
Don't know how the e-mail was activated??
 
  • #10
Did you post and then delete? I have instant email notification activated for responses to my threads. Here's what I received in my inbox:

1. As to continuity:
f is continuous, when restricting our limiting process upon the subset specified.
That is why it can be "partially" differentiable there!

2. However, neither of the two subsets I specify are connected sets (they consist, so to speak, of isolated points), something that is peculiar for the one-variable case, but not for higher dimensional cases (the x-axis is a connected set, and so is the y-axis).


3. Well, [tex]\frac{ \partial f }{ \partial x } \frac{ \partial f}{ \partial x}[/tex] constrains us to only look at how f behaves along the x-axis, (or along an axis parallell to that).
It does NOT take into account how we may approach a point further along the x-axis by LEAVING the x-axis, and then rejoin it at some other point.
And it is precisely this restricted perspective of the partial derivative(s) that make it possible that all partial derivatives exists, even though the function remains non-differentiable at some point.
 
Last edited:
  • #11
That's probably it.

Anyhow.

What is the truly essential idea about "continuity"?

It is that if you are "close enough" in your argument space of values, you'll be "close enough" in your space of function values.

But, if your argument space is not a line, but a plane, being "close enough" to some point is to be within some circle of sufficiently tiny radius to that point.
Is that clear?

Take the function [tex]f(x,y)=x^{2}+y^{2}[/tex]
How can we prove that this is continuous everywhere?

Well, if we pick a point [tex](x_[0},y_{0}[/tex], any other point in the plane can be represented as [tex]x_{0}+r\cos\theta,y_{0}+r\sin\theta[/tex].

Now, we look at the difference between the assumed limit [tex]L=x_{0}^{2}+y_{0}^{2}[/tex] (i.e, what it needs to be if continuous), and the general expression of f:

[tex]|(x_{0}+r\cos\theta)^{2}+(y_{0}+r\sin\theta)^{2}-x_{0}^{2}-y_{0}^{2}|=r|2x_{0}\cos\theta+2y_{0}\sin\theta+r|[/tex], an expression that:

Strictly vanishes when r goes to zero, wholly independent of the angle (which will be different for different paths towards our point)!

Thus, we have proven continuity of the function.Now, tell me if this is OK, and ikf it is, we can go on with the derivatives for higher-dimensional functions.
 
  • #12
I think I follow your explanation, though I doubt I could personally prove the continuity of other functions. But the basic idea in this case is that lim r->0 = x^2 + y^2. Conceptually this means that, for any point within the function's range is continuous with every other point within an arbitrarily large radius within a certain tolerance for error. And you could formalize this in the same way that a limit is precisely defined for single variable functions by showing a necessary relationship between the variance of (x,y) and f(x,y).
 

Similar threads

  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 6 ·
Replies
6
Views
4K
  • · Replies 4 ·
Replies
4
Views
2K
  • · Replies 3 ·
Replies
3
Views
4K
  • · Replies 6 ·
Replies
6
Views
3K
  • · Replies 9 ·
Replies
9
Views
3K
  • · Replies 9 ·
Replies
9
Views
4K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 3 ·
Replies
3
Views
3K
  • · Replies 21 ·
Replies
21
Views
6K