Suppose that f is a differentiable real function in

Click For Summary

Homework Help Overview

The discussion revolves around proving that the derivative of a differentiable real function at a local maximum is zero. The function is defined in an open set of ℝn, and participants are exploring the implications of this definition in the context of local maxima.

Discussion Character

  • Conceptual clarification, Mathematical reasoning, Assumption checking

Approaches and Questions Raised

  • Participants discuss the relationship between the definition of differentiability and the condition of a local maximum. There are attempts to connect the concept of a local maximum to the derivative being zero, with some participants suggesting the use of the fundamental theorem of calculus to illustrate this relationship.

Discussion Status

There are multiple interpretations being explored regarding the proof. Some participants have offered hints and guidance on how to approach the problem, particularly focusing on the implications of the local maximum and the definition of the derivative. However, there is no explicit consensus on the method to be used.

Contextual Notes

Some participants note that the problem statement is framed in higher dimensions, while certain suggestions seem to apply only to one-dimensional cases. This raises questions about the applicability of specific methods across different dimensions.

Jamin2112
Messages
973
Reaction score
12

Homework Statement



Suppose that f is a differentiable real function in an open set E (which is a subset of) ℝn, and that that f has a local maximum at a point x in E. Prove that f'(x)=0.

Homework Equations



Definition. Suppose E is an open set in ℝn, f maps E into ℝm, and x is an element of E. If there exists a linear transformation A of ℝn into ℝm such that

limh0 |f(x+h)-f(x)-Ah|/|h|=0,

then f is differentiable at x, and we write

f'(x)=A.​

The Attempt at a Solution



f has a local max at x, so there exists a δ>0 such that f(y)≤f(y) for all y at which |y-x|<δ. Let h=x+y. Whenever |h|≤δ we have f(x)≥f(x+h) or f(x)≥f(x-h).

Now I need to use some algebraic cut-and-paste wizardry to end up with |f(x+h)-f(x)|/|h|. Could I please get an inconspicuous suggestion from the peanut gallery here?
 
Physics news on Phys.org
your problem statement includes f'(x)=0 your relavent equation has a similar form of f'(x)=A, if you show that A = 0 at a local maxima then you should have your answer. Hint: Make sure that you understand the physical meaning of a local maxima.
 
jake929 said:
your problem statement includes f'(x)=0 your relavent equation has a similar form of f'(x)=A, if you show that A = 0 at a local maxima then you should have your answer. Hint: Make sure that you understand the physical meaning of a local maxima.

Did not the first sentence of my attempt capture the concept of a local maxima?
 
I will paraphrase your problem statement and maybe that will help you understand what I mean. The problem is asking you to prove that the slope of a line at a local maxima is 0 using the fundamental theorem of calculus. This is something that will be assumed for later proofs. If you look at the fundamental theorem it will show an equation that looks like what you have seen in pre-calculus classes for calculating the slope (y2-y1)/(x2-x1) = slope. Try and show that the fundamental theorem yields a slope of zero at the maxima. Post your work if you want further help.
 
jake929, what you are suggesting is valid only in one dimension.
 
Yes, and even though the problem statement is in two dimensions the definition is in one
 

Similar threads

  • · Replies 5 ·
Replies
5
Views
2K
  • · Replies 8 ·
Replies
8
Views
1K
  • · Replies 11 ·
Replies
11
Views
3K
  • · Replies 18 ·
Replies
18
Views
3K
  • · Replies 10 ·
Replies
10
Views
3K
Replies
20
Views
5K
  • · Replies 17 ·
Replies
17
Views
4K
Replies
2
Views
2K
Replies
30
Views
3K
  • · Replies 3 ·
Replies
3
Views
2K