Undergrad Uniform convergence of difference quotient in higher dimensions

Click For Summary
The discussion focuses on the uniform convergence of the difference quotient for differentiable functions in higher dimensions. It establishes that for a differentiable function f: R^d → R^d, the limit of the difference quotient approaches zero uniformly on a compact subset K under certain conditions. The mean value theorem for vector-valued functions is referenced as a potential tool for proving this convergence in higher dimensions. A specific approach is suggested involving the construction of a new function F(t) to facilitate the proof. The conversation highlights the need for continuous differentiability of f to ensure uniform convergence.
psie
Messages
315
Reaction score
40
TL;DR
In ##\mathbb R## with ##f:[a,b]\to\mathbb R##, one can show that the difference quotient in the definition of the derivative of ##f## converges uniformly on ##[a,b]## to the derivative of ##f##. I wonder, is this true for ##\mathbb R^d## with ##d>1## too?
Let ##f:\mathbb R^d\to\mathbb R^d## and let ##f## be differentiable on an open subset ##U\subset\mathbb R^d##. We then have $$\lim_{h\to0}\frac{|f(h+a)-f(a)-Df(a)(h)|}{|h|}=0,\quad \forall a\in U.$$ ##Df(a)## is the linear transformation at ##a##, i.e. the derivative of ##f##, acting on ##h##. Let now ##K\subset U## be compact. Under what conditions on ##f## does the fraction above converge uniformly to ##0## on ##K##?

I'm partly reading this answer (and also my own textbook, where this claim seems to be made) and it seems to hold for ##d=1## with the help of the mean value theorem and the assumption that ##f## is continuously differentiable, but I have no clue how one would go about it for ##d>1##, since I don't think any mean value theorem/inequality holds for vector-valued functions of a vector variable. If you know a proof, I'd be happy to hear about it. :smile:
 
Physics news on Phys.org
psie said:
since I don't think any mean value theorem/inequality holds for vector-valued functions of a vector variable.
There is the mean value theorem for vector valued functions. See https://mathresearch.utsa.edu/wiki/index.php?title=Mean-Value_Theorems_for_Vector_Valued_Functions.

If you look at their proof you will see that you can, in similar way, define a new function ## F(t)=\textbf{a}\textbf{f}(\textbf{x}+t\textbf{J}) ## where ## \textbf{a}\in\mathbb{R}^m ## and ## \textbf{J}\in\mathbb{R}^n ## is a vector of ones. ## F’(t) ## will be ## \textbf{a}\textbf{f}’(\textbf{x}+t\textbf{J})\textbf{J} ##. By using $$ |\frac{F(t+h)-F(t)}{h}-F’(t)|\lt\epsilon $$ for ## t=0 ## you can get the next $$ |\frac{\textbf{a}\textbf{f}(\textbf{x}+h\textbf{J})-\textbf{a}\textbf{f}(\textbf{x})}{h}-\textbf{a}\textbf{f}'(\textbf{x})\textbf{J}|\lt\epsilon $$.
 
We all know the definition of n-dimensional topological manifold uses open sets and homeomorphisms onto the image as open set in ##\mathbb R^n##. It should be possible to reformulate the definition of n-dimensional topological manifold using closed sets on the manifold's topology and on ##\mathbb R^n## ? I'm positive for this. Perhaps the definition of smooth manifold would be problematic, though.

Similar threads

  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 0 ·
Replies
0
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K
  • · Replies 11 ·
Replies
11
Views
2K
  • · Replies 22 ·
Replies
22
Views
4K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 11 ·
Replies
11
Views
2K
  • · Replies 4 ·
Replies
4
Views
3K
  • · Replies 5 ·
Replies
5
Views
960