How is Equation (1) Equivalent to the Derivative Definition in Theorem 7.1?

Click For Summary
SUMMARY

The discussion centers on the equivalence of Equation (1) and the derivative definition in Theorem 7.1 from A. I. Markushevich's "Theory of Functions of a Complex Variable." Equation (1) states that the derivative \( f'_E(z_0) \) is defined as \( \frac{ f(z) - f(z_0) }{ z - z_0 } \), while the theorem asserts that \( \frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon(z, z_0) \) as \( \Delta z \to 0 \). The participants clarify that the limit must be explicitly taken to establish the derivative, emphasizing the importance of the triangle inequality in estimating differences between definitions. The conversation also highlights the generalization of derivative concepts in complex analysis.

PREREQUISITES
  • Understanding of complex functions and derivatives
  • Familiarity with limits and the epsilon-delta definition of continuity
  • Knowledge of the triangle inequality in mathematical analysis
  • Basic concepts of complex analysis as presented in Markushevich's work
NEXT STEPS
  • Study the epsilon-delta definition of limits in calculus
  • Explore the triangle inequality and its applications in analysis
  • Read Chapter 7 of "Theory of Functions of a Complex Variable" by A. I. Markushevich
  • Investigate the relationship between single-variable and multivariable derivatives
USEFUL FOR

Mathematics students, educators, and researchers interested in complex analysis, particularly those studying derivatives and their definitions in various contexts.

Math Amateur
Gold Member
MHB
Messages
3,920
Reaction score
48
I am reading the book: "Theory of Functions of a Complex Variable" by A. I. Markushevich (Part 1) ...

I need some help with an aspect of the proof of Theorem 7.1 ...The statement of Theorem 7.1 reads as follows:
View attachment 9330At the start of the above proof by Markushevich we read the following:

"If $$f(z)$$ has a derivative $$f'_E(z_0)$$ at $$z_0$$, then by definition

$$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$

where $$\epsilon ( z, z_0 ) \to 0$$ as $$\Delta z \to 0$$. ... ... "Now previously in Equation 7.1 at the start of Chapter 7, Markushevich has defined $$f'_E(z_0)$$ as follows:

$$f'_E(z_0) = \frac{ f(z) - f(z_0) }{ z - z_0 } = \frac{ \Delta_E f(z) }{ \Delta z }$$ ... ... ... (1)How exactly (formally and rigorously) is equation (1) exactly the same as $$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$ ...... strictly speaking, shouldn't Markushevich be deriving $$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$ ... from equation (1) ... Peter
 

Attachments

  • Markushevich - Theorem 7.1 and Proof ... .png
    Markushevich - Theorem 7.1 and Proof ... .png
    22 KB · Views: 167
Physics news on Phys.org
Peter said:
I am reading the book: "Theory of Functions of a Complex Variable" by A. I. Markushevich (Part 1) ...
At the start of the above proof by Markushevich we read the following:

"If $$f(z)$$ has a derivative $$f'_E(z_0)$$ at $$z_0$$, then by definition

$$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$

where $$\epsilon ( z, z_0 ) \to 0$$ as $$\Delta z \to 0$$. ... ... "Now previously in Equation 7.1 at the start of Chapter 7, Markushevich has defined $$f'_E(z_0)$$ as follows:

$$f'_E(z_0) = \frac{ f(z) - f(z_0) }{ z - z_0 } = \frac{ \Delta_E f(z) }{ \Delta z }$$ ... ... ... (1)How exactly (formally and rigorously) is equation (1) exactly the same as $$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$ ...... strictly speaking, shouldn't Markushevich be deriving $$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$ ... from equation (1) ... Peter

I think I know what is being said, though the notation is a bit all over the place here. E.g. you wrote
$$f'_E(z_0) = \frac{ f(z) - f(z_0) }{ z - z_0 } = \frac{ \Delta_E f(z) }{ \Delta z }$$
but I didn't see a limit explicitly taken so it doesn't seem like a definition of a derivative that I'm familiar with.

Anyway, as is often the case, why not try to estimate the difference between those two definitions, and use the oh so important triangle inequality.

This gives

$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - \big(f'_E(z_0) + \epsilon ( z, z_0 )\big) \Big \vert \leq \Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert + \Big \vert \epsilon ( z, z_0 ) \Big \vert $

by triangle inequality. Now pass limits
i.e. consider for any $\epsilon \gt 0$ we can select a $\delta_1$ neighborhood (i.e. for all $\big \vert z - z_0 \big \vert \lt \delta_1$) such that
$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert \lt \frac{\epsilon}{2}$

and $\delta_2$ neighborhood such that
$ \Big \vert \epsilon ( z, z_0 ) \Big \vert \lt \frac{\epsilon}{2}$

and select $\delta = \min \big(\delta_1, \delta_2\big)$ so you have

$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - \big(f'_E(z_0) + \epsilon ( z, z_0 )\big) \Big \vert \leq \Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert + \Big \vert \epsilon ( z, z_0 ) \Big \vert \lt \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon$
 
steep said:
I think I know what is being said, though the notation is a bit all over the place here. E.g. you wrote
$$f'_E(z_0) = \frac{ f(z) - f(z_0) }{ z - z_0 } = \frac{ \Delta_E f(z) }{ \Delta z }$$
but I didn't see a limit explicitly taken so it doesn't seem like a definition of a derivative that I'm familiar with.

Anyway, as is often the case, why not try to estimate the difference between those two definitions, and use the oh so important triangle inequality.

This gives

$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - \big(f'_E(z_0) + \epsilon ( z, z_0 )\big) \Big \vert \leq \Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert + \Big \vert \epsilon ( z, z_0 ) \Big \vert $

by triangle inequality. Now pass limits
i.e. consider for any $\epsilon \gt 0$ we can select a $\delta_1$ neighborhood (i.e. for all $\big \vert z - z_0 \big \vert \lt \delta_1$) such that
$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert \lt \frac{\epsilon}{2}$

and $\delta_2$ neighborhood such that
$ \Big \vert \epsilon ( z, z_0 ) \Big \vert \lt \frac{\epsilon}{2}$

and select $\delta = \min \big(\delta_1, \delta_2\big)$ so you have

$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - \big(f'_E(z_0) + \epsilon ( z, z_0 )\big) \Big \vert \leq \Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert + \Big \vert \epsilon ( z, z_0 ) \Big \vert \lt \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon$
Hi steep...

Thanks so much for your post ...

I am still reflecting on what you have written ...I must apologize for a serious typo in equation (1) ...

I wrote ...$$f'_E(z_0) = \frac{ f(z) - f(z_0) }{ z - z_0 } = \frac{ \Delta_E f(z) }{ \Delta z }$$ ... ... ... (1)when I should have written $$ f'_E(z_0) = \lim_{ z \to z_0} \frac{ f(z) - f(z_0) }{ z - z_0 } = \lim_{ \Delta z \to 0} \frac{ \Delta_E f(z) }{ \Delta z }$$ ... ... ... (1)I should have also posted the beginnings of Markushevich's start to Chapter 7 to give readers access o his definitions ... so I am posting that now ... as follows:View attachment 9332
View attachment 9333
Hope that helps ...

Peter
 

Attachments

  • Markushevich - 1 - Start of Ch. 7, Section 28 ... PART 1 .png
    Markushevich - 1 - Start of Ch. 7, Section 28 ... PART 1 .png
    25.4 KB · Views: 132
  • Markushevich - 2 - Start of Ch. 7, Section 28 ... PART 2 ... .png
    Markushevich - 2 - Start of Ch. 7, Section 28 ... PART 2 ... .png
    17.5 KB · Views: 152
btw it seems worth pointing out that the (limiting) difference quotient is in some sense the 'original' derivative definition for single variable calc.

However the second definition introduced here is, in effect, that the derivative (if it exists) is the best linear approximation of a function over a sufficiently small neighborhood, period. This definition / interpretation is one that generalizes to higher dimensions. And since complex analysis is one variable analysis, but is 'kind of like' multivariable analysis ($\mathbb R^2$ looms), any perceived difference between the definitions is a good thing to dwell on.

Another thing-- a long overdue release / update to Beardon's "Complex Analysis: The Argument Principle" is coming next month as a Dover book. I think you may have a bigger math library than me but I thought I'd mention it.
 
steep said:
btw it seems worth pointing out that the (limiting) difference quotient is in some sense the 'original' derivative definition for single variable calc.

However the second definition introduced here is, in effect, that the derivative (if it exists) is the best linear approximation of a function over a sufficiently small neighborhood, period. This definition / interpretation is one that generalizes to higher dimensions. And since complex analysis is one variable analysis, but is 'kind of like' multivariable analysis ($\mathbb R^2$ looms), any perceived difference between the definitions is a good thing to dwell on.

Another thing-- a long overdue release / update to Beardon's "Complex Analysis: The Argument Principle" is coming next month as a Dover book. I think you may have a bigger math library than me but I thought I'd mention it.

Thanks for your most helpful posts, steep ...

I'll definitely keep a watch out for the release of Beardon's book ... I find being able consult a number of texts treatment of mathematical topics is helpful to learning ...

Thanks again ...

Peter
 

Similar threads

  • · Replies 0 ·
Replies
0
Views
2K
  • · Replies 1 ·
Replies
1
Views
1K
  • · Replies 7 ·
Replies
7
Views
2K
Replies
2
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 4 ·
Replies
4
Views
4K
  • · Replies 2 ·
Replies
2
Views
4K
Replies
3
Views
2K
  • · Replies 2 ·
Replies
2
Views
2K
  • · Replies 1 ·
Replies
1
Views
2K