MHB How is Equation (1) Equivalent to the Derivative Definition in Theorem 7.1?

Math Amateur
Gold Member
MHB
Messages
3,920
Reaction score
48
I am reading the book: "Theory of Functions of a Complex Variable" by A. I. Markushevich (Part 1) ...

I need some help with an aspect of the proof of Theorem 7.1 ...The statement of Theorem 7.1 reads as follows:
View attachment 9330At the start of the above proof by Markushevich we read the following:

"If $$f(z)$$ has a derivative $$f'_E(z_0)$$ at $$z_0$$, then by definition

$$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$

where $$\epsilon ( z, z_0 ) \to 0$$ as $$\Delta z \to 0$$. ... ... "Now previously in Equation 7.1 at the start of Chapter 7, Markushevich has defined $$f'_E(z_0)$$ as follows:

$$f'_E(z_0) = \frac{ f(z) - f(z_0) }{ z - z_0 } = \frac{ \Delta_E f(z) }{ \Delta z }$$ ... ... ... (1)How exactly (formally and rigorously) is equation (1) exactly the same as $$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$ ...... strictly speaking, shouldn't Markushevich be deriving $$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$ ... from equation (1) ... Peter
 

Attachments

  • Markushevich - Theorem 7.1 and Proof ... .png
    Markushevich - Theorem 7.1 and Proof ... .png
    22 KB · Views: 154
Physics news on Phys.org
Peter said:
I am reading the book: "Theory of Functions of a Complex Variable" by A. I. Markushevich (Part 1) ...
At the start of the above proof by Markushevich we read the following:

"If $$f(z)$$ has a derivative $$f'_E(z_0)$$ at $$z_0$$, then by definition

$$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$

where $$\epsilon ( z, z_0 ) \to 0$$ as $$\Delta z \to 0$$. ... ... "Now previously in Equation 7.1 at the start of Chapter 7, Markushevich has defined $$f'_E(z_0)$$ as follows:

$$f'_E(z_0) = \frac{ f(z) - f(z_0) }{ z - z_0 } = \frac{ \Delta_E f(z) }{ \Delta z }$$ ... ... ... (1)How exactly (formally and rigorously) is equation (1) exactly the same as $$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$ ...... strictly speaking, shouldn't Markushevich be deriving $$\frac{ \Delta_E f(z) }{ \Delta z } = f'_E(z_0) + \epsilon ( z, z_0 )$$ ... from equation (1) ... Peter

I think I know what is being said, though the notation is a bit all over the place here. E.g. you wrote
$$f'_E(z_0) = \frac{ f(z) - f(z_0) }{ z - z_0 } = \frac{ \Delta_E f(z) }{ \Delta z }$$
but I didn't see a limit explicitly taken so it doesn't seem like a definition of a derivative that I'm familiar with.

Anyway, as is often the case, why not try to estimate the difference between those two definitions, and use the oh so important triangle inequality.

This gives

$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - \big(f'_E(z_0) + \epsilon ( z, z_0 )\big) \Big \vert \leq \Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert + \Big \vert \epsilon ( z, z_0 ) \Big \vert $

by triangle inequality. Now pass limits
i.e. consider for any $\epsilon \gt 0$ we can select a $\delta_1$ neighborhood (i.e. for all $\big \vert z - z_0 \big \vert \lt \delta_1$) such that
$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert \lt \frac{\epsilon}{2}$

and $\delta_2$ neighborhood such that
$ \Big \vert \epsilon ( z, z_0 ) \Big \vert \lt \frac{\epsilon}{2}$

and select $\delta = \min \big(\delta_1, \delta_2\big)$ so you have

$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - \big(f'_E(z_0) + \epsilon ( z, z_0 )\big) \Big \vert \leq \Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert + \Big \vert \epsilon ( z, z_0 ) \Big \vert \lt \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon$
 
steep said:
I think I know what is being said, though the notation is a bit all over the place here. E.g. you wrote
$$f'_E(z_0) = \frac{ f(z) - f(z_0) }{ z - z_0 } = \frac{ \Delta_E f(z) }{ \Delta z }$$
but I didn't see a limit explicitly taken so it doesn't seem like a definition of a derivative that I'm familiar with.

Anyway, as is often the case, why not try to estimate the difference between those two definitions, and use the oh so important triangle inequality.

This gives

$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - \big(f'_E(z_0) + \epsilon ( z, z_0 )\big) \Big \vert \leq \Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert + \Big \vert \epsilon ( z, z_0 ) \Big \vert $

by triangle inequality. Now pass limits
i.e. consider for any $\epsilon \gt 0$ we can select a $\delta_1$ neighborhood (i.e. for all $\big \vert z - z_0 \big \vert \lt \delta_1$) such that
$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert \lt \frac{\epsilon}{2}$

and $\delta_2$ neighborhood such that
$ \Big \vert \epsilon ( z, z_0 ) \Big \vert \lt \frac{\epsilon}{2}$

and select $\delta = \min \big(\delta_1, \delta_2\big)$ so you have

$\Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - \big(f'_E(z_0) + \epsilon ( z, z_0 )\big) \Big \vert \leq \Big \vert \frac{ f(z) - f(z_0) }{ z - z_0 } - f'_E(z_0) \Big \vert + \Big \vert \epsilon ( z, z_0 ) \Big \vert \lt \frac{\epsilon}{2} + \frac{\epsilon}{2} = \epsilon$
Hi steep...

Thanks so much for your post ...

I am still reflecting on what you have written ...I must apologize for a serious typo in equation (1) ...

I wrote ...$$f'_E(z_0) = \frac{ f(z) - f(z_0) }{ z - z_0 } = \frac{ \Delta_E f(z) }{ \Delta z }$$ ... ... ... (1)when I should have written $$ f'_E(z_0) = \lim_{ z \to z_0} \frac{ f(z) - f(z_0) }{ z - z_0 } = \lim_{ \Delta z \to 0} \frac{ \Delta_E f(z) }{ \Delta z }$$ ... ... ... (1)I should have also posted the beginnings of Markushevich's start to Chapter 7 to give readers access o his definitions ... so I am posting that now ... as follows:View attachment 9332
View attachment 9333
Hope that helps ...

Peter
 

Attachments

  • Markushevich - 1 - Start of Ch. 7, Section 28 ... PART 1 .png
    Markushevich - 1 - Start of Ch. 7, Section 28 ... PART 1 .png
    25.4 KB · Views: 115
  • Markushevich - 2 - Start of Ch. 7, Section 28 ... PART 2 ... .png
    Markushevich - 2 - Start of Ch. 7, Section 28 ... PART 2 ... .png
    17.5 KB · Views: 132
btw it seems worth pointing out that the (limiting) difference quotient is in some sense the 'original' derivative definition for single variable calc.

However the second definition introduced here is, in effect, that the derivative (if it exists) is the best linear approximation of a function over a sufficiently small neighborhood, period. This definition / interpretation is one that generalizes to higher dimensions. And since complex analysis is one variable analysis, but is 'kind of like' multivariable analysis ($\mathbb R^2$ looms), any perceived difference between the definitions is a good thing to dwell on.

Another thing-- a long overdue release / update to Beardon's "Complex Analysis: The Argument Principle" is coming next month as a Dover book. I think you may have a bigger math library than me but I thought I'd mention it.
 
steep said:
btw it seems worth pointing out that the (limiting) difference quotient is in some sense the 'original' derivative definition for single variable calc.

However the second definition introduced here is, in effect, that the derivative (if it exists) is the best linear approximation of a function over a sufficiently small neighborhood, period. This definition / interpretation is one that generalizes to higher dimensions. And since complex analysis is one variable analysis, but is 'kind of like' multivariable analysis ($\mathbb R^2$ looms), any perceived difference between the definitions is a good thing to dwell on.

Another thing-- a long overdue release / update to Beardon's "Complex Analysis: The Argument Principle" is coming next month as a Dover book. I think you may have a bigger math library than me but I thought I'd mention it.

Thanks for your most helpful posts, steep ...

I'll definitely keep a watch out for the release of Beardon's book ... I find being able consult a number of texts treatment of mathematical topics is helpful to learning ...

Thanks again ...

Peter
 
Back
Top